00:00:00.000 Started by upstream project "autotest-spdk-master-vs-dpdk-v23.11" build number 1041 00:00:00.000 originally caused by: 00:00:00.000 Started by upstream project "nightly-trigger" build number 3703 00:00:00.000 originally caused by: 00:00:00.000 Started by timer 00:00:00.000 Started by timer 00:00:00.020 Checking out git https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool into /var/jenkins_home/workspace/nvmf-tcp-phy-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4 to read jbp/jenkins/jjb-config/jobs/autotest-downstream/autotest-phy.groovy 00:00:00.021 The recommended git tool is: git 00:00:00.021 using credential 00000000-0000-0000-0000-000000000002 00:00:00.024 > git rev-parse --resolve-git-dir /var/jenkins_home/workspace/nvmf-tcp-phy-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4/jbp/.git # timeout=10 00:00:00.036 Fetching changes from the remote Git repository 00:00:00.039 > git config remote.origin.url https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool # timeout=10 00:00:00.051 Using shallow fetch with depth 1 00:00:00.051 Fetching upstream changes from https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool 00:00:00.051 > git --version # timeout=10 00:00:00.066 > git --version # 'git version 2.39.2' 00:00:00.066 using GIT_ASKPASS to set credentials SPDKCI HTTPS Credentials 00:00:00.083 Setting http proxy: proxy-dmz.intel.com:911 00:00:00.083 > git fetch --tags --force --progress --depth=1 -- https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool refs/heads/master # timeout=5 00:00:02.514 > git rev-parse origin/FETCH_HEAD^{commit} # timeout=10 00:00:02.525 > git rev-parse FETCH_HEAD^{commit} # timeout=10 00:00:02.537 Checking out Revision db4637e8b949f278f369ec13f70585206ccd9507 (FETCH_HEAD) 00:00:02.537 > git config core.sparsecheckout # timeout=10 00:00:02.547 > git read-tree -mu HEAD # timeout=10 00:00:02.563 > git checkout -f db4637e8b949f278f369ec13f70585206ccd9507 # timeout=5 00:00:02.589 Commit message: "jenkins/jjb-config: Add missing SPDK_TEST_NVME_INTERRUPT flag" 00:00:02.589 > git rev-list --no-walk db4637e8b949f278f369ec13f70585206ccd9507 # timeout=10 00:00:02.709 [Pipeline] Start of Pipeline 00:00:02.723 [Pipeline] library 00:00:02.724 Loading library shm_lib@master 00:00:02.724 Library shm_lib@master is cached. Copying from home. 00:00:02.740 [Pipeline] node 00:00:02.757 Running on GP11 in /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:00:02.759 [Pipeline] { 00:00:02.769 [Pipeline] catchError 00:00:02.771 [Pipeline] { 00:00:02.783 [Pipeline] wrap 00:00:02.792 [Pipeline] { 00:00:02.799 [Pipeline] stage 00:00:02.801 [Pipeline] { (Prologue) 00:00:03.027 [Pipeline] sh 00:00:03.713 + logger -p user.info -t JENKINS-CI 00:00:03.746 [Pipeline] echo 00:00:03.747 Node: GP11 00:00:03.754 [Pipeline] sh 00:00:04.101 [Pipeline] setCustomBuildProperty 00:00:04.111 [Pipeline] echo 00:00:04.113 Cleanup processes 00:00:04.118 [Pipeline] sh 00:00:04.406 + sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:00:04.406 14891 sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:00:04.437 [Pipeline] sh 00:00:04.731 ++ sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:00:04.731 ++ grep -v 'sudo pgrep' 00:00:04.731 ++ awk '{print $1}' 00:00:04.731 + sudo kill -9 00:00:04.731 + true 00:00:04.744 [Pipeline] cleanWs 00:00:04.754 [WS-CLEANUP] Deleting project workspace... 00:00:04.754 [WS-CLEANUP] Deferred wipeout is used... 00:00:04.765 [WS-CLEANUP] done 00:00:04.770 [Pipeline] setCustomBuildProperty 00:00:04.783 [Pipeline] sh 00:00:05.071 + sudo git config --global --replace-all safe.directory '*' 00:00:05.182 [Pipeline] httpRequest 00:00:06.721 [Pipeline] echo 00:00:06.722 Sorcerer 10.211.164.101 is alive 00:00:06.729 [Pipeline] retry 00:00:06.730 [Pipeline] { 00:00:06.740 [Pipeline] httpRequest 00:00:06.745 HttpMethod: GET 00:00:06.745 URL: http://10.211.164.101/packages/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:06.746 Sending request to url: http://10.211.164.101/packages/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:06.756 Response Code: HTTP/1.1 200 OK 00:00:06.756 Success: Status code 200 is in the accepted range: 200,404 00:00:06.757 Saving response body to /var/jenkins/workspace/nvmf-tcp-phy-autotest/jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:09.265 [Pipeline] } 00:00:09.277 [Pipeline] // retry 00:00:09.282 [Pipeline] sh 00:00:09.573 + tar --no-same-owner -xf jbp_db4637e8b949f278f369ec13f70585206ccd9507.tar.gz 00:00:09.588 [Pipeline] httpRequest 00:00:09.949 [Pipeline] echo 00:00:09.950 Sorcerer 10.211.164.101 is alive 00:00:09.960 [Pipeline] retry 00:00:09.962 [Pipeline] { 00:00:09.971 [Pipeline] httpRequest 00:00:09.975 HttpMethod: GET 00:00:09.976 URL: http://10.211.164.101/packages/spdk_a5e6ecf28fd8e9a86690362af173cd2cf51891ee.tar.gz 00:00:09.977 Sending request to url: http://10.211.164.101/packages/spdk_a5e6ecf28fd8e9a86690362af173cd2cf51891ee.tar.gz 00:00:10.000 Response Code: HTTP/1.1 200 OK 00:00:10.000 Success: Status code 200 is in the accepted range: 200,404 00:00:10.001 Saving response body to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk_a5e6ecf28fd8e9a86690362af173cd2cf51891ee.tar.gz 00:04:27.349 [Pipeline] } 00:04:27.369 [Pipeline] // retry 00:04:27.377 [Pipeline] sh 00:04:27.676 + tar --no-same-owner -xf spdk_a5e6ecf28fd8e9a86690362af173cd2cf51891ee.tar.gz 00:04:31.011 [Pipeline] sh 00:04:31.308 + git -C spdk log --oneline -n5 00:04:31.308 a5e6ecf28 lib/reduce: Data copy logic in thin read operations 00:04:31.308 a333974e5 nvme/rdma: Flush queued send WRs when disconnecting a qpair 00:04:31.308 2b8672176 nvme/rdma: Prevent submitting new recv WR when disconnecting 00:04:31.308 e2dfdf06c accel/mlx5: Register post_poller handler 00:04:31.308 3c8001115 accel/mlx5: More precise condition to update DB 00:04:31.329 [Pipeline] withCredentials 00:04:31.342 > git --version # timeout=10 00:04:31.354 > git --version # 'git version 2.39.2' 00:04:31.381 Masking supported pattern matches of $GIT_PASSWORD or $GIT_ASKPASS 00:04:31.383 [Pipeline] { 00:04:31.395 [Pipeline] retry 00:04:31.397 [Pipeline] { 00:04:31.415 [Pipeline] sh 00:04:32.019 + git ls-remote http://dpdk.org/git/dpdk-stable v23.11 00:04:32.292 [Pipeline] } 00:04:32.309 [Pipeline] // retry 00:04:32.313 [Pipeline] } 00:04:32.329 [Pipeline] // withCredentials 00:04:32.335 [Pipeline] httpRequest 00:04:32.738 [Pipeline] echo 00:04:32.739 Sorcerer 10.211.164.101 is alive 00:04:32.749 [Pipeline] retry 00:04:32.751 [Pipeline] { 00:04:32.764 [Pipeline] httpRequest 00:04:32.769 HttpMethod: GET 00:04:32.770 URL: http://10.211.164.101/packages/dpdk_d15625009dced269fcec27fc81dd74fd58d54cdb.tar.gz 00:04:32.771 Sending request to url: http://10.211.164.101/packages/dpdk_d15625009dced269fcec27fc81dd74fd58d54cdb.tar.gz 00:04:32.775 Response Code: HTTP/1.1 200 OK 00:04:32.776 Success: Status code 200 is in the accepted range: 200,404 00:04:32.776 Saving response body to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk_d15625009dced269fcec27fc81dd74fd58d54cdb.tar.gz 00:04:37.285 [Pipeline] } 00:04:37.302 [Pipeline] // retry 00:04:37.309 [Pipeline] sh 00:04:37.607 + tar --no-same-owner -xf dpdk_d15625009dced269fcec27fc81dd74fd58d54cdb.tar.gz 00:04:39.004 [Pipeline] sh 00:04:39.297 + git -C dpdk log --oneline -n5 00:04:39.297 eeb0605f11 version: 23.11.0 00:04:39.297 238778122a doc: update release notes for 23.11 00:04:39.297 46aa6b3cfc doc: fix description of RSS features 00:04:39.297 dd88f51a57 devtools: forbid DPDK API in cnxk base driver 00:04:39.297 7e421ae345 devtools: support skipping forbid rule check 00:04:39.309 [Pipeline] } 00:04:39.324 [Pipeline] // stage 00:04:39.332 [Pipeline] stage 00:04:39.335 [Pipeline] { (Prepare) 00:04:39.354 [Pipeline] writeFile 00:04:39.368 [Pipeline] sh 00:04:39.659 + logger -p user.info -t JENKINS-CI 00:04:39.672 [Pipeline] sh 00:04:39.959 + logger -p user.info -t JENKINS-CI 00:04:39.971 [Pipeline] sh 00:04:40.257 + cat autorun-spdk.conf 00:04:40.257 SPDK_RUN_FUNCTIONAL_TEST=1 00:04:40.257 SPDK_TEST_NVMF=1 00:04:40.257 SPDK_TEST_NVME_CLI=1 00:04:40.257 SPDK_TEST_NVMF_TRANSPORT=tcp 00:04:40.257 SPDK_TEST_NVMF_NICS=e810 00:04:40.257 SPDK_TEST_VFIOUSER=1 00:04:40.257 SPDK_RUN_UBSAN=1 00:04:40.257 NET_TYPE=phy 00:04:40.257 SPDK_TEST_NATIVE_DPDK=v23.11 00:04:40.257 SPDK_RUN_EXTERNAL_DPDK=/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build 00:04:40.266 RUN_NIGHTLY=1 00:04:40.270 [Pipeline] readFile 00:04:40.313 [Pipeline] withEnv 00:04:40.316 [Pipeline] { 00:04:40.327 [Pipeline] sh 00:04:40.620 + set -ex 00:04:40.620 + [[ -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/autorun-spdk.conf ]] 00:04:40.620 + source /var/jenkins/workspace/nvmf-tcp-phy-autotest/autorun-spdk.conf 00:04:40.620 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:04:40.620 ++ SPDK_TEST_NVMF=1 00:04:40.620 ++ SPDK_TEST_NVME_CLI=1 00:04:40.620 ++ SPDK_TEST_NVMF_TRANSPORT=tcp 00:04:40.620 ++ SPDK_TEST_NVMF_NICS=e810 00:04:40.620 ++ SPDK_TEST_VFIOUSER=1 00:04:40.620 ++ SPDK_RUN_UBSAN=1 00:04:40.620 ++ NET_TYPE=phy 00:04:40.620 ++ SPDK_TEST_NATIVE_DPDK=v23.11 00:04:40.620 ++ SPDK_RUN_EXTERNAL_DPDK=/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build 00:04:40.620 ++ RUN_NIGHTLY=1 00:04:40.620 + case $SPDK_TEST_NVMF_NICS in 00:04:40.620 + DRIVERS=ice 00:04:40.620 + [[ tcp == \r\d\m\a ]] 00:04:40.620 + [[ -n ice ]] 00:04:40.620 + sudo rmmod mlx4_ib mlx5_ib irdma i40iw iw_cxgb4 00:04:40.620 rmmod: ERROR: Module mlx4_ib is not currently loaded 00:04:43.930 rmmod: ERROR: Module irdma is not currently loaded 00:04:43.930 rmmod: ERROR: Module i40iw is not currently loaded 00:04:43.930 rmmod: ERROR: Module iw_cxgb4 is not currently loaded 00:04:43.930 + true 00:04:43.930 + for D in $DRIVERS 00:04:43.930 + sudo modprobe ice 00:04:43.930 + exit 0 00:04:43.940 [Pipeline] } 00:04:43.954 [Pipeline] // withEnv 00:04:43.958 [Pipeline] } 00:04:43.972 [Pipeline] // stage 00:04:43.981 [Pipeline] catchError 00:04:43.983 [Pipeline] { 00:04:43.997 [Pipeline] timeout 00:04:43.997 Timeout set to expire in 1 hr 0 min 00:04:43.999 [Pipeline] { 00:04:44.014 [Pipeline] stage 00:04:44.016 [Pipeline] { (Tests) 00:04:44.031 [Pipeline] sh 00:04:44.324 + jbp/jenkins/jjb-config/jobs/scripts/autoruner.sh /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:04:44.324 ++ readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:04:44.324 + DIR_ROOT=/var/jenkins/workspace/nvmf-tcp-phy-autotest 00:04:44.324 + [[ -n /var/jenkins/workspace/nvmf-tcp-phy-autotest ]] 00:04:44.324 + DIR_SPDK=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:04:44.324 + DIR_OUTPUT=/var/jenkins/workspace/nvmf-tcp-phy-autotest/output 00:04:44.324 + [[ -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk ]] 00:04:44.324 + [[ ! -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/output ]] 00:04:44.324 + mkdir -p /var/jenkins/workspace/nvmf-tcp-phy-autotest/output 00:04:44.324 + [[ -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/output ]] 00:04:44.324 + [[ nvmf-tcp-phy-autotest == pkgdep-* ]] 00:04:44.324 + cd /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:04:44.324 + source /etc/os-release 00:04:44.324 ++ NAME='Fedora Linux' 00:04:44.324 ++ VERSION='39 (Cloud Edition)' 00:04:44.324 ++ ID=fedora 00:04:44.324 ++ VERSION_ID=39 00:04:44.324 ++ VERSION_CODENAME= 00:04:44.324 ++ PLATFORM_ID=platform:f39 00:04:44.324 ++ PRETTY_NAME='Fedora Linux 39 (Cloud Edition)' 00:04:44.324 ++ ANSI_COLOR='0;38;2;60;110;180' 00:04:44.324 ++ LOGO=fedora-logo-icon 00:04:44.324 ++ CPE_NAME=cpe:/o:fedoraproject:fedora:39 00:04:44.324 ++ HOME_URL=https://fedoraproject.org/ 00:04:44.324 ++ DOCUMENTATION_URL=https://docs.fedoraproject.org/en-US/fedora/f39/system-administrators-guide/ 00:04:44.324 ++ SUPPORT_URL=https://ask.fedoraproject.org/ 00:04:44.324 ++ BUG_REPORT_URL=https://bugzilla.redhat.com/ 00:04:44.324 ++ REDHAT_BUGZILLA_PRODUCT=Fedora 00:04:44.324 ++ REDHAT_BUGZILLA_PRODUCT_VERSION=39 00:04:44.324 ++ REDHAT_SUPPORT_PRODUCT=Fedora 00:04:44.324 ++ REDHAT_SUPPORT_PRODUCT_VERSION=39 00:04:44.324 ++ SUPPORT_END=2024-11-12 00:04:44.324 ++ VARIANT='Cloud Edition' 00:04:44.324 ++ VARIANT_ID=cloud 00:04:44.324 + uname -a 00:04:44.324 Linux spdk-gp-11 6.8.9-200.fc39.x86_64 #1 SMP PREEMPT_DYNAMIC Wed Jul 24 03:04:40 UTC 2024 x86_64 GNU/Linux 00:04:44.324 + sudo /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh status 00:04:45.284 Hugepages 00:04:45.284 node hugesize free / total 00:04:45.284 node0 1048576kB 0 / 0 00:04:45.284 node0 2048kB 0 / 0 00:04:45.284 node1 1048576kB 0 / 0 00:04:45.284 node1 2048kB 0 / 0 00:04:45.284 00:04:45.284 Type BDF Vendor Device NUMA Driver Device Block devices 00:04:45.284 I/OAT 0000:00:04.0 8086 0e20 0 ioatdma - - 00:04:45.284 I/OAT 0000:00:04.1 8086 0e21 0 ioatdma - - 00:04:45.284 I/OAT 0000:00:04.2 8086 0e22 0 ioatdma - - 00:04:45.284 I/OAT 0000:00:04.3 8086 0e23 0 ioatdma - - 00:04:45.284 I/OAT 0000:00:04.4 8086 0e24 0 ioatdma - - 00:04:45.284 I/OAT 0000:00:04.5 8086 0e25 0 ioatdma - - 00:04:45.284 I/OAT 0000:00:04.6 8086 0e26 0 ioatdma - - 00:04:45.284 I/OAT 0000:00:04.7 8086 0e27 0 ioatdma - - 00:04:45.284 I/OAT 0000:80:04.0 8086 0e20 1 ioatdma - - 00:04:45.284 I/OAT 0000:80:04.1 8086 0e21 1 ioatdma - - 00:04:45.284 I/OAT 0000:80:04.2 8086 0e22 1 ioatdma - - 00:04:45.284 I/OAT 0000:80:04.3 8086 0e23 1 ioatdma - - 00:04:45.284 I/OAT 0000:80:04.4 8086 0e24 1 ioatdma - - 00:04:45.284 I/OAT 0000:80:04.5 8086 0e25 1 ioatdma - - 00:04:45.284 I/OAT 0000:80:04.6 8086 0e26 1 ioatdma - - 00:04:45.284 I/OAT 0000:80:04.7 8086 0e27 1 ioatdma - - 00:04:45.544 NVMe 0000:88:00.0 8086 0a54 1 nvme nvme0 nvme0n1 00:04:45.544 + rm -f /tmp/spdk-ld-path 00:04:45.544 + source autorun-spdk.conf 00:04:45.544 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:04:45.544 ++ SPDK_TEST_NVMF=1 00:04:45.544 ++ SPDK_TEST_NVME_CLI=1 00:04:45.544 ++ SPDK_TEST_NVMF_TRANSPORT=tcp 00:04:45.544 ++ SPDK_TEST_NVMF_NICS=e810 00:04:45.544 ++ SPDK_TEST_VFIOUSER=1 00:04:45.544 ++ SPDK_RUN_UBSAN=1 00:04:45.544 ++ NET_TYPE=phy 00:04:45.544 ++ SPDK_TEST_NATIVE_DPDK=v23.11 00:04:45.544 ++ SPDK_RUN_EXTERNAL_DPDK=/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build 00:04:45.544 ++ RUN_NIGHTLY=1 00:04:45.544 + (( SPDK_TEST_NVME_CMB == 1 || SPDK_TEST_NVME_PMR == 1 )) 00:04:45.544 + [[ -n '' ]] 00:04:45.544 + sudo git config --global --add safe.directory /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:04:45.544 + for M in /var/spdk/build-*-manifest.txt 00:04:45.544 + [[ -f /var/spdk/build-kernel-manifest.txt ]] 00:04:45.544 + cp /var/spdk/build-kernel-manifest.txt /var/jenkins/workspace/nvmf-tcp-phy-autotest/output/ 00:04:45.544 + for M in /var/spdk/build-*-manifest.txt 00:04:45.544 + [[ -f /var/spdk/build-pkg-manifest.txt ]] 00:04:45.544 + cp /var/spdk/build-pkg-manifest.txt /var/jenkins/workspace/nvmf-tcp-phy-autotest/output/ 00:04:45.544 + for M in /var/spdk/build-*-manifest.txt 00:04:45.544 + [[ -f /var/spdk/build-repo-manifest.txt ]] 00:04:45.544 + cp /var/spdk/build-repo-manifest.txt /var/jenkins/workspace/nvmf-tcp-phy-autotest/output/ 00:04:45.544 ++ uname 00:04:45.544 + [[ Linux == \L\i\n\u\x ]] 00:04:45.544 + sudo dmesg -T 00:04:45.544 + sudo dmesg --clear 00:04:45.544 + dmesg_pid=16232 00:04:45.544 + [[ Fedora Linux == FreeBSD ]] 00:04:45.544 + export UNBIND_ENTIRE_IOMMU_GROUP=yes 00:04:45.544 + UNBIND_ENTIRE_IOMMU_GROUP=yes 00:04:45.544 + [[ -e /var/spdk/dependencies/vhost/spdk_test_image.qcow2 ]] 00:04:45.544 + sudo dmesg -Tw 00:04:45.544 + [[ -x /usr/src/fio-static/fio ]] 00:04:45.544 + export FIO_BIN=/usr/src/fio-static/fio 00:04:45.544 + FIO_BIN=/usr/src/fio-static/fio 00:04:45.544 + [[ '' == \/\v\a\r\/\j\e\n\k\i\n\s\/\w\o\r\k\s\p\a\c\e\/\n\v\m\f\-\t\c\p\-\p\h\y\-\a\u\t\o\t\e\s\t\/\q\e\m\u\_\v\f\i\o\/* ]] 00:04:45.544 + [[ ! -v VFIO_QEMU_BIN ]] 00:04:45.544 + [[ -e /usr/local/qemu/vfio-user-latest ]] 00:04:45.544 + export VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:04:45.544 + VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:04:45.544 + [[ -e /usr/local/qemu/vanilla-latest ]] 00:04:45.544 + export QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:04:45.544 + QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:04:45.544 + spdk/autorun.sh /var/jenkins/workspace/nvmf-tcp-phy-autotest/autorun-spdk.conf 00:04:45.544 17:20:27 -- common/autotest_common.sh@1710 -- $ [[ n == y ]] 00:04:45.544 17:20:27 -- spdk/autorun.sh@20 -- $ source /var/jenkins/workspace/nvmf-tcp-phy-autotest/autorun-spdk.conf 00:04:45.544 17:20:27 -- nvmf-tcp-phy-autotest/autorun-spdk.conf@1 -- $ SPDK_RUN_FUNCTIONAL_TEST=1 00:04:45.544 17:20:27 -- nvmf-tcp-phy-autotest/autorun-spdk.conf@2 -- $ SPDK_TEST_NVMF=1 00:04:45.544 17:20:27 -- nvmf-tcp-phy-autotest/autorun-spdk.conf@3 -- $ SPDK_TEST_NVME_CLI=1 00:04:45.544 17:20:27 -- nvmf-tcp-phy-autotest/autorun-spdk.conf@4 -- $ SPDK_TEST_NVMF_TRANSPORT=tcp 00:04:45.544 17:20:27 -- nvmf-tcp-phy-autotest/autorun-spdk.conf@5 -- $ SPDK_TEST_NVMF_NICS=e810 00:04:45.544 17:20:27 -- nvmf-tcp-phy-autotest/autorun-spdk.conf@6 -- $ SPDK_TEST_VFIOUSER=1 00:04:45.544 17:20:27 -- nvmf-tcp-phy-autotest/autorun-spdk.conf@7 -- $ SPDK_RUN_UBSAN=1 00:04:45.544 17:20:27 -- nvmf-tcp-phy-autotest/autorun-spdk.conf@8 -- $ NET_TYPE=phy 00:04:45.544 17:20:27 -- nvmf-tcp-phy-autotest/autorun-spdk.conf@9 -- $ SPDK_TEST_NATIVE_DPDK=v23.11 00:04:45.544 17:20:27 -- nvmf-tcp-phy-autotest/autorun-spdk.conf@10 -- $ SPDK_RUN_EXTERNAL_DPDK=/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build 00:04:45.544 17:20:27 -- nvmf-tcp-phy-autotest/autorun-spdk.conf@11 -- $ RUN_NIGHTLY=1 00:04:45.544 17:20:27 -- spdk/autorun.sh@22 -- $ trap 'timing_finish || exit 1' EXIT 00:04:45.544 17:20:27 -- spdk/autorun.sh@25 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/autobuild.sh /var/jenkins/workspace/nvmf-tcp-phy-autotest/autorun-spdk.conf 00:04:45.544 17:20:27 -- common/autotest_common.sh@1710 -- $ [[ n == y ]] 00:04:45.544 17:20:27 -- common/autobuild_common.sh@15 -- $ source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:04:45.544 17:20:27 -- scripts/common.sh@15 -- $ shopt -s extglob 00:04:45.544 17:20:27 -- scripts/common.sh@544 -- $ [[ -e /bin/wpdk_common.sh ]] 00:04:45.544 17:20:27 -- scripts/common.sh@552 -- $ [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:04:45.544 17:20:27 -- scripts/common.sh@553 -- $ source /etc/opt/spdk-pkgdep/paths/export.sh 00:04:45.544 17:20:27 -- paths/export.sh@2 -- $ PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:45.544 17:20:27 -- paths/export.sh@3 -- $ PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:45.544 17:20:27 -- paths/export.sh@4 -- $ PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:45.544 17:20:27 -- paths/export.sh@5 -- $ export PATH 00:04:45.544 17:20:27 -- paths/export.sh@6 -- $ echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:45.544 17:20:27 -- common/autobuild_common.sh@492 -- $ out=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output 00:04:45.544 17:20:27 -- common/autobuild_common.sh@493 -- $ date +%s 00:04:45.544 17:20:27 -- common/autobuild_common.sh@493 -- $ mktemp -dt spdk_1733502027.XXXXXX 00:04:45.544 17:20:27 -- common/autobuild_common.sh@493 -- $ SPDK_WORKSPACE=/tmp/spdk_1733502027.XqRWWj 00:04:45.544 17:20:27 -- common/autobuild_common.sh@495 -- $ [[ -n '' ]] 00:04:45.544 17:20:27 -- common/autobuild_common.sh@499 -- $ '[' -n v23.11 ']' 00:04:45.544 17:20:27 -- common/autobuild_common.sh@500 -- $ dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build 00:04:45.544 17:20:27 -- common/autobuild_common.sh@500 -- $ scanbuild_exclude=' --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk' 00:04:45.544 17:20:27 -- common/autobuild_common.sh@506 -- $ scanbuild_exclude+=' --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/xnvme --exclude /tmp' 00:04:45.544 17:20:27 -- common/autobuild_common.sh@508 -- $ scanbuild='scan-build -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/scan-build-tmp --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/xnvme --exclude /tmp --status-bugs' 00:04:45.544 17:20:27 -- common/autobuild_common.sh@509 -- $ get_config_params 00:04:45.544 17:20:27 -- common/autotest_common.sh@409 -- $ xtrace_disable 00:04:45.544 17:20:27 -- common/autotest_common.sh@10 -- $ set +x 00:04:45.544 17:20:27 -- common/autobuild_common.sh@509 -- $ config_params='--enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-coverage --with-ublk --with-vfio-user --with-dpdk=/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build' 00:04:45.544 17:20:27 -- common/autobuild_common.sh@511 -- $ start_monitor_resources 00:04:45.544 17:20:27 -- pm/common@17 -- $ local monitor 00:04:45.544 17:20:27 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:04:45.544 17:20:27 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:04:45.544 17:20:27 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:04:45.544 17:20:27 -- pm/common@21 -- $ date +%s 00:04:45.544 17:20:27 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:04:45.544 17:20:27 -- pm/common@21 -- $ date +%s 00:04:45.544 17:20:27 -- pm/common@25 -- $ sleep 1 00:04:45.544 17:20:27 -- pm/common@21 -- $ date +%s 00:04:45.544 17:20:27 -- pm/common@21 -- $ date +%s 00:04:45.544 17:20:27 -- pm/common@21 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-load -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autobuild.sh.1733502027 00:04:45.544 17:20:27 -- pm/common@21 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-vmstat -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autobuild.sh.1733502027 00:04:45.544 17:20:27 -- pm/common@21 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-temp -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autobuild.sh.1733502027 00:04:45.544 17:20:27 -- pm/common@21 -- $ sudo -E /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-bmc-pm -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autobuild.sh.1733502027 00:04:45.805 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autobuild.sh.1733502027_collect-vmstat.pm.log 00:04:45.805 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autobuild.sh.1733502027_collect-cpu-temp.pm.log 00:04:45.805 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autobuild.sh.1733502027_collect-cpu-load.pm.log 00:04:45.805 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autobuild.sh.1733502027_collect-bmc-pm.bmc.pm.log 00:04:46.747 17:20:28 -- common/autobuild_common.sh@512 -- $ trap stop_monitor_resources EXIT 00:04:46.747 17:20:28 -- spdk/autobuild.sh@11 -- $ SPDK_TEST_AUTOBUILD= 00:04:46.747 17:20:28 -- spdk/autobuild.sh@12 -- $ umask 022 00:04:46.747 17:20:28 -- spdk/autobuild.sh@13 -- $ cd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:04:46.747 17:20:28 -- spdk/autobuild.sh@16 -- $ date -u 00:04:46.747 Fri Dec 6 04:20:28 PM UTC 2024 00:04:46.747 17:20:28 -- spdk/autobuild.sh@17 -- $ git describe --tags 00:04:46.747 v25.01-pre-303-ga5e6ecf28 00:04:46.747 17:20:28 -- spdk/autobuild.sh@19 -- $ '[' 0 -eq 1 ']' 00:04:46.747 17:20:28 -- spdk/autobuild.sh@23 -- $ '[' 1 -eq 1 ']' 00:04:46.747 17:20:28 -- spdk/autobuild.sh@24 -- $ run_test ubsan echo 'using ubsan' 00:04:46.747 17:20:28 -- common/autotest_common.sh@1105 -- $ '[' 3 -le 1 ']' 00:04:46.747 17:20:28 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:04:46.747 17:20:28 -- common/autotest_common.sh@10 -- $ set +x 00:04:46.747 ************************************ 00:04:46.747 START TEST ubsan 00:04:46.747 ************************************ 00:04:46.747 17:20:28 ubsan -- common/autotest_common.sh@1129 -- $ echo 'using ubsan' 00:04:46.747 using ubsan 00:04:46.747 00:04:46.747 real 0m0.000s 00:04:46.747 user 0m0.000s 00:04:46.747 sys 0m0.000s 00:04:46.747 17:20:28 ubsan -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:04:46.747 17:20:28 ubsan -- common/autotest_common.sh@10 -- $ set +x 00:04:46.747 ************************************ 00:04:46.747 END TEST ubsan 00:04:46.747 ************************************ 00:04:46.747 17:20:28 -- spdk/autobuild.sh@27 -- $ '[' -n v23.11 ']' 00:04:46.747 17:20:28 -- spdk/autobuild.sh@28 -- $ build_native_dpdk 00:04:46.747 17:20:28 -- common/autobuild_common.sh@449 -- $ run_test build_native_dpdk _build_native_dpdk 00:04:46.747 17:20:28 -- common/autotest_common.sh@1105 -- $ '[' 2 -le 1 ']' 00:04:46.747 17:20:28 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:04:46.747 17:20:28 -- common/autotest_common.sh@10 -- $ set +x 00:04:46.747 ************************************ 00:04:46.747 START TEST build_native_dpdk 00:04:46.747 ************************************ 00:04:46.747 17:20:28 build_native_dpdk -- common/autotest_common.sh@1129 -- $ _build_native_dpdk 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@48 -- $ local external_dpdk_dir 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@49 -- $ local external_dpdk_base_dir 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@50 -- $ local compiler_version 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@51 -- $ local compiler 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@52 -- $ local dpdk_kmods 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@53 -- $ local repo=dpdk 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@55 -- $ compiler=gcc 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@61 -- $ export CC=gcc 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@61 -- $ CC=gcc 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@63 -- $ [[ gcc != *clang* ]] 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@63 -- $ [[ gcc != *gcc* ]] 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@68 -- $ gcc -dumpversion 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@68 -- $ compiler_version=13 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@69 -- $ compiler_version=13 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@70 -- $ external_dpdk_dir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@71 -- $ dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@71 -- $ external_dpdk_base_dir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@73 -- $ [[ ! -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk ]] 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@82 -- $ orgdir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@83 -- $ git -C /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk log --oneline -n 5 00:04:46.747 eeb0605f11 version: 23.11.0 00:04:46.747 238778122a doc: update release notes for 23.11 00:04:46.747 46aa6b3cfc doc: fix description of RSS features 00:04:46.747 dd88f51a57 devtools: forbid DPDK API in cnxk base driver 00:04:46.747 7e421ae345 devtools: support skipping forbid rule check 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@85 -- $ dpdk_cflags='-fPIC -g -fcommon' 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@86 -- $ dpdk_ldflags= 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@87 -- $ dpdk_ver=23.11.0 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@89 -- $ [[ gcc == *gcc* ]] 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@89 -- $ [[ 13 -ge 5 ]] 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@90 -- $ dpdk_cflags+=' -Werror' 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@93 -- $ [[ gcc == *gcc* ]] 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@93 -- $ [[ 13 -ge 10 ]] 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@94 -- $ dpdk_cflags+=' -Wno-stringop-overflow' 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@102 -- $ DPDK_DRIVERS=("bus" "bus/pci" "bus/vdev" "mempool/ring" "net/i40e" "net/i40e/base" "power/acpi" "power/amd_pstate" "power/cppc" "power/intel_pstate" "power/intel_uncore" "power/kvm_vm") 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@103 -- $ local mlx5_libs_added=n 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@104 -- $ [[ 0 -eq 1 ]] 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@104 -- $ [[ 0 -eq 1 ]] 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@146 -- $ [[ 0 -eq 1 ]] 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@174 -- $ cd /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@175 -- $ uname -s 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@175 -- $ '[' Linux = Linux ']' 00:04:46.747 17:20:28 build_native_dpdk -- common/autobuild_common.sh@176 -- $ lt 23.11.0 21.11.0 00:04:46.747 17:20:28 build_native_dpdk -- scripts/common.sh@373 -- $ cmp_versions 23.11.0 '<' 21.11.0 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=<' 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=3 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@345 -- $ : 1 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 23 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@353 -- $ local d=23 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 23 =~ ^[0-9]+$ ]] 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@355 -- $ echo 23 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=23 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 21 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@353 -- $ local d=21 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 21 =~ ^[0-9]+$ ]] 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@355 -- $ echo 21 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=21 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@367 -- $ return 1 00:04:46.748 17:20:28 build_native_dpdk -- common/autobuild_common.sh@180 -- $ patch -p1 00:04:46.748 patching file config/rte_config.h 00:04:46.748 Hunk #1 succeeded at 60 (offset 1 line). 00:04:46.748 17:20:28 build_native_dpdk -- common/autobuild_common.sh@183 -- $ lt 23.11.0 24.07.0 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@373 -- $ cmp_versions 23.11.0 '<' 24.07.0 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=<' 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=3 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@345 -- $ : 1 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 23 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@353 -- $ local d=23 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 23 =~ ^[0-9]+$ ]] 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@355 -- $ echo 23 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=23 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 24 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@353 -- $ local d=24 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 24 =~ ^[0-9]+$ ]] 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@355 -- $ echo 24 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=24 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@368 -- $ (( ver1[v] < ver2[v] )) 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@368 -- $ return 0 00:04:46.748 17:20:28 build_native_dpdk -- common/autobuild_common.sh@184 -- $ patch -p1 00:04:46.748 patching file lib/pcapng/rte_pcapng.c 00:04:46.748 17:20:28 build_native_dpdk -- common/autobuild_common.sh@186 -- $ ge 23.11.0 24.07.0 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@376 -- $ cmp_versions 23.11.0 '>=' 24.07.0 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@336 -- $ IFS=.-: 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@336 -- $ read -ra ver1 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@337 -- $ IFS=.-: 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@337 -- $ read -ra ver2 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@338 -- $ local 'op=>=' 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@340 -- $ ver1_l=3 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@341 -- $ ver2_l=3 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@344 -- $ case "$op" in 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@348 -- $ : 1 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@364 -- $ (( v = 0 )) 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@365 -- $ decimal 23 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@353 -- $ local d=23 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 23 =~ ^[0-9]+$ ]] 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@355 -- $ echo 23 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@365 -- $ ver1[v]=23 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@366 -- $ decimal 24 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@353 -- $ local d=24 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@354 -- $ [[ 24 =~ ^[0-9]+$ ]] 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@355 -- $ echo 24 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@366 -- $ ver2[v]=24 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@368 -- $ (( ver1[v] < ver2[v] )) 00:04:46.748 17:20:28 build_native_dpdk -- scripts/common.sh@368 -- $ return 1 00:04:46.748 17:20:28 build_native_dpdk -- common/autobuild_common.sh@190 -- $ dpdk_kmods=false 00:04:46.748 17:20:28 build_native_dpdk -- common/autobuild_common.sh@191 -- $ uname -s 00:04:46.748 17:20:28 build_native_dpdk -- common/autobuild_common.sh@191 -- $ '[' Linux = FreeBSD ']' 00:04:46.748 17:20:28 build_native_dpdk -- common/autobuild_common.sh@195 -- $ printf %s, bus bus/pci bus/vdev mempool/ring net/i40e net/i40e/base power/acpi power/amd_pstate power/cppc power/intel_pstate power/intel_uncore power/kvm_vm 00:04:46.748 17:20:28 build_native_dpdk -- common/autobuild_common.sh@195 -- $ meson build-tmp --prefix=/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build --libdir lib -Denable_docs=false -Denable_kmods=false -Dtests=false -Dc_link_args= '-Dc_args=-fPIC -g -fcommon -Werror -Wno-stringop-overflow' -Dmachine=native -Denable_drivers=bus,bus/pci,bus/vdev,mempool/ring,net/i40e,net/i40e/base,power/acpi,power/amd_pstate,power/cppc,power/intel_pstate,power/intel_uncore,power/kvm_vm, 00:04:53.326 The Meson build system 00:04:53.326 Version: 1.5.0 00:04:53.326 Source dir: /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk 00:04:53.326 Build dir: /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build-tmp 00:04:53.326 Build type: native build 00:04:53.326 Program cat found: YES (/usr/bin/cat) 00:04:53.326 Project name: DPDK 00:04:53.326 Project version: 23.11.0 00:04:53.326 C compiler for the host machine: gcc (gcc 13.3.1 "gcc (GCC) 13.3.1 20240522 (Red Hat 13.3.1-1)") 00:04:53.326 C linker for the host machine: gcc ld.bfd 2.40-14 00:04:53.326 Host machine cpu family: x86_64 00:04:53.326 Host machine cpu: x86_64 00:04:53.326 Message: ## Building in Developer Mode ## 00:04:53.326 Program pkg-config found: YES (/usr/bin/pkg-config) 00:04:53.327 Program check-symbols.sh found: YES (/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/buildtools/check-symbols.sh) 00:04:53.327 Program options-ibverbs-static.sh found: YES (/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/buildtools/options-ibverbs-static.sh) 00:04:53.327 Program python3 found: YES (/usr/bin/python3) 00:04:53.327 Program cat found: YES (/usr/bin/cat) 00:04:53.327 config/meson.build:113: WARNING: The "machine" option is deprecated. Please use "cpu_instruction_set" instead. 00:04:53.327 Compiler for C supports arguments -march=native: YES 00:04:53.327 Checking for size of "void *" : 8 00:04:53.327 Checking for size of "void *" : 8 (cached) 00:04:53.327 Library m found: YES 00:04:53.327 Library numa found: YES 00:04:53.327 Has header "numaif.h" : YES 00:04:53.327 Library fdt found: NO 00:04:53.327 Library execinfo found: NO 00:04:53.327 Has header "execinfo.h" : YES 00:04:53.327 Found pkg-config: YES (/usr/bin/pkg-config) 1.9.5 00:04:53.327 Run-time dependency libarchive found: NO (tried pkgconfig) 00:04:53.327 Run-time dependency libbsd found: NO (tried pkgconfig) 00:04:53.327 Run-time dependency jansson found: NO (tried pkgconfig) 00:04:53.327 Run-time dependency openssl found: YES 3.1.1 00:04:53.327 Run-time dependency libpcap found: YES 1.10.4 00:04:53.327 Has header "pcap.h" with dependency libpcap: YES 00:04:53.327 Compiler for C supports arguments -Wcast-qual: YES 00:04:53.327 Compiler for C supports arguments -Wdeprecated: YES 00:04:53.327 Compiler for C supports arguments -Wformat: YES 00:04:53.327 Compiler for C supports arguments -Wformat-nonliteral: NO 00:04:53.327 Compiler for C supports arguments -Wformat-security: NO 00:04:53.327 Compiler for C supports arguments -Wmissing-declarations: YES 00:04:53.327 Compiler for C supports arguments -Wmissing-prototypes: YES 00:04:53.327 Compiler for C supports arguments -Wnested-externs: YES 00:04:53.327 Compiler for C supports arguments -Wold-style-definition: YES 00:04:53.327 Compiler for C supports arguments -Wpointer-arith: YES 00:04:53.327 Compiler for C supports arguments -Wsign-compare: YES 00:04:53.327 Compiler for C supports arguments -Wstrict-prototypes: YES 00:04:53.327 Compiler for C supports arguments -Wundef: YES 00:04:53.327 Compiler for C supports arguments -Wwrite-strings: YES 00:04:53.327 Compiler for C supports arguments -Wno-address-of-packed-member: YES 00:04:53.327 Compiler for C supports arguments -Wno-packed-not-aligned: YES 00:04:53.327 Compiler for C supports arguments -Wno-missing-field-initializers: YES 00:04:53.327 Compiler for C supports arguments -Wno-zero-length-bounds: YES 00:04:53.327 Program objdump found: YES (/usr/bin/objdump) 00:04:53.327 Compiler for C supports arguments -mavx512f: YES 00:04:53.327 Checking if "AVX512 checking" compiles: YES 00:04:53.327 Fetching value of define "__SSE4_2__" : 1 00:04:53.327 Fetching value of define "__AES__" : 1 00:04:53.327 Fetching value of define "__AVX__" : 1 00:04:53.327 Fetching value of define "__AVX2__" : (undefined) 00:04:53.327 Fetching value of define "__AVX512BW__" : (undefined) 00:04:53.327 Fetching value of define "__AVX512CD__" : (undefined) 00:04:53.327 Fetching value of define "__AVX512DQ__" : (undefined) 00:04:53.327 Fetching value of define "__AVX512F__" : (undefined) 00:04:53.327 Fetching value of define "__AVX512VL__" : (undefined) 00:04:53.327 Fetching value of define "__PCLMUL__" : 1 00:04:53.327 Fetching value of define "__RDRND__" : 1 00:04:53.327 Fetching value of define "__RDSEED__" : (undefined) 00:04:53.327 Fetching value of define "__VPCLMULQDQ__" : (undefined) 00:04:53.327 Fetching value of define "__znver1__" : (undefined) 00:04:53.327 Fetching value of define "__znver2__" : (undefined) 00:04:53.327 Fetching value of define "__znver3__" : (undefined) 00:04:53.327 Fetching value of define "__znver4__" : (undefined) 00:04:53.327 Compiler for C supports arguments -Wno-format-truncation: YES 00:04:53.327 Message: lib/log: Defining dependency "log" 00:04:53.327 Message: lib/kvargs: Defining dependency "kvargs" 00:04:53.327 Message: lib/telemetry: Defining dependency "telemetry" 00:04:53.327 Checking for function "getentropy" : NO 00:04:53.327 Message: lib/eal: Defining dependency "eal" 00:04:53.327 Message: lib/ring: Defining dependency "ring" 00:04:53.327 Message: lib/rcu: Defining dependency "rcu" 00:04:53.327 Message: lib/mempool: Defining dependency "mempool" 00:04:53.327 Message: lib/mbuf: Defining dependency "mbuf" 00:04:53.327 Fetching value of define "__PCLMUL__" : 1 (cached) 00:04:53.327 Fetching value of define "__AVX512F__" : (undefined) (cached) 00:04:53.327 Compiler for C supports arguments -mpclmul: YES 00:04:53.327 Compiler for C supports arguments -maes: YES 00:04:53.327 Compiler for C supports arguments -mavx512f: YES (cached) 00:04:53.327 Compiler for C supports arguments -mavx512bw: YES 00:04:53.327 Compiler for C supports arguments -mavx512dq: YES 00:04:53.327 Compiler for C supports arguments -mavx512vl: YES 00:04:53.327 Compiler for C supports arguments -mvpclmulqdq: YES 00:04:53.327 Compiler for C supports arguments -mavx2: YES 00:04:53.327 Compiler for C supports arguments -mavx: YES 00:04:53.327 Message: lib/net: Defining dependency "net" 00:04:53.327 Message: lib/meter: Defining dependency "meter" 00:04:53.327 Message: lib/ethdev: Defining dependency "ethdev" 00:04:53.327 Message: lib/pci: Defining dependency "pci" 00:04:53.327 Message: lib/cmdline: Defining dependency "cmdline" 00:04:53.327 Message: lib/metrics: Defining dependency "metrics" 00:04:53.327 Message: lib/hash: Defining dependency "hash" 00:04:53.327 Message: lib/timer: Defining dependency "timer" 00:04:53.327 Fetching value of define "__AVX512F__" : (undefined) (cached) 00:04:53.327 Fetching value of define "__AVX512VL__" : (undefined) (cached) 00:04:53.327 Fetching value of define "__AVX512CD__" : (undefined) (cached) 00:04:53.327 Fetching value of define "__AVX512BW__" : (undefined) (cached) 00:04:53.327 Compiler for C supports arguments -mavx512f -mavx512vl -mavx512cd -mavx512bw: YES 00:04:53.327 Message: lib/acl: Defining dependency "acl" 00:04:53.327 Message: lib/bbdev: Defining dependency "bbdev" 00:04:53.327 Message: lib/bitratestats: Defining dependency "bitratestats" 00:04:53.327 Run-time dependency libelf found: YES 0.191 00:04:53.327 Message: lib/bpf: Defining dependency "bpf" 00:04:53.327 Message: lib/cfgfile: Defining dependency "cfgfile" 00:04:53.327 Message: lib/compressdev: Defining dependency "compressdev" 00:04:53.327 Message: lib/cryptodev: Defining dependency "cryptodev" 00:04:53.327 Message: lib/distributor: Defining dependency "distributor" 00:04:53.327 Message: lib/dmadev: Defining dependency "dmadev" 00:04:53.327 Message: lib/efd: Defining dependency "efd" 00:04:53.327 Message: lib/eventdev: Defining dependency "eventdev" 00:04:53.327 Message: lib/dispatcher: Defining dependency "dispatcher" 00:04:53.327 Message: lib/gpudev: Defining dependency "gpudev" 00:04:53.327 Message: lib/gro: Defining dependency "gro" 00:04:53.327 Message: lib/gso: Defining dependency "gso" 00:04:53.327 Message: lib/ip_frag: Defining dependency "ip_frag" 00:04:53.327 Message: lib/jobstats: Defining dependency "jobstats" 00:04:53.327 Message: lib/latencystats: Defining dependency "latencystats" 00:04:53.327 Message: lib/lpm: Defining dependency "lpm" 00:04:53.327 Fetching value of define "__AVX512F__" : (undefined) (cached) 00:04:53.327 Fetching value of define "__AVX512DQ__" : (undefined) (cached) 00:04:53.327 Fetching value of define "__AVX512IFMA__" : (undefined) 00:04:53.327 Compiler for C supports arguments -mavx512f -mavx512dq -mavx512ifma: YES 00:04:53.327 Message: lib/member: Defining dependency "member" 00:04:53.327 Message: lib/pcapng: Defining dependency "pcapng" 00:04:53.327 Compiler for C supports arguments -Wno-cast-qual: YES 00:04:53.327 Message: lib/power: Defining dependency "power" 00:04:53.327 Message: lib/rawdev: Defining dependency "rawdev" 00:04:53.327 Message: lib/regexdev: Defining dependency "regexdev" 00:04:53.327 Message: lib/mldev: Defining dependency "mldev" 00:04:53.327 Message: lib/rib: Defining dependency "rib" 00:04:53.327 Message: lib/reorder: Defining dependency "reorder" 00:04:53.327 Message: lib/sched: Defining dependency "sched" 00:04:53.327 Message: lib/security: Defining dependency "security" 00:04:53.327 Message: lib/stack: Defining dependency "stack" 00:04:53.327 Has header "linux/userfaultfd.h" : YES 00:04:53.327 Has header "linux/vduse.h" : YES 00:04:53.327 Message: lib/vhost: Defining dependency "vhost" 00:04:53.327 Message: lib/ipsec: Defining dependency "ipsec" 00:04:53.327 Message: lib/pdcp: Defining dependency "pdcp" 00:04:53.327 Fetching value of define "__AVX512F__" : (undefined) (cached) 00:04:53.327 Fetching value of define "__AVX512DQ__" : (undefined) (cached) 00:04:53.327 Compiler for C supports arguments -mavx512f -mavx512dq: YES 00:04:53.327 Compiler for C supports arguments -mavx512bw: YES (cached) 00:04:53.327 Message: lib/fib: Defining dependency "fib" 00:04:53.327 Message: lib/port: Defining dependency "port" 00:04:53.327 Message: lib/pdump: Defining dependency "pdump" 00:04:53.327 Message: lib/table: Defining dependency "table" 00:04:53.327 Message: lib/pipeline: Defining dependency "pipeline" 00:04:53.327 Message: lib/graph: Defining dependency "graph" 00:04:53.327 Message: lib/node: Defining dependency "node" 00:04:55.238 Compiler for C supports arguments -Wno-format-truncation: YES (cached) 00:04:55.238 Message: drivers/bus/pci: Defining dependency "bus_pci" 00:04:55.238 Message: drivers/bus/vdev: Defining dependency "bus_vdev" 00:04:55.238 Message: drivers/mempool/ring: Defining dependency "mempool_ring" 00:04:55.238 Compiler for C supports arguments -Wno-sign-compare: YES 00:04:55.238 Compiler for C supports arguments -Wno-unused-value: YES 00:04:55.238 Compiler for C supports arguments -Wno-format: YES 00:04:55.238 Compiler for C supports arguments -Wno-format-security: YES 00:04:55.238 Compiler for C supports arguments -Wno-format-nonliteral: YES 00:04:55.238 Compiler for C supports arguments -Wno-strict-aliasing: YES 00:04:55.238 Compiler for C supports arguments -Wno-unused-but-set-variable: YES 00:04:55.238 Compiler for C supports arguments -Wno-unused-parameter: YES 00:04:55.238 Fetching value of define "__AVX512F__" : (undefined) (cached) 00:04:55.238 Compiler for C supports arguments -mavx512f: YES (cached) 00:04:55.238 Compiler for C supports arguments -mavx512bw: YES (cached) 00:04:55.238 Compiler for C supports arguments -march=skylake-avx512: YES 00:04:55.238 Message: drivers/net/i40e: Defining dependency "net_i40e" 00:04:55.238 Has header "sys/epoll.h" : YES 00:04:55.238 Program doxygen found: YES (/usr/local/bin/doxygen) 00:04:55.238 Configuring doxy-api-html.conf using configuration 00:04:55.238 Configuring doxy-api-man.conf using configuration 00:04:55.238 Program mandb found: YES (/usr/bin/mandb) 00:04:55.238 Program sphinx-build found: NO 00:04:55.238 Configuring rte_build_config.h using configuration 00:04:55.238 Message: 00:04:55.238 ================= 00:04:55.238 Applications Enabled 00:04:55.238 ================= 00:04:55.238 00:04:55.238 apps: 00:04:55.238 dumpcap, graph, pdump, proc-info, test-acl, test-bbdev, test-cmdline, test-compress-perf, 00:04:55.238 test-crypto-perf, test-dma-perf, test-eventdev, test-fib, test-flow-perf, test-gpudev, test-mldev, test-pipeline, 00:04:55.238 test-pmd, test-regex, test-sad, test-security-perf, 00:04:55.238 00:04:55.238 Message: 00:04:55.238 ================= 00:04:55.238 Libraries Enabled 00:04:55.238 ================= 00:04:55.238 00:04:55.238 libs: 00:04:55.238 log, kvargs, telemetry, eal, ring, rcu, mempool, mbuf, 00:04:55.238 net, meter, ethdev, pci, cmdline, metrics, hash, timer, 00:04:55.238 acl, bbdev, bitratestats, bpf, cfgfile, compressdev, cryptodev, distributor, 00:04:55.238 dmadev, efd, eventdev, dispatcher, gpudev, gro, gso, ip_frag, 00:04:55.238 jobstats, latencystats, lpm, member, pcapng, power, rawdev, regexdev, 00:04:55.238 mldev, rib, reorder, sched, security, stack, vhost, ipsec, 00:04:55.238 pdcp, fib, port, pdump, table, pipeline, graph, node, 00:04:55.238 00:04:55.238 00:04:55.238 Message: 00:04:55.238 =============== 00:04:55.238 Drivers Enabled 00:04:55.238 =============== 00:04:55.238 00:04:55.238 common: 00:04:55.238 00:04:55.238 bus: 00:04:55.238 pci, vdev, 00:04:55.238 mempool: 00:04:55.238 ring, 00:04:55.238 dma: 00:04:55.238 00:04:55.238 net: 00:04:55.238 i40e, 00:04:55.238 raw: 00:04:55.238 00:04:55.238 crypto: 00:04:55.238 00:04:55.238 compress: 00:04:55.238 00:04:55.238 regex: 00:04:55.238 00:04:55.238 ml: 00:04:55.238 00:04:55.238 vdpa: 00:04:55.238 00:04:55.238 event: 00:04:55.238 00:04:55.238 baseband: 00:04:55.238 00:04:55.238 gpu: 00:04:55.238 00:04:55.238 00:04:55.238 Message: 00:04:55.238 ================= 00:04:55.238 Content Skipped 00:04:55.238 ================= 00:04:55.238 00:04:55.238 apps: 00:04:55.238 00:04:55.238 libs: 00:04:55.238 00:04:55.238 drivers: 00:04:55.238 common/cpt: not in enabled drivers build config 00:04:55.238 common/dpaax: not in enabled drivers build config 00:04:55.238 common/iavf: not in enabled drivers build config 00:04:55.238 common/idpf: not in enabled drivers build config 00:04:55.238 common/mvep: not in enabled drivers build config 00:04:55.238 common/octeontx: not in enabled drivers build config 00:04:55.238 bus/auxiliary: not in enabled drivers build config 00:04:55.239 bus/cdx: not in enabled drivers build config 00:04:55.239 bus/dpaa: not in enabled drivers build config 00:04:55.239 bus/fslmc: not in enabled drivers build config 00:04:55.239 bus/ifpga: not in enabled drivers build config 00:04:55.239 bus/platform: not in enabled drivers build config 00:04:55.239 bus/vmbus: not in enabled drivers build config 00:04:55.239 common/cnxk: not in enabled drivers build config 00:04:55.239 common/mlx5: not in enabled drivers build config 00:04:55.239 common/nfp: not in enabled drivers build config 00:04:55.239 common/qat: not in enabled drivers build config 00:04:55.239 common/sfc_efx: not in enabled drivers build config 00:04:55.239 mempool/bucket: not in enabled drivers build config 00:04:55.239 mempool/cnxk: not in enabled drivers build config 00:04:55.239 mempool/dpaa: not in enabled drivers build config 00:04:55.239 mempool/dpaa2: not in enabled drivers build config 00:04:55.239 mempool/octeontx: not in enabled drivers build config 00:04:55.239 mempool/stack: not in enabled drivers build config 00:04:55.239 dma/cnxk: not in enabled drivers build config 00:04:55.239 dma/dpaa: not in enabled drivers build config 00:04:55.239 dma/dpaa2: not in enabled drivers build config 00:04:55.239 dma/hisilicon: not in enabled drivers build config 00:04:55.239 dma/idxd: not in enabled drivers build config 00:04:55.239 dma/ioat: not in enabled drivers build config 00:04:55.239 dma/skeleton: not in enabled drivers build config 00:04:55.239 net/af_packet: not in enabled drivers build config 00:04:55.239 net/af_xdp: not in enabled drivers build config 00:04:55.239 net/ark: not in enabled drivers build config 00:04:55.239 net/atlantic: not in enabled drivers build config 00:04:55.239 net/avp: not in enabled drivers build config 00:04:55.239 net/axgbe: not in enabled drivers build config 00:04:55.239 net/bnx2x: not in enabled drivers build config 00:04:55.239 net/bnxt: not in enabled drivers build config 00:04:55.239 net/bonding: not in enabled drivers build config 00:04:55.239 net/cnxk: not in enabled drivers build config 00:04:55.239 net/cpfl: not in enabled drivers build config 00:04:55.239 net/cxgbe: not in enabled drivers build config 00:04:55.239 net/dpaa: not in enabled drivers build config 00:04:55.239 net/dpaa2: not in enabled drivers build config 00:04:55.239 net/e1000: not in enabled drivers build config 00:04:55.239 net/ena: not in enabled drivers build config 00:04:55.239 net/enetc: not in enabled drivers build config 00:04:55.239 net/enetfec: not in enabled drivers build config 00:04:55.239 net/enic: not in enabled drivers build config 00:04:55.239 net/failsafe: not in enabled drivers build config 00:04:55.239 net/fm10k: not in enabled drivers build config 00:04:55.239 net/gve: not in enabled drivers build config 00:04:55.239 net/hinic: not in enabled drivers build config 00:04:55.239 net/hns3: not in enabled drivers build config 00:04:55.239 net/iavf: not in enabled drivers build config 00:04:55.239 net/ice: not in enabled drivers build config 00:04:55.239 net/idpf: not in enabled drivers build config 00:04:55.239 net/igc: not in enabled drivers build config 00:04:55.239 net/ionic: not in enabled drivers build config 00:04:55.239 net/ipn3ke: not in enabled drivers build config 00:04:55.239 net/ixgbe: not in enabled drivers build config 00:04:55.239 net/mana: not in enabled drivers build config 00:04:55.239 net/memif: not in enabled drivers build config 00:04:55.239 net/mlx4: not in enabled drivers build config 00:04:55.239 net/mlx5: not in enabled drivers build config 00:04:55.239 net/mvneta: not in enabled drivers build config 00:04:55.239 net/mvpp2: not in enabled drivers build config 00:04:55.239 net/netvsc: not in enabled drivers build config 00:04:55.239 net/nfb: not in enabled drivers build config 00:04:55.239 net/nfp: not in enabled drivers build config 00:04:55.239 net/ngbe: not in enabled drivers build config 00:04:55.239 net/null: not in enabled drivers build config 00:04:55.239 net/octeontx: not in enabled drivers build config 00:04:55.239 net/octeon_ep: not in enabled drivers build config 00:04:55.239 net/pcap: not in enabled drivers build config 00:04:55.239 net/pfe: not in enabled drivers build config 00:04:55.239 net/qede: not in enabled drivers build config 00:04:55.239 net/ring: not in enabled drivers build config 00:04:55.239 net/sfc: not in enabled drivers build config 00:04:55.239 net/softnic: not in enabled drivers build config 00:04:55.239 net/tap: not in enabled drivers build config 00:04:55.239 net/thunderx: not in enabled drivers build config 00:04:55.239 net/txgbe: not in enabled drivers build config 00:04:55.239 net/vdev_netvsc: not in enabled drivers build config 00:04:55.239 net/vhost: not in enabled drivers build config 00:04:55.239 net/virtio: not in enabled drivers build config 00:04:55.239 net/vmxnet3: not in enabled drivers build config 00:04:55.239 raw/cnxk_bphy: not in enabled drivers build config 00:04:55.239 raw/cnxk_gpio: not in enabled drivers build config 00:04:55.239 raw/dpaa2_cmdif: not in enabled drivers build config 00:04:55.239 raw/ifpga: not in enabled drivers build config 00:04:55.239 raw/ntb: not in enabled drivers build config 00:04:55.239 raw/skeleton: not in enabled drivers build config 00:04:55.239 crypto/armv8: not in enabled drivers build config 00:04:55.239 crypto/bcmfs: not in enabled drivers build config 00:04:55.239 crypto/caam_jr: not in enabled drivers build config 00:04:55.239 crypto/ccp: not in enabled drivers build config 00:04:55.239 crypto/cnxk: not in enabled drivers build config 00:04:55.239 crypto/dpaa_sec: not in enabled drivers build config 00:04:55.239 crypto/dpaa2_sec: not in enabled drivers build config 00:04:55.239 crypto/ipsec_mb: not in enabled drivers build config 00:04:55.239 crypto/mlx5: not in enabled drivers build config 00:04:55.239 crypto/mvsam: not in enabled drivers build config 00:04:55.239 crypto/nitrox: not in enabled drivers build config 00:04:55.239 crypto/null: not in enabled drivers build config 00:04:55.239 crypto/octeontx: not in enabled drivers build config 00:04:55.239 crypto/openssl: not in enabled drivers build config 00:04:55.239 crypto/scheduler: not in enabled drivers build config 00:04:55.239 crypto/uadk: not in enabled drivers build config 00:04:55.239 crypto/virtio: not in enabled drivers build config 00:04:55.239 compress/isal: not in enabled drivers build config 00:04:55.239 compress/mlx5: not in enabled drivers build config 00:04:55.239 compress/octeontx: not in enabled drivers build config 00:04:55.239 compress/zlib: not in enabled drivers build config 00:04:55.239 regex/mlx5: not in enabled drivers build config 00:04:55.239 regex/cn9k: not in enabled drivers build config 00:04:55.239 ml/cnxk: not in enabled drivers build config 00:04:55.239 vdpa/ifc: not in enabled drivers build config 00:04:55.239 vdpa/mlx5: not in enabled drivers build config 00:04:55.239 vdpa/nfp: not in enabled drivers build config 00:04:55.239 vdpa/sfc: not in enabled drivers build config 00:04:55.239 event/cnxk: not in enabled drivers build config 00:04:55.239 event/dlb2: not in enabled drivers build config 00:04:55.239 event/dpaa: not in enabled drivers build config 00:04:55.239 event/dpaa2: not in enabled drivers build config 00:04:55.239 event/dsw: not in enabled drivers build config 00:04:55.239 event/opdl: not in enabled drivers build config 00:04:55.239 event/skeleton: not in enabled drivers build config 00:04:55.239 event/sw: not in enabled drivers build config 00:04:55.239 event/octeontx: not in enabled drivers build config 00:04:55.239 baseband/acc: not in enabled drivers build config 00:04:55.239 baseband/fpga_5gnr_fec: not in enabled drivers build config 00:04:55.239 baseband/fpga_lte_fec: not in enabled drivers build config 00:04:55.239 baseband/la12xx: not in enabled drivers build config 00:04:55.239 baseband/null: not in enabled drivers build config 00:04:55.239 baseband/turbo_sw: not in enabled drivers build config 00:04:55.239 gpu/cuda: not in enabled drivers build config 00:04:55.239 00:04:55.239 00:04:55.239 Build targets in project: 220 00:04:55.239 00:04:55.239 DPDK 23.11.0 00:04:55.239 00:04:55.239 User defined options 00:04:55.239 libdir : lib 00:04:55.239 prefix : /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build 00:04:55.239 c_args : -fPIC -g -fcommon -Werror -Wno-stringop-overflow 00:04:55.239 c_link_args : 00:04:55.239 enable_docs : false 00:04:55.239 enable_drivers: bus,bus/pci,bus/vdev,mempool/ring,net/i40e,net/i40e/base,power/acpi,power/amd_pstate,power/cppc,power/intel_pstate,power/intel_uncore,power/kvm_vm, 00:04:55.239 enable_kmods : false 00:04:55.239 machine : native 00:04:55.239 tests : false 00:04:55.239 00:04:55.239 Found ninja-1.11.1.git.kitware.jobserver-1 at /usr/local/bin/ninja 00:04:55.239 WARNING: Running the setup command as `meson [options]` instead of `meson setup [options]` is ambiguous and deprecated. 00:04:55.239 17:20:36 build_native_dpdk -- common/autobuild_common.sh@199 -- $ ninja -C /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build-tmp -j48 00:04:55.239 ninja: Entering directory `/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build-tmp' 00:04:55.239 [1/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hypervisor.c.o 00:04:55.239 [2/710] Compiling C object lib/librte_eal.a.p/eal_common_rte_version.c.o 00:04:55.239 [3/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_errno.c.o 00:04:55.239 [4/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_cpuflags.c.o 00:04:55.239 [5/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_class.c.o 00:04:55.239 [6/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_debug.c.o 00:04:55.239 [7/710] Compiling C object lib/librte_eal.a.p/eal_common_rte_reciprocal.c.o 00:04:55.239 [8/710] Compiling C object lib/librte_log.a.p/log_log_linux.c.o 00:04:55.239 [9/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hexdump.c.o 00:04:55.239 [10/710] Compiling C object lib/librte_eal.a.p/eal_unix_eal_firmware.c.o 00:04:55.239 [11/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_string_fns.c.o 00:04:55.239 [12/710] Compiling C object lib/librte_eal.a.p/eal_unix_eal_debug.c.o 00:04:55.239 [13/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_uuid.c.o 00:04:55.239 [14/710] Compiling C object lib/librte_kvargs.a.p/kvargs_rte_kvargs.c.o 00:04:55.239 [15/710] Linking static target lib/librte_kvargs.a 00:04:55.508 [16/710] Compiling C object lib/librte_eal.a.p/eal_linux_eal_cpuflags.c.o 00:04:55.508 [17/710] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_data.c.o 00:04:55.508 [18/710] Compiling C object lib/librte_eal.a.p/eal_unix_rte_thread.c.o 00:04:55.508 [19/710] Compiling C object lib/librte_log.a.p/log_log.c.o 00:04:55.508 [20/710] Linking static target lib/librte_log.a 00:04:55.508 [21/710] Compiling C object lib/librte_eal.a.p/eal_linux_eal_thread.c.o 00:04:55.776 [22/710] Generating lib/kvargs.sym_chk with a custom command (wrapped by meson to capture output) 00:04:56.353 [23/710] Compiling C object lib/librte_eal.a.p/eal_unix_eal_file.c.o 00:04:56.353 [24/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_config.c.o 00:04:56.353 [25/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_timer.c.o 00:04:56.353 [26/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_tailqs.c.o 00:04:56.353 [27/710] Compiling C object lib/librte_eal.a.p/eal_common_rte_keepalive.c.o 00:04:56.353 [28/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_launch.c.o 00:04:56.353 [29/710] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_timer.c.o 00:04:56.353 [30/710] Compiling C object lib/librte_eal.a.p/eal_unix_eal_filesystem.c.o 00:04:56.353 [31/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_mcfg.c.o 00:04:56.353 [32/710] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cpuflags.c.o 00:04:56.353 [33/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_bus.c.o 00:04:56.353 [34/710] Compiling C object lib/librte_eal.a.p/eal_x86_rte_spinlock.c.o 00:04:56.353 [35/710] Compiling C object lib/librte_eal.a.p/eal_x86_rte_hypervisor.c.o 00:04:56.353 [36/710] Generating lib/log.sym_chk with a custom command (wrapped by meson to capture output) 00:04:56.353 [37/710] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_memory.c.o 00:04:56.353 [38/710] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio_mp_sync.c.o 00:04:56.353 [39/710] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_thread.c.o 00:04:56.353 [40/710] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_legacy.c.o 00:04:56.353 [41/710] Compiling C object lib/librte_eal.a.p/eal_common_rte_random.c.o 00:04:56.353 [42/710] Compiling C object lib/librte_eal.a.p/eal_linux_eal_lcore.c.o 00:04:56.353 [43/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_ctf.c.o 00:04:56.353 [44/710] Linking target lib/librte_log.so.24.0 00:04:56.353 [45/710] Compiling C object lib/librte_eal.a.p/eal_common_hotplug_mp.c.o 00:04:56.353 [46/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_interrupts.c.o 00:04:56.353 [47/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memalloc.c.o 00:04:56.353 [48/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_devargs.c.o 00:04:56.353 [49/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_points.c.o 00:04:56.622 [50/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace.c.o 00:04:56.622 [51/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_thread.c.o 00:04:56.622 [52/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_utils.c.o 00:04:56.622 [53/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dynmem.c.o 00:04:56.623 [54/710] Compiling C object lib/librte_eal.a.p/eal_common_malloc_mp.c.o 00:04:56.623 [55/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memzone.c.o 00:04:56.623 [56/710] Compiling C object lib/librte_eal.a.p/eal_common_malloc_elem.c.o 00:04:56.623 [57/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dev.c.o 00:04:56.623 [58/710] Compiling C object lib/librte_eal.a.p/eal_linux_eal_alarm.c.o 00:04:56.623 [59/710] Compiling C object lib/librte_eal.a.p/eal_linux_eal_dev.c.o 00:04:56.623 [60/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_lcore.c.o 00:04:56.623 [61/710] Compiling C object lib/librte_eal.a.p/eal_linux_eal_timer.c.o 00:04:56.623 [62/710] Compiling C object lib/librte_eal.a.p/eal_linux_eal_hugepage_info.c.o 00:04:56.925 [63/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_proc.c.o 00:04:56.925 [64/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_fbarray.c.o 00:04:56.925 [65/710] Compiling C object lib/librte_eal.a.p/eal_linux_eal.c.o 00:04:56.925 [66/710] Compiling C object lib/librte_eal.a.p/eal_common_malloc_heap.c.o 00:04:56.925 [67/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memory.c.o 00:04:57.187 [68/710] Compiling C object lib/librte_pci.a.p/pci_rte_pci.c.o 00:04:57.187 [69/710] Linking static target lib/librte_pci.a 00:04:57.187 [70/710] Compiling C object lib/librte_eal.a.p/eal_common_rte_service.c.o 00:04:57.187 [71/710] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline.c.o 00:04:57.187 [72/710] Compiling C object lib/librte_eal.a.p/eal_linux_eal_interrupts.c.o 00:04:57.454 [73/710] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_cirbuf.c.o 00:04:57.454 [74/710] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memalloc.c.o 00:04:57.454 [75/710] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memory.c.o 00:04:57.454 [76/710] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cycles.c.o 00:04:57.454 [77/710] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse.c.o 00:04:57.454 [78/710] Compiling C object lib/librte_eal.a.p/eal_x86_rte_power_intrinsics.c.o 00:04:57.454 [79/710] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_portlist.c.o 00:04:57.454 [80/710] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_num.c.o 00:04:57.454 [81/710] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_socket.c.o 00:04:57.454 [82/710] Generating lib/pci.sym_chk with a custom command (wrapped by meson to capture output) 00:04:57.454 [83/710] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_pool_ops.c.o 00:04:57.454 [84/710] Compiling C object lib/librte_ring.a.p/ring_rte_ring.c.o 00:04:57.454 [85/710] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_vt100.c.o 00:04:57.454 [86/710] Compiling C object lib/net/libnet_crc_avx512_lib.a.p/net_crc_avx512.c.o 00:04:57.454 [87/710] Linking static target lib/librte_ring.a 00:04:57.454 [88/710] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_string.c.o 00:04:57.454 [89/710] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops_default.c.o 00:04:57.718 [90/710] Linking static target lib/net/libnet_crc_avx512_lib.a 00:04:57.718 [91/710] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_os_unix.c.o 00:04:57.718 [92/710] Compiling C object lib/librte_net.a.p/net_rte_net_crc.c.o 00:04:57.718 [93/710] Compiling C object lib/librte_net.a.p/net_net_crc_sse.c.o 00:04:57.718 [94/710] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_ptype.c.o 00:04:57.718 [95/710] Compiling C object lib/librte_meter.a.p/meter_rte_meter.c.o 00:04:57.718 [96/710] Linking static target lib/librte_meter.a 00:04:57.718 [97/710] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops.c.o 00:04:57.718 [98/710] Compiling C object lib/librte_mempool.a.p/mempool_mempool_trace_points.c.o 00:04:57.718 [99/710] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_profile.c.o 00:04:57.718 [100/710] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry.c.o 00:04:57.718 [101/710] Compiling C object lib/librte_eal.a.p/eal_common_rte_malloc.c.o 00:04:57.718 [102/710] Linking static target lib/librte_telemetry.a 00:04:57.979 [103/710] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_options.c.o 00:04:57.979 [104/710] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_rdline.c.o 00:04:57.979 [105/710] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_class_eth.c.o 00:04:57.979 [106/710] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_dyn.c.o 00:04:57.979 [107/710] Compiling C object lib/librte_net.a.p/net_rte_ether.c.o 00:04:57.979 [108/710] Compiling C object lib/librte_net.a.p/net_rte_net.c.o 00:04:57.979 [109/710] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_telemetry.c.o 00:04:57.979 [110/710] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_private.c.o 00:04:57.979 [111/710] Generating lib/ring.sym_chk with a custom command (wrapped by meson to capture output) 00:04:57.979 [112/710] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio.c.o 00:04:57.979 [113/710] Linking static target lib/librte_eal.a 00:04:57.979 [114/710] Generating lib/meter.sym_chk with a custom command (wrapped by meson to capture output) 00:04:57.979 [115/710] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_driver.c.o 00:04:58.242 [116/710] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_cman.c.o 00:04:58.242 [117/710] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8472.c.o 00:04:58.242 [118/710] Compiling C object lib/librte_net.a.p/net_rte_arp.c.o 00:04:58.242 [119/710] Linking static target lib/librte_net.a 00:04:58.242 [120/710] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_common.c.o 00:04:58.242 [121/710] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_etheraddr.c.o 00:04:58.242 [122/710] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8079.c.o 00:04:58.507 [123/710] Generating lib/telemetry.sym_chk with a custom command (wrapped by meson to capture output) 00:04:58.507 [124/710] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_ipaddr.c.o 00:04:58.507 [125/710] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool.c.o 00:04:58.507 [126/710] Linking static target lib/librte_cmdline.a 00:04:58.507 [127/710] Linking static target lib/librte_mempool.a 00:04:58.507 [128/710] Compiling C object lib/librte_cfgfile.a.p/cfgfile_rte_cfgfile.c.o 00:04:58.791 [129/710] Linking static target lib/librte_cfgfile.a 00:04:58.791 [130/710] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8636.c.o 00:04:58.791 [131/710] Generating lib/net.sym_chk with a custom command (wrapped by meson to capture output) 00:04:58.791 [132/710] Compiling C object lib/librte_hash.a.p/hash_rte_fbk_hash.c.o 00:04:58.791 [133/710] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_mtr.c.o 00:04:58.791 [134/710] Compiling C object lib/librte_metrics.a.p/metrics_rte_metrics.c.o 00:04:58.791 [135/710] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_telemetry.c.o 00:04:58.791 [136/710] Compiling C object lib/librte_metrics.a.p/metrics_rte_metrics_telemetry.c.o 00:04:58.791 [137/710] Compiling C object lib/librte_acl.a.p/acl_tb_mem.c.o 00:04:58.791 [138/710] Linking static target lib/librte_metrics.a 00:04:59.054 [139/710] Compiling C object lib/librte_acl.a.p/acl_rte_acl.c.o 00:04:59.054 [140/710] Compiling C object lib/librte_rcu.a.p/rcu_rte_rcu_qsbr.c.o 00:04:59.054 [141/710] Linking static target lib/librte_rcu.a 00:04:59.054 [142/710] Compiling C object lib/librte_bpf.a.p/bpf_bpf.c.o 00:04:59.054 [143/710] Compiling C object lib/librte_bpf.a.p/bpf_bpf_stub.c.o 00:04:59.054 [144/710] Compiling C object lib/librte_bpf.a.p/bpf_bpf_load.c.o 00:04:59.054 [145/710] Compiling C object lib/librte_bitratestats.a.p/bitratestats_rte_bitrate.c.o 00:04:59.054 [146/710] Linking static target lib/librte_bitratestats.a 00:04:59.054 [147/710] Compiling C object lib/librte_bpf.a.p/bpf_bpf_dump.c.o 00:04:59.322 [148/710] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_tm.c.o 00:04:59.322 [149/710] Generating symbol file lib/librte_log.so.24.0.p/librte_log.so.24.0.symbols 00:04:59.322 [150/710] Generating lib/cfgfile.sym_chk with a custom command (wrapped by meson to capture output) 00:04:59.322 [151/710] Linking target lib/librte_telemetry.so.24.0 00:04:59.322 [152/710] Linking target lib/librte_kvargs.so.24.0 00:04:59.322 [153/710] Compiling C object lib/librte_acl.a.p/acl_acl_gen.c.o 00:04:59.322 [154/710] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev_pmd.c.o 00:04:59.322 [155/710] Compiling C object lib/librte_timer.a.p/timer_rte_timer.c.o 00:04:59.322 [156/710] Compiling C object lib/librte_bpf.a.p/bpf_bpf_load_elf.c.o 00:04:59.322 [157/710] Linking static target lib/librte_timer.a 00:04:59.322 [158/710] Compiling C object lib/librte_acl.a.p/acl_acl_run_scalar.c.o 00:04:59.322 [159/710] Generating lib/metrics.sym_chk with a custom command (wrapped by meson to capture output) 00:04:59.593 [160/710] Generating lib/rcu.sym_chk with a custom command (wrapped by meson to capture output) 00:04:59.593 [161/710] Generating lib/bitratestats.sym_chk with a custom command (wrapped by meson to capture output) 00:04:59.593 [162/710] Generating symbol file lib/librte_kvargs.so.24.0.p/librte_kvargs.so.24.0.symbols 00:04:59.593 [163/710] Generating symbol file lib/librte_telemetry.so.24.0.p/librte_telemetry.so.24.0.symbols 00:04:59.593 [164/710] Generating lib/mempool.sym_chk with a custom command (wrapped by meson to capture output) 00:04:59.593 [165/710] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_pmd.c.o 00:04:59.593 [166/710] Compiling C object lib/librte_bpf.a.p/bpf_bpf_convert.c.o 00:04:59.593 [167/710] Compiling C object lib/librte_bpf.a.p/bpf_bpf_exec.c.o 00:04:59.593 [168/710] Compiling C object lib/librte_bbdev.a.p/bbdev_rte_bbdev.c.o 00:04:59.593 [169/710] Linking static target lib/librte_bbdev.a 00:04:59.593 [170/710] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev.c.o 00:04:59.858 [171/710] Generating lib/cmdline.sym_chk with a custom command (wrapped by meson to capture output) 00:04:59.858 [172/710] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_trace_points.c.o 00:04:59.858 [173/710] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor_match_sse.c.o 00:04:59.858 [174/710] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev_trace_points.c.o 00:05:00.126 [175/710] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_comp.c.o 00:05:00.126 [176/710] Linking static target lib/librte_compressdev.a 00:05:00.126 [177/710] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_trace_points.c.o 00:05:00.126 [178/710] Generating lib/timer.sym_chk with a custom command (wrapped by meson to capture output) 00:05:00.126 [179/710] Compiling C object lib/librte_eventdev.a.p/eventdev_eventdev_private.c.o 00:05:00.126 [180/710] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor_single.c.o 00:05:00.126 [181/710] Compiling C object lib/librte_bpf.a.p/bpf_bpf_pkt.c.o 00:05:00.391 [182/710] Compiling C object lib/librte_bpf.a.p/bpf_bpf_validate.c.o 00:05:00.391 [183/710] Compiling C object lib/librte_hash.a.p/hash_rte_thash.c.o 00:05:00.660 [184/710] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_ring.c.o 00:05:00.660 [185/710] Compiling C object lib/librte_distributor.a.p/distributor_rte_distributor.c.o 00:05:00.660 [186/710] Linking static target lib/librte_distributor.a 00:05:00.660 [187/710] Generating lib/bbdev.sym_chk with a custom command (wrapped by meson to capture output) 00:05:00.660 [188/710] Compiling C object lib/librte_bpf.a.p/bpf_bpf_jit_x86.c.o 00:05:00.660 [189/710] Linking static target lib/librte_bpf.a 00:05:00.660 [190/710] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev.c.o 00:05:00.660 [191/710] Linking static target lib/librte_dmadev.a 00:05:00.660 [192/710] Compiling C object lib/librte_eventdev.a.p/eventdev_eventdev_trace_points.c.o 00:05:00.927 [193/710] Generating lib/compressdev.sym_chk with a custom command (wrapped by meson to capture output) 00:05:00.927 [194/710] Compiling C object lib/librte_gso.a.p/gso_gso_udp4.c.o 00:05:00.927 [195/710] Compiling C object lib/librte_gso.a.p/gso_gso_tcp4.c.o 00:05:00.927 [196/710] Compiling C object lib/librte_gso.a.p/gso_gso_tunnel_udp4.c.o 00:05:00.927 [197/710] Compiling C object lib/librte_gso.a.p/gso_gso_tunnel_tcp4.c.o 00:05:00.927 [198/710] Compiling C object lib/librte_gro.a.p/gro_gro_tcp6.c.o 00:05:00.927 [199/710] Compiling C object lib/librte_dispatcher.a.p/dispatcher_rte_dispatcher.c.o 00:05:01.192 [200/710] Generating lib/distributor.sym_chk with a custom command (wrapped by meson to capture output) 00:05:01.192 [201/710] Linking static target lib/librte_dispatcher.a 00:05:01.192 [202/710] Compiling C object lib/librte_gro.a.p/gro_gro_tcp4.c.o 00:05:01.192 [203/710] Compiling C object lib/librte_power.a.p/power_guest_channel.c.o 00:05:01.192 [204/710] Compiling C object lib/librte_gro.a.p/gro_gro_udp4.c.o 00:05:01.192 [205/710] Compiling C object lib/librte_gso.a.p/gso_rte_gso.c.o 00:05:01.192 [206/710] Compiling C object lib/librte_gro.a.p/gro_rte_gro.c.o 00:05:01.192 [207/710] Compiling C object lib/librte_gro.a.p/gro_gro_vxlan_tcp4.c.o 00:05:01.192 [208/710] Compiling C object lib/librte_power.a.p/power_power_common.c.o 00:05:01.192 [209/710] Generating lib/bpf.sym_chk with a custom command (wrapped by meson to capture output) 00:05:01.192 [210/710] Compiling C object lib/librte_gro.a.p/gro_gro_vxlan_udp4.c.o 00:05:01.192 [211/710] Compiling C object lib/librte_gpudev.a.p/gpudev_gpudev.c.o 00:05:01.192 [212/710] Linking static target lib/librte_gro.a 00:05:01.192 [213/710] Linking static target lib/librte_gpudev.a 00:05:01.192 [214/710] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv4_reassembly.c.o 00:05:01.192 [215/710] Compiling C object lib/librte_power.a.p/power_power_kvm_vm.c.o 00:05:01.457 [216/710] Compiling C object lib/librte_acl.a.p/acl_acl_bld.c.o 00:05:01.457 [217/710] Compiling C object lib/librte_jobstats.a.p/jobstats_rte_jobstats.c.o 00:05:01.457 [218/710] Generating lib/dmadev.sym_chk with a custom command (wrapped by meson to capture output) 00:05:01.457 [219/710] Linking static target lib/librte_jobstats.a 00:05:01.724 [220/710] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv6_reassembly.c.o 00:05:01.724 [221/710] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_dma_adapter.c.o 00:05:01.724 [222/710] Generating lib/gro.sym_chk with a custom command (wrapped by meson to capture output) 00:05:01.724 [223/710] Generating lib/dispatcher.sym_chk with a custom command (wrapped by meson to capture output) 00:05:01.724 [224/710] Compiling C object lib/librte_acl.a.p/acl_acl_run_sse.c.o 00:05:01.724 [225/710] Compiling C object lib/librte_latencystats.a.p/latencystats_rte_latencystats.c.o 00:05:01.724 [226/710] Linking static target lib/librte_latencystats.a 00:05:01.994 [227/710] Compiling C object lib/librte_ip_frag.a.p/ip_frag_ip_frag_internal.c.o 00:05:01.994 [228/710] Generating lib/jobstats.sym_chk with a custom command (wrapped by meson to capture output) 00:05:01.994 [229/710] Compiling C object lib/librte_member.a.p/member_rte_member.c.o 00:05:01.994 [230/710] Compiling C object lib/member/libsketch_avx512_tmp.a.p/rte_member_sketch_avx512.c.o 00:05:01.994 [231/710] Linking static target lib/member/libsketch_avx512_tmp.a 00:05:01.994 [232/710] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv6_fragmentation.c.o 00:05:01.994 [233/710] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ip_frag_common.c.o 00:05:01.994 [234/710] Compiling C object lib/librte_ip_frag.a.p/ip_frag_rte_ipv4_fragmentation.c.o 00:05:02.260 [235/710] Compiling C object lib/librte_lpm.a.p/lpm_rte_lpm.c.o 00:05:02.260 [236/710] Linking static target lib/librte_ip_frag.a 00:05:02.260 [237/710] Generating lib/latencystats.sym_chk with a custom command (wrapped by meson to capture output) 00:05:02.260 [238/710] Compiling C object lib/librte_sched.a.p/sched_rte_approx.c.o 00:05:02.260 [239/710] Compiling C object lib/librte_power.a.p/power_rte_power_uncore.c.o 00:05:02.260 [240/710] Compiling C object lib/librte_power.a.p/power_rte_power.c.o 00:05:02.527 [241/710] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_crypto_adapter.c.o 00:05:02.527 [242/710] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_eth_tx_adapter.c.o 00:05:02.527 [243/710] Compiling C object lib/librte_vhost.a.p/vhost_fd_man.c.o 00:05:02.527 [244/710] Compiling C object lib/librte_mldev.a.p/mldev_rte_mldev_pmd.c.o 00:05:02.528 [245/710] Compiling C object lib/librte_mldev.a.p/mldev_mldev_utils.c.o 00:05:02.528 [246/710] Generating lib/ip_frag.sym_chk with a custom command (wrapped by meson to capture output) 00:05:02.792 [247/710] Generating lib/gpudev.sym_chk with a custom command (wrapped by meson to capture output) 00:05:02.792 [248/710] Compiling C object lib/librte_gso.a.p/gso_gso_common.c.o 00:05:02.792 [249/710] Linking static target lib/librte_gso.a 00:05:02.792 [250/710] Compiling C object lib/librte_regexdev.a.p/regexdev_rte_regexdev.c.o 00:05:02.792 [251/710] Linking static target lib/librte_regexdev.a 00:05:02.792 [252/710] Compiling C object lib/librte_power.a.p/power_rte_power_pmd_mgmt.c.o 00:05:02.792 [253/710] Compiling C object lib/librte_member.a.p/member_rte_member_vbf.c.o 00:05:03.060 [254/710] Compiling C object lib/librte_rawdev.a.p/rawdev_rte_rawdev.c.o 00:05:03.060 [255/710] Linking static target lib/librte_rawdev.a 00:05:03.060 [256/710] Compiling C object lib/librte_mldev.a.p/mldev_mldev_utils_scalar_bfloat16.c.o 00:05:03.060 [257/710] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_eventdev.c.o 00:05:03.060 [258/710] Compiling C object lib/librte_efd.a.p/efd_rte_efd.c.o 00:05:03.060 [259/710] Linking static target lib/librte_efd.a 00:05:03.060 [260/710] Compiling C object lib/librte_mldev.a.p/mldev_rte_mldev.c.o 00:05:03.060 [261/710] Generating lib/gso.sym_chk with a custom command (wrapped by meson to capture output) 00:05:03.060 [262/710] Compiling C object lib/librte_sched.a.p/sched_rte_pie.c.o 00:05:03.060 [263/710] Compiling C object lib/librte_sched.a.p/sched_rte_red.c.o 00:05:03.060 [264/710] Compiling C object lib/librte_mldev.a.p/mldev_mldev_utils_scalar.c.o 00:05:03.060 [265/710] Linking static target lib/librte_mldev.a 00:05:03.342 [266/710] Compiling C object lib/librte_pcapng.a.p/pcapng_rte_pcapng.c.o 00:05:03.342 [267/710] Linking static target lib/librte_pcapng.a 00:05:03.342 [268/710] Compiling C object lib/librte_stack.a.p/stack_rte_stack.c.o 00:05:03.342 [269/710] Compiling C object lib/acl/libavx2_tmp.a.p/acl_run_avx2.c.o 00:05:03.342 [270/710] Linking static target lib/acl/libavx2_tmp.a 00:05:03.342 [271/710] Compiling C object lib/librte_stack.a.p/stack_rte_stack_std.c.o 00:05:03.342 [272/710] Compiling C object lib/librte_stack.a.p/stack_rte_stack_lf.c.o 00:05:03.342 [273/710] Linking static target lib/librte_stack.a 00:05:03.342 [274/710] Compiling C object lib/librte_power.a.p/power_power_acpi_cpufreq.c.o 00:05:03.342 [275/710] Compiling C object lib/librte_member.a.p/member_rte_member_ht.c.o 00:05:03.342 [276/710] Compiling C object lib/librte_lpm.a.p/lpm_rte_lpm6.c.o 00:05:03.342 [277/710] Generating lib/efd.sym_chk with a custom command (wrapped by meson to capture output) 00:05:03.342 [278/710] Linking static target lib/librte_lpm.a 00:05:03.605 [279/710] Compiling C object lib/librte_power.a.p/power_power_intel_uncore.c.o 00:05:03.605 [280/710] Compiling C object lib/librte_power.a.p/power_power_amd_pstate_cpufreq.c.o 00:05:03.605 [281/710] Generating lib/pcapng.sym_chk with a custom command (wrapped by meson to capture output) 00:05:03.605 [282/710] Compiling C object lib/librte_hash.a.p/hash_rte_cuckoo_hash.c.o 00:05:03.605 [283/710] Linking static target lib/librte_hash.a 00:05:03.605 [284/710] Compiling C object lib/librte_rib.a.p/rib_rte_rib.c.o 00:05:03.605 [285/710] Generating lib/stack.sym_chk with a custom command (wrapped by meson to capture output) 00:05:03.605 [286/710] Generating lib/rawdev.sym_chk with a custom command (wrapped by meson to capture output) 00:05:03.605 [287/710] Compiling C object lib/librte_power.a.p/power_power_cppc_cpufreq.c.o 00:05:03.871 [288/710] Compiling C object lib/acl/libavx512_tmp.a.p/acl_run_avx512.c.o 00:05:03.871 [289/710] Linking static target lib/acl/libavx512_tmp.a 00:05:03.871 [290/710] Compiling C object lib/librte_reorder.a.p/reorder_rte_reorder.c.o 00:05:03.871 [291/710] Linking static target lib/librte_acl.a 00:05:03.871 [292/710] Linking static target lib/librte_reorder.a 00:05:03.871 [293/710] Compiling C object lib/librte_vhost.a.p/vhost_vdpa.c.o 00:05:03.871 [294/710] Compiling C object lib/librte_power.a.p/power_power_pstate_cpufreq.c.o 00:05:03.871 [295/710] Linking static target lib/librte_power.a 00:05:03.871 [296/710] Generating lib/regexdev.sym_chk with a custom command (wrapped by meson to capture output) 00:05:04.138 [297/710] Generating lib/lpm.sym_chk with a custom command (wrapped by meson to capture output) 00:05:04.138 [298/710] Compiling C object lib/librte_vhost.a.p/vhost_iotlb.c.o 00:05:04.138 [299/710] Compiling C object lib/librte_security.a.p/security_rte_security.c.o 00:05:04.138 [300/710] Linking static target lib/librte_security.a 00:05:04.138 [301/710] Compiling C object lib/librte_ipsec.a.p/ipsec_ses.c.o 00:05:04.402 [302/710] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net_ctrl.c.o 00:05:04.402 [303/710] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf.c.o 00:05:04.402 [304/710] Compiling C object lib/librte_vhost.a.p/vhost_vduse.c.o 00:05:04.402 [305/710] Linking static target lib/librte_mbuf.a 00:05:04.402 [306/710] Generating lib/acl.sym_chk with a custom command (wrapped by meson to capture output) 00:05:04.402 [307/710] Compiling C object lib/librte_vhost.a.p/vhost_socket.c.o 00:05:04.402 [308/710] Compiling C object lib/librte_ipsec.a.p/ipsec_ipsec_telemetry.c.o 00:05:04.402 [309/710] Compiling C object lib/librte_pdcp.a.p/pdcp_pdcp_crypto.c.o 00:05:04.402 [310/710] Generating lib/reorder.sym_chk with a custom command (wrapped by meson to capture output) 00:05:04.402 [311/710] Compiling C object lib/librte_pdcp.a.p/pdcp_pdcp_reorder.c.o 00:05:04.402 [312/710] Compiling C object lib/librte_pdcp.a.p/pdcp_pdcp_ctrl_pdu.c.o 00:05:04.402 [313/710] Compiling C object lib/librte_rib.a.p/rib_rte_rib6.c.o 00:05:04.402 [314/710] Generating lib/hash.sym_chk with a custom command (wrapped by meson to capture output) 00:05:04.402 [315/710] Linking static target lib/librte_rib.a 00:05:04.668 [316/710] Compiling C object lib/librte_pdcp.a.p/pdcp_pdcp_cnt.c.o 00:05:04.668 [317/710] Compiling C object lib/librte_ipsec.a.p/ipsec_sa.c.o 00:05:04.668 [318/710] Compiling C object lib/fib/libtrie_avx512_tmp.a.p/trie_avx512.c.o 00:05:04.668 [319/710] Linking static target lib/fib/libtrie_avx512_tmp.a 00:05:04.668 [320/710] Compiling C object lib/librte_table.a.p/table_rte_swx_keycmp.c.o 00:05:04.668 [321/710] Compiling C object lib/fib/libdir24_8_avx512_tmp.a.p/dir24_8_avx512.c.o 00:05:04.668 [322/710] Linking static target lib/fib/libdir24_8_avx512_tmp.a 00:05:04.668 [323/710] Generating lib/security.sym_chk with a custom command (wrapped by meson to capture output) 00:05:04.668 [324/710] Compiling C object lib/librte_fib.a.p/fib_rte_fib.c.o 00:05:04.929 [325/710] Compiling C object lib/librte_fib.a.p/fib_rte_fib6.c.o 00:05:04.929 [326/710] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_flow.c.o 00:05:04.929 [327/710] Generating lib/power.sym_chk with a custom command (wrapped by meson to capture output) 00:05:05.196 [328/710] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_eth_rx_adapter.c.o 00:05:05.196 [329/710] Generating lib/mbuf.sym_chk with a custom command (wrapped by meson to capture output) 00:05:05.196 [330/710] Generating lib/rib.sym_chk with a custom command (wrapped by meson to capture output) 00:05:05.196 [331/710] Generating lib/mldev.sym_chk with a custom command (wrapped by meson to capture output) 00:05:05.462 [332/710] Compiling C object lib/librte_pdcp.a.p/pdcp_rte_pdcp.c.o 00:05:05.462 [333/710] Compiling C object lib/librte_port.a.p/port_rte_port_sched.c.o 00:05:05.462 [334/710] Compiling C object lib/librte_member.a.p/member_rte_member_sketch.c.o 00:05:05.732 [335/710] Linking static target lib/librte_member.a 00:05:05.732 [336/710] Compiling C object lib/librte_ipsec.a.p/ipsec_ipsec_sad.c.o 00:05:05.732 [337/710] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev.c.o 00:05:05.732 [338/710] Compiling C object lib/librte_eventdev.a.p/eventdev_rte_event_timer_adapter.c.o 00:05:05.732 [339/710] Linking static target lib/librte_ethdev.a 00:05:05.732 [340/710] Linking static target lib/librte_eventdev.a 00:05:05.997 [341/710] Compiling C object lib/librte_vhost.a.p/vhost_vhost_user.c.o 00:05:05.997 [342/710] Compiling C object lib/librte_port.a.p/port_rte_port_fd.c.o 00:05:05.997 [343/710] Compiling C object lib/librte_port.a.p/port_rte_port_frag.c.o 00:05:05.997 [344/710] Compiling C object lib/librte_table.a.p/table_rte_swx_table_selector.c.o 00:05:05.997 [345/710] Compiling C object lib/librte_port.a.p/port_rte_port_ethdev.c.o 00:05:05.997 [346/710] Compiling C object lib/librte_cryptodev.a.p/cryptodev_rte_cryptodev.c.o 00:05:05.997 [347/710] Compiling C object lib/librte_table.a.p/table_rte_table_array.c.o 00:05:05.997 [348/710] Linking static target lib/librte_cryptodev.a 00:05:05.997 [349/710] Compiling C object lib/librte_table.a.p/table_rte_swx_table_wm.c.o 00:05:05.997 [350/710] Compiling C object lib/librte_table.a.p/table_rte_swx_table_learner.c.o 00:05:05.998 [351/710] Compiling C object lib/librte_fib.a.p/fib_trie.c.o 00:05:05.998 [352/710] Generating lib/member.sym_chk with a custom command (wrapped by meson to capture output) 00:05:05.998 [353/710] Compiling C object lib/librte_table.a.p/table_rte_table_hash_cuckoo.c.o 00:05:05.998 [354/710] Compiling C object lib/librte_sched.a.p/sched_rte_sched.c.o 00:05:05.998 [355/710] Linking static target lib/librte_sched.a 00:05:05.998 [356/710] Compiling C object lib/librte_table.a.p/table_rte_swx_table_em.c.o 00:05:06.267 [357/710] Compiling C object lib/librte_port.a.p/port_rte_port_ras.c.o 00:05:06.267 [358/710] Compiling C object lib/librte_fib.a.p/fib_dir24_8.c.o 00:05:06.267 [359/710] Linking static target lib/librte_fib.a 00:05:06.267 [360/710] Compiling C object lib/librte_table.a.p/table_rte_table_lpm.c.o 00:05:06.267 [361/710] Compiling C object lib/librte_table.a.p/table_rte_table_acl.c.o 00:05:06.538 [362/710] Compiling C object lib/librte_port.a.p/port_rte_swx_port_ethdev.c.o 00:05:06.538 [363/710] Compiling C object lib/librte_table.a.p/table_rte_table_stub.c.o 00:05:06.538 [364/710] Compiling C object lib/librte_port.a.p/port_rte_swx_port_fd.c.o 00:05:06.538 [365/710] Compiling C object lib/librte_port.a.p/port_rte_port_sym_crypto.c.o 00:05:06.538 [366/710] Compiling C object lib/librte_table.a.p/table_rte_table_lpm_ipv6.c.o 00:05:06.538 [367/710] Compiling C object lib/librte_vhost.a.p/vhost_vhost.c.o 00:05:06.801 [368/710] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_port_in_action.c.o 00:05:06.801 [369/710] Compiling C object lib/librte_port.a.p/port_rte_port_eventdev.c.o 00:05:06.801 [370/710] Generating lib/fib.sym_chk with a custom command (wrapped by meson to capture output) 00:05:06.801 [371/710] Generating lib/sched.sym_chk with a custom command (wrapped by meson to capture output) 00:05:06.801 [372/710] Compiling C object lib/librte_node.a.p/node_null.c.o 00:05:06.801 [373/710] Compiling C object lib/librte_ipsec.a.p/ipsec_esp_outb.c.o 00:05:07.069 [374/710] Compiling C object lib/librte_pdump.a.p/pdump_rte_pdump.c.o 00:05:07.069 [375/710] Linking static target lib/librte_pdump.a 00:05:07.069 [376/710] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key8.c.o 00:05:07.069 [377/710] Compiling C object lib/librte_port.a.p/port_rte_swx_port_ring.c.o 00:05:07.337 [378/710] Compiling C object lib/librte_graph.a.p/graph_graph_ops.c.o 00:05:07.337 [379/710] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_params.c.o 00:05:07.337 [380/710] Compiling C object lib/librte_graph.a.p/graph_graph_debug.c.o 00:05:07.337 [381/710] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key16.c.o 00:05:07.337 [382/710] Compiling C object lib/librte_graph.a.p/graph_rte_graph_worker.c.o 00:05:07.337 [383/710] Compiling C object lib/librte_graph.a.p/graph_node.c.o 00:05:07.337 [384/710] Compiling C object lib/librte_table.a.p/table_rte_table_hash_lru.c.o 00:05:07.337 [385/710] Compiling C object lib/librte_graph.a.p/graph_graph_populate.c.o 00:05:07.337 [386/710] Compiling C object lib/librte_table.a.p/table_rte_table_hash_ext.c.o 00:05:07.337 [387/710] Generating lib/pdump.sym_chk with a custom command (wrapped by meson to capture output) 00:05:07.337 [388/710] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev_params.c.o 00:05:07.337 [389/710] Compiling C object lib/librte_node.a.p/node_ethdev_ctrl.c.o 00:05:07.337 [390/710] Compiling C object lib/librte_graph.a.p/graph_graph_pcap.c.o 00:05:07.610 [391/710] Compiling C object lib/librte_graph.a.p/graph_graph.c.o 00:05:07.610 [392/710] Compiling C object lib/librte_table.a.p/table_rte_table_hash_key32.c.o 00:05:07.610 [393/710] Linking static target lib/librte_table.a 00:05:07.610 [394/710] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_pipeline.c.o 00:05:07.874 [395/710] Compiling C object lib/librte_ipsec.a.p/ipsec_esp_inb.c.o 00:05:07.874 [396/710] Linking static target lib/librte_ipsec.a 00:05:07.874 [397/710] Compiling C object lib/librte_node.a.p/node_log.c.o 00:05:07.874 [398/710] Generating lib/cryptodev.sym_chk with a custom command (wrapped by meson to capture output) 00:05:08.141 [399/710] Compiling C object lib/librte_node.a.p/node_pkt_drop.c.o 00:05:08.141 [400/710] Compiling C object lib/librte_node.a.p/node_kernel_tx.c.o 00:05:08.141 [401/710] Compiling C object lib/librte_port.a.p/port_rte_swx_port_source_sink.c.o 00:05:08.410 [402/710] Generating lib/ipsec.sym_chk with a custom command (wrapped by meson to capture output) 00:05:08.410 [403/710] Compiling C object lib/librte_graph.a.p/graph_graph_stats.c.o 00:05:08.410 [404/710] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common_uio.c.o 00:05:08.673 [405/710] Compiling C object lib/librte_node.a.p/node_ethdev_rx.c.o 00:05:08.673 [406/710] Compiling C object lib/librte_node.a.p/node_ethdev_tx.c.o 00:05:08.673 [407/710] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common.c.o 00:05:08.673 [408/710] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_diag.c.o 00:05:08.673 [409/710] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_uio.c.o 00:05:08.674 [410/710] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci.c.o 00:05:08.943 [411/710] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_ipsec.c.o 00:05:08.943 [412/710] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev.c.o 00:05:08.943 [413/710] Linking static target drivers/libtmp_rte_bus_vdev.a 00:05:08.943 [414/710] Generating lib/table.sym_chk with a custom command (wrapped by meson to capture output) 00:05:08.943 [415/710] Compiling C object lib/librte_port.a.p/port_rte_port_ring.c.o 00:05:08.943 [416/710] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_hmc.c.o 00:05:09.221 [417/710] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_vfio.c.o 00:05:09.221 [418/710] Linking static target drivers/libtmp_rte_bus_pci.a 00:05:09.221 [419/710] Generating lib/eal.sym_chk with a custom command (wrapped by meson to capture output) 00:05:09.221 [420/710] Compiling C object lib/librte_port.a.p/port_rte_port_source_sink.c.o 00:05:09.221 [421/710] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_dcb.c.o 00:05:09.221 [422/710] Linking static target lib/librte_port.a 00:05:09.221 [423/710] Linking target lib/librte_eal.so.24.0 00:05:09.491 [424/710] Generating lib/eventdev.sym_chk with a custom command (wrapped by meson to capture output) 00:05:09.491 [425/710] Generating drivers/rte_bus_vdev.pmd.c with a custom command 00:05:09.491 [426/710] Compiling C object drivers/librte_bus_vdev.a.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:05:09.491 [427/710] Linking static target drivers/librte_bus_vdev.a 00:05:09.491 [428/710] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_ctl.c.o 00:05:09.491 [429/710] Compiling C object drivers/librte_bus_vdev.so.24.0.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:05:09.491 [430/710] Generating symbol file lib/librte_eal.so.24.0.p/librte_eal.so.24.0.symbols 00:05:09.491 [431/710] Compiling C object lib/librte_graph.a.p/graph_rte_graph_model_mcore_dispatch.c.o 00:05:09.757 [432/710] Linking target lib/librte_ring.so.24.0 00:05:09.757 [433/710] Generating drivers/rte_bus_pci.pmd.c with a custom command 00:05:09.757 [434/710] Linking target lib/librte_meter.so.24.0 00:05:09.757 [435/710] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_lan_hmc.c.o 00:05:09.757 [436/710] Linking target lib/librte_pci.so.24.0 00:05:09.757 [437/710] Compiling C object lib/librte_node.a.p/node_ip4_reassembly.c.o 00:05:09.757 [438/710] Linking target lib/librte_timer.so.24.0 00:05:09.757 [439/710] Compiling C object lib/librte_node.a.p/node_kernel_rx.c.o 00:05:09.757 [440/710] Linking target lib/librte_acl.so.24.0 00:05:09.757 [441/710] Linking target lib/librte_cfgfile.so.24.0 00:05:09.757 [442/710] Generating drivers/rte_bus_vdev.sym_chk with a custom command (wrapped by meson to capture output) 00:05:09.757 [443/710] Linking target lib/librte_dmadev.so.24.0 00:05:10.027 [444/710] Generating symbol file lib/librte_ring.so.24.0.p/librte_ring.so.24.0.symbols 00:05:10.027 [445/710] Generating symbol file lib/librte_meter.so.24.0.p/librte_meter.so.24.0.symbols 00:05:10.027 [446/710] Generating symbol file lib/librte_pci.so.24.0.p/librte_pci.so.24.0.symbols 00:05:10.027 [447/710] Linking target lib/librte_rcu.so.24.0 00:05:10.027 [448/710] Linking target lib/librte_jobstats.so.24.0 00:05:10.027 [449/710] Linking target lib/librte_mempool.so.24.0 00:05:10.027 [450/710] Compiling C object app/dpdk-graph.p/graph_cli.c.o 00:05:10.027 [451/710] Generating symbol file lib/librte_timer.so.24.0.p/librte_timer.so.24.0.symbols 00:05:10.027 [452/710] Linking static target lib/librte_graph.a 00:05:10.027 [453/710] Compiling C object drivers/librte_bus_pci.a.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:05:10.027 [454/710] Linking target lib/librte_rawdev.so.24.0 00:05:10.027 [455/710] Linking static target drivers/librte_bus_pci.a 00:05:10.027 [456/710] Linking target lib/librte_stack.so.24.0 00:05:10.027 [457/710] Generating symbol file lib/librte_acl.so.24.0.p/librte_acl.so.24.0.symbols 00:05:10.027 [458/710] Compiling C object drivers/libtmp_rte_mempool_ring.a.p/mempool_ring_rte_mempool_ring.c.o 00:05:10.027 [459/710] Compiling C object drivers/librte_bus_pci.so.24.0.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:05:10.027 [460/710] Linking static target drivers/libtmp_rte_mempool_ring.a 00:05:10.027 [461/710] Linking target drivers/librte_bus_vdev.so.24.0 00:05:10.027 [462/710] Compiling C object app/dpdk-graph.p/graph_conn.c.o 00:05:10.027 [463/710] Generating symbol file lib/librte_dmadev.so.24.0.p/librte_dmadev.so.24.0.symbols 00:05:10.291 [464/710] Compiling C object app/dpdk-graph.p/graph_ethdev_rx.c.o 00:05:10.291 [465/710] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_vf_representor.c.o 00:05:10.291 [466/710] Generating lib/port.sym_chk with a custom command (wrapped by meson to capture output) 00:05:10.291 [467/710] Generating symbol file lib/librte_rcu.so.24.0.p/librte_rcu.so.24.0.symbols 00:05:10.291 [468/710] Generating symbol file lib/librte_mempool.so.24.0.p/librte_mempool.so.24.0.symbols 00:05:10.291 [469/710] Linking target lib/librte_mbuf.so.24.0 00:05:10.291 [470/710] Generating symbol file drivers/librte_bus_vdev.so.24.0.p/librte_bus_vdev.so.24.0.symbols 00:05:10.559 [471/710] Linking target lib/librte_rib.so.24.0 00:05:10.559 [472/710] Generating drivers/rte_mempool_ring.pmd.c with a custom command 00:05:10.559 [473/710] Compiling C object drivers/librte_mempool_ring.a.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:05:10.559 [474/710] Linking static target drivers/librte_mempool_ring.a 00:05:10.559 [475/710] Compiling C object drivers/librte_mempool_ring.so.24.0.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:05:10.559 [476/710] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_hash.c.o 00:05:10.559 [477/710] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_tm.c.o 00:05:10.559 [478/710] Compiling C object app/dpdk-graph.p/graph_ip4_route.c.o 00:05:10.559 [479/710] Compiling C object app/dpdk-graph.p/graph_graph.c.o 00:05:10.559 [480/710] Linking target drivers/librte_mempool_ring.so.24.0 00:05:10.559 [481/710] Generating symbol file lib/librte_mbuf.so.24.0.p/librte_mbuf.so.24.0.symbols 00:05:10.559 [482/710] Compiling C object app/dpdk-graph.p/graph_ethdev.c.o 00:05:10.559 [483/710] Compiling C object app/dpdk-graph.p/graph_ip6_route.c.o 00:05:10.828 [484/710] Compiling C object app/dpdk-dumpcap.p/dumpcap_main.c.o 00:05:10.828 [485/710] Linking target lib/librte_net.so.24.0 00:05:10.828 [486/710] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_pf.c.o 00:05:10.828 [487/710] Linking target lib/librte_bbdev.so.24.0 00:05:10.828 [488/710] Compiling C object app/dpdk-test-cmdline.p/test-cmdline_cmdline_test.c.o 00:05:10.828 [489/710] Generating symbol file lib/librte_rib.so.24.0.p/librte_rib.so.24.0.symbols 00:05:10.828 [490/710] Compiling C object app/dpdk-test-cmdline.p/test-cmdline_commands.c.o 00:05:10.828 [491/710] Linking target lib/librte_compressdev.so.24.0 00:05:10.828 [492/710] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_recycle_mbufs_vec_common.c.o 00:05:10.828 [493/710] Linking target lib/librte_cryptodev.so.24.0 00:05:10.828 [494/710] Linking target lib/librte_distributor.so.24.0 00:05:10.828 [495/710] Linking target lib/librte_gpudev.so.24.0 00:05:10.828 [496/710] Linking target lib/librte_mldev.so.24.0 00:05:10.828 [497/710] Linking target lib/librte_regexdev.so.24.0 00:05:10.828 [498/710] Linking target lib/librte_reorder.so.24.0 00:05:10.828 [499/710] Linking target lib/librte_sched.so.24.0 00:05:10.828 [500/710] Generating drivers/rte_bus_pci.sym_chk with a custom command (wrapped by meson to capture output) 00:05:10.828 [501/710] Linking target lib/librte_fib.so.24.0 00:05:10.828 [502/710] Linking target drivers/librte_bus_pci.so.24.0 00:05:10.828 [503/710] Compiling C object app/dpdk-graph.p/graph_l3fwd.c.o 00:05:10.828 [504/710] Generating symbol file lib/librte_net.so.24.0.p/librte_net.so.24.0.symbols 00:05:10.828 [505/710] Compiling C object lib/librte_node.a.p/node_ip4_local.c.o 00:05:11.094 [506/710] Generating symbol file lib/librte_cryptodev.so.24.0.p/librte_cryptodev.so.24.0.symbols 00:05:11.094 [507/710] Generating lib/graph.sym_chk with a custom command (wrapped by meson to capture output) 00:05:11.094 [508/710] Linking target lib/librte_hash.so.24.0 00:05:11.094 [509/710] Linking target lib/librte_cmdline.so.24.0 00:05:11.094 [510/710] Generating symbol file lib/librte_reorder.so.24.0.p/librte_reorder.so.24.0.symbols 00:05:11.094 [511/710] Generating symbol file lib/librte_sched.so.24.0.p/librte_sched.so.24.0.symbols 00:05:11.094 [512/710] Linking target lib/librte_security.so.24.0 00:05:11.094 [513/710] Generating symbol file drivers/librte_bus_pci.so.24.0.p/librte_bus_pci.so.24.0.symbols 00:05:11.364 [514/710] Generating symbol file lib/librte_hash.so.24.0.p/librte_hash.so.24.0.symbols 00:05:11.364 [515/710] Compiling C object app/dpdk-graph.p/graph_mempool.c.o 00:05:11.364 [516/710] Generating symbol file lib/librte_security.so.24.0.p/librte_security.so.24.0.symbols 00:05:11.364 [517/710] Compiling C object app/dpdk-graph.p/graph_main.c.o 00:05:11.364 [518/710] Linking target lib/librte_lpm.so.24.0 00:05:11.364 [519/710] Linking target lib/librte_efd.so.24.0 00:05:11.364 [520/710] Linking target lib/librte_member.so.24.0 00:05:11.364 [521/710] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_pipeline_spec.c.o 00:05:11.364 [522/710] Linking target lib/librte_ipsec.so.24.0 00:05:11.364 [523/710] Compiling C object app/dpdk-graph.p/graph_utils.c.o 00:05:11.628 [524/710] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_main.c.o 00:05:11.628 [525/710] Generating symbol file lib/librte_lpm.so.24.0.p/librte_lpm.so.24.0.symbols 00:05:11.628 [526/710] Generating symbol file lib/librte_ipsec.so.24.0.p/librte_ipsec.so.24.0.symbols 00:05:11.628 [527/710] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_fdir.c.o 00:05:11.894 [528/710] Compiling C object drivers/net/i40e/libi40e_avx512_lib.a.p/i40e_rxtx_vec_avx512.c.o 00:05:11.894 [529/710] Linking static target drivers/net/i40e/libi40e_avx512_lib.a 00:05:11.894 [530/710] Compiling C object lib/librte_node.a.p/node_udp4_input.c.o 00:05:11.894 [531/710] Compiling C object app/dpdk-graph.p/graph_neigh.c.o 00:05:11.894 [532/710] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_test.c.o 00:05:12.160 [533/710] Compiling C object app/dpdk-test-acl.p/test-acl_main.c.o 00:05:12.160 [534/710] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_adminq.c.o 00:05:12.160 [535/710] Compiling C object drivers/net/i40e/libi40e_avx2_lib.a.p/i40e_rxtx_vec_avx2.c.o 00:05:12.160 [536/710] Linking static target drivers/net/i40e/libi40e_avx2_lib.a 00:05:12.160 [537/710] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_parser.c.o 00:05:12.425 [538/710] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_common.c.o 00:05:12.425 [539/710] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_options_parse.c.o 00:05:12.425 [540/710] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_main.c.o 00:05:12.425 [541/710] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_common.c.o 00:05:12.691 [542/710] Compiling C object app/dpdk-pdump.p/pdump_main.c.o 00:05:12.691 [543/710] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_flow.c.o 00:05:12.691 [544/710] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_vectors.c.o 00:05:12.691 [545/710] Compiling C object drivers/net/i40e/base/libi40e_base.a.p/i40e_nvm.c.o 00:05:12.691 [546/710] Compiling C object lib/librte_node.a.p/node_ip6_lookup.c.o 00:05:12.691 [547/710] Linking static target drivers/net/i40e/base/libi40e_base.a 00:05:12.954 [548/710] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_options_parsing.c.o 00:05:12.954 [549/710] Compiling C object app/dpdk-proc-info.p/proc-info_main.c.o 00:05:12.954 [550/710] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_ops.c.o 00:05:13.221 [551/710] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_vector_parsing.c.o 00:05:13.221 [552/710] Compiling C object app/dpdk-test-dma-perf.p/test-dma-perf_main.c.o 00:05:13.221 [553/710] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_main.c.o 00:05:13.221 [554/710] Compiling C object app/dpdk-test-mldev.p/test-mldev_ml_test.c.o 00:05:13.221 [555/710] Compiling C object lib/librte_node.a.p/node_pkt_cls.c.o 00:05:13.221 [556/710] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_main.c.o 00:05:13.486 [557/710] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev_vector.c.o 00:05:13.486 [558/710] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_evt_options.c.o 00:05:13.486 [559/710] Compiling C object app/dpdk-test-mldev.p/test-mldev_parser.c.o 00:05:13.754 [560/710] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_common.c.o 00:05:14.021 [561/710] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_rte_pmd_i40e.c.o 00:05:14.021 [562/710] Generating lib/ethdev.sym_chk with a custom command (wrapped by meson to capture output) 00:05:14.021 [563/710] Linking target lib/librte_ethdev.so.24.0 00:05:14.290 [564/710] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_flow_gen.c.o 00:05:14.290 [565/710] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_atq.c.o 00:05:14.290 [566/710] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_items_gen.c.o 00:05:14.290 [567/710] Compiling C object app/dpdk-test-mldev.p/test-mldev_ml_main.c.o 00:05:14.290 [568/710] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_order_queue.c.o 00:05:14.290 [569/710] Generating symbol file lib/librte_ethdev.so.24.0.p/librte_ethdev.so.24.0.symbols 00:05:14.556 [570/710] Compiling C object app/dpdk-test-gpudev.p/test-gpudev_main.c.o 00:05:14.566 [571/710] Linking target lib/librte_metrics.so.24.0 00:05:14.566 [572/710] Compiling C object app/dpdk-test-mldev.p/test-mldev_ml_options.c.o 00:05:14.566 [573/710] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_device_ops.c.o 00:05:14.566 [574/710] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_model_common.c.o 00:05:14.566 [575/710] Linking target lib/librte_bpf.so.24.0 00:05:14.566 [576/710] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_actions_gen.c.o 00:05:14.566 [577/710] Linking target lib/librte_eventdev.so.24.0 00:05:14.566 [578/710] Linking target lib/librte_gro.so.24.0 00:05:14.566 [579/710] Linking target lib/librte_gso.so.24.0 00:05:14.566 [580/710] Linking target lib/librte_ip_frag.so.24.0 00:05:14.566 [581/710] Linking target lib/librte_pcapng.so.24.0 00:05:14.566 [582/710] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_inference_ordered.c.o 00:05:14.566 [583/710] Linking target lib/librte_power.so.24.0 00:05:14.566 [584/710] Generating symbol file lib/librte_metrics.so.24.0.p/librte_metrics.so.24.0.symbols 00:05:14.833 [585/710] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_inference_interleave.c.o 00:05:14.833 [586/710] Linking target lib/librte_bitratestats.so.24.0 00:05:14.833 [587/710] Generating symbol file lib/librte_bpf.so.24.0.p/librte_bpf.so.24.0.symbols 00:05:14.833 [588/710] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_rxtx_vec_sse.c.o 00:05:14.833 [589/710] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_model_ops.c.o 00:05:14.833 [590/710] Generating symbol file lib/librte_eventdev.so.24.0.p/librte_eventdev.so.24.0.symbols 00:05:14.833 [591/710] Linking target lib/librte_latencystats.so.24.0 00:05:14.833 [592/710] Generating symbol file lib/librte_ip_frag.so.24.0.p/librte_ip_frag.so.24.0.symbols 00:05:14.833 [593/710] Compiling C object app/dpdk-test-dma-perf.p/test-dma-perf_benchmark.c.o 00:05:14.833 [594/710] Linking target lib/librte_dispatcher.so.24.0 00:05:14.834 [595/710] Generating symbol file lib/librte_pcapng.so.24.0.p/librte_pcapng.so.24.0.symbols 00:05:14.834 [596/710] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_atq.c.o 00:05:14.834 [597/710] Compiling C object lib/librte_node.a.p/node_ip4_lookup.c.o 00:05:14.834 [598/710] Linking target lib/librte_port.so.24.0 00:05:15.102 [599/710] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_stats.c.o 00:05:15.102 [600/710] Linking target lib/librte_pdump.so.24.0 00:05:15.102 [601/710] Linking target lib/librte_graph.so.24.0 00:05:15.102 [602/710] Compiling C object lib/librte_pdcp.a.p/pdcp_pdcp_process.c.o 00:05:15.102 [603/710] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_queue.c.o 00:05:15.102 [604/710] Linking static target lib/librte_pdcp.a 00:05:15.102 [605/710] Generating symbol file lib/librte_port.so.24.0.p/librte_port.so.24.0.symbols 00:05:15.102 [606/710] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_pipeline_common.c.o 00:05:15.367 [607/710] Compiling C object app/dpdk-test-fib.p/test-fib_main.c.o 00:05:15.367 [608/710] Generating symbol file lib/librte_graph.so.24.0.p/librte_graph.so.24.0.symbols 00:05:15.367 [609/710] Linking target lib/librte_table.so.24.0 00:05:15.367 [610/710] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_common.c.o 00:05:15.367 [611/710] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_config.c.o 00:05:15.633 [612/710] Generating symbol file lib/librte_table.so.24.0.p/librte_table.so.24.0.symbols 00:05:15.633 [613/710] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_init.c.o 00:05:15.633 [614/710] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_acl.c.o 00:05:15.633 [615/710] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_main.c.o 00:05:15.633 [616/710] Generating lib/pdcp.sym_chk with a custom command (wrapped by meson to capture output) 00:05:15.633 [617/710] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_lpm.c.o 00:05:15.900 [618/710] Linking target lib/librte_pdcp.so.24.0 00:05:15.900 [619/710] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev.c.o 00:05:15.900 [620/710] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_stub.c.o 00:05:15.900 [621/710] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_hash.c.o 00:05:15.900 [622/710] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_pmd_cyclecount.c.o 00:05:16.163 [623/710] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_pipeline_lpm_ipv6.c.o 00:05:16.163 [624/710] Compiling C object app/dpdk-test-flow-perf.p/test-flow-perf_main.c.o 00:05:16.163 [625/710] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_verify.c.o 00:05:16.163 [626/710] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_throughput.c.o 00:05:16.163 [627/710] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_cman.c.o 00:05:16.430 [628/710] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_rxtx.c.o 00:05:16.430 [629/710] Compiling C object app/dpdk-testpmd.p/test-pmd_cmd_flex_item.c.o 00:05:16.430 [630/710] Compiling C object app/dpdk-testpmd.p/test-pmd_5tswap.c.o 00:05:16.691 [631/710] Compiling C object app/dpdk-testpmd.p/test-pmd_flowgen.c.o 00:05:16.691 [632/710] Compiling C object app/dpdk-testpmd.p/test-pmd_iofwd.c.o 00:05:16.691 [633/710] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_throughput.c.o 00:05:16.691 [634/710] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_mtr.c.o 00:05:16.951 [635/710] Compiling C object app/dpdk-testpmd.p/test-pmd_macfwd.c.o 00:05:16.951 [636/710] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_tm.c.o 00:05:16.951 [637/710] Compiling C object app/dpdk-testpmd.p/test-pmd_rxonly.c.o 00:05:16.951 [638/710] Compiling C object app/dpdk-testpmd.p/test-pmd_macswap.c.o 00:05:16.951 [639/710] Compiling C object app/dpdk-testpmd.p/test-pmd_recycle_mbufs.c.o 00:05:17.211 [640/710] Compiling C object app/dpdk-testpmd.p/test-pmd_shared_rxq_fwd.c.o 00:05:17.211 [641/710] Compiling C object app/dpdk-test-pipeline.p/test-pipeline_runtime.c.o 00:05:17.211 [642/710] Compiling C object app/dpdk-testpmd.p/test-pmd_bpf_cmd.c.o 00:05:17.211 [643/710] Compiling C object app/dpdk-testpmd.p/test-pmd_ieee1588fwd.c.o 00:05:17.211 [644/710] Compiling C object app/dpdk-test-crypto-perf.p/test-crypto-perf_cperf_test_latency.c.o 00:05:17.470 [645/710] Compiling C object app/dpdk-testpmd.p/test-pmd_icmpecho.c.o 00:05:17.470 [646/710] Compiling C object app/dpdk-testpmd.p/test-pmd_parameters.c.o 00:05:17.470 [647/710] Compiling C object app/dpdk-test-sad.p/test-sad_main.c.o 00:05:17.729 [648/710] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_atq.c.o 00:05:17.729 [649/710] Compiling C object app/dpdk-test-security-perf.p/test-security-perf_test_security_perf.c.o 00:05:17.729 [650/710] Compiling C object app/dpdk-testpmd.p/test-pmd_util.c.o 00:05:17.729 [651/710] Compiling C object drivers/libtmp_rte_net_i40e.a.p/net_i40e_i40e_ethdev.c.o 00:05:17.729 [652/710] Linking static target drivers/libtmp_rte_net_i40e.a 00:05:17.729 [653/710] Compiling C object app/dpdk-testpmd.p/.._drivers_net_i40e_i40e_testpmd.c.o 00:05:17.989 [654/710] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_common.c.o 00:05:17.989 [655/710] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_queue.c.o 00:05:17.989 [656/710] Compiling C object app/dpdk-test-security-perf.p/test_test_cryptodev_security_ipsec.c.o 00:05:17.989 [657/710] Compiling C object app/dpdk-test-regex.p/test-regex_main.c.o 00:05:17.989 [658/710] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_cyclecount.c.o 00:05:17.989 [659/710] Compiling C object app/dpdk-testpmd.p/test-pmd_csumonly.c.o 00:05:18.248 [660/710] Generating drivers/rte_net_i40e.pmd.c with a custom command 00:05:18.248 [661/710] Compiling C object drivers/librte_net_i40e.so.24.0.p/meson-generated_.._rte_net_i40e.pmd.c.o 00:05:18.248 [662/710] Compiling C object drivers/librte_net_i40e.a.p/meson-generated_.._rte_net_i40e.pmd.c.o 00:05:18.248 [663/710] Linking static target drivers/librte_net_i40e.a 00:05:18.816 [664/710] Compiling C object app/dpdk-test-compress-perf.p/test-compress-perf_comp_perf_test_verify.c.o 00:05:18.816 [665/710] Generating drivers/rte_net_i40e.sym_chk with a custom command (wrapped by meson to capture output) 00:05:18.816 [666/710] Compiling C object app/dpdk-testpmd.p/test-pmd_testpmd.c.o 00:05:18.816 [667/710] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_swx_pipeline.c.o 00:05:18.816 [668/710] Linking target drivers/librte_net_i40e.so.24.0 00:05:19.074 [669/710] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline.c.o 00:05:19.332 [670/710] Compiling C object app/dpdk-testpmd.p/test-pmd_noisy_vnf.c.o 00:05:19.591 [671/710] Compiling C object lib/librte_node.a.p/node_ip4_rewrite.c.o 00:05:19.849 [672/710] Compiling C object lib/librte_node.a.p/node_ip6_rewrite.c.o 00:05:19.849 [673/710] Linking static target lib/librte_node.a 00:05:20.108 [674/710] Compiling C object app/dpdk-testpmd.p/test-pmd_txonly.c.o 00:05:20.108 [675/710] Generating lib/node.sym_chk with a custom command (wrapped by meson to capture output) 00:05:20.108 [676/710] Linking target lib/librte_node.so.24.0 00:05:21.044 [677/710] Compiling C object app/dpdk-test-eventdev.p/test-eventdev_test_perf_common.c.o 00:05:21.304 [678/710] Compiling C object app/dpdk-test-mldev.p/test-mldev_test_inference_common.c.o 00:05:21.304 [679/710] Compiling C object app/dpdk-testpmd.p/test-pmd_config.c.o 00:05:23.210 [680/710] Compiling C object app/dpdk-testpmd.p/test-pmd_cmdline_flow.c.o 00:05:23.787 [681/710] Compiling C object app/dpdk-test-bbdev.p/test-bbdev_test_bbdev_perf.c.o 00:05:29.058 [682/710] Compiling C object lib/librte_vhost.a.p/vhost_vhost_crypto.c.o 00:06:01.174 [683/710] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net.c.o 00:06:01.174 [684/710] Linking static target lib/librte_vhost.a 00:06:01.174 [685/710] Generating lib/vhost.sym_chk with a custom command (wrapped by meson to capture output) 00:06:01.174 [686/710] Linking target lib/librte_vhost.so.24.0 00:06:11.159 [687/710] Compiling C object lib/librte_pipeline.a.p/pipeline_rte_table_action.c.o 00:06:11.159 [688/710] Linking static target lib/librte_pipeline.a 00:06:11.730 [689/710] Linking target app/dpdk-test-acl 00:06:11.730 [690/710] Linking target app/dpdk-test-cmdline 00:06:11.730 [691/710] Linking target app/dpdk-proc-info 00:06:11.730 [692/710] Linking target app/dpdk-dumpcap 00:06:11.730 [693/710] Linking target app/dpdk-test-dma-perf 00:06:11.730 [694/710] Linking target app/dpdk-pdump 00:06:11.730 [695/710] Linking target app/dpdk-test-gpudev 00:06:11.730 [696/710] Linking target app/dpdk-test-regex 00:06:11.730 [697/710] Linking target app/dpdk-test-sad 00:06:11.730 [698/710] Linking target app/dpdk-test-fib 00:06:11.730 [699/710] Linking target app/dpdk-test-mldev 00:06:11.730 [700/710] Linking target app/dpdk-test-flow-perf 00:06:11.730 [701/710] Linking target app/dpdk-graph 00:06:11.730 [702/710] Linking target app/dpdk-test-security-perf 00:06:11.730 [703/710] Linking target app/dpdk-test-pipeline 00:06:11.730 [704/710] Linking target app/dpdk-test-crypto-perf 00:06:11.730 [705/710] Linking target app/dpdk-test-bbdev 00:06:11.730 [706/710] Linking target app/dpdk-test-compress-perf 00:06:11.730 [707/710] Linking target app/dpdk-test-eventdev 00:06:11.730 [708/710] Linking target app/dpdk-testpmd 00:06:14.271 [709/710] Generating lib/pipeline.sym_chk with a custom command (wrapped by meson to capture output) 00:06:14.271 [710/710] Linking target lib/librte_pipeline.so.24.0 00:06:14.271 17:21:55 build_native_dpdk -- common/autobuild_common.sh@201 -- $ uname -s 00:06:14.271 17:21:55 build_native_dpdk -- common/autobuild_common.sh@201 -- $ [[ Linux == \F\r\e\e\B\S\D ]] 00:06:14.271 17:21:55 build_native_dpdk -- common/autobuild_common.sh@214 -- $ ninja -C /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build-tmp -j48 install 00:06:14.271 ninja: Entering directory `/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build-tmp' 00:06:14.271 [0/1] Installing files. 00:06:14.271 Installing subdir /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples 00:06:14.271 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/l3fwd.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.271 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/l3fwd_em.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.271 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/em_default_v4.cfg to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.271 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/l3fwd_sse.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.271 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/l3fwd_em_hlm_neon.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.271 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/l3fwd_route.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/l3fwd_common.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/l3fwd_acl.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/l3fwd_event.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/l3fwd_lpm.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/l3fwd_acl.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/l3fwd_em_hlm_sse.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/lpm_route_parse.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/l3fwd_altivec.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/l3fwd_lpm_altivec.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/l3fwd_fib.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/l3fwd_event_generic.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/lpm_default_v6.cfg to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/l3fwd_event_internal_port.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/l3fwd_em_sequential.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/l3fwd_event.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/l3fwd_em_hlm.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/l3fwd_lpm.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/l3fwd_em.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/em_route_parse.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/lpm_default_v4.cfg to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/l3fwd_lpm_sse.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/l3fwd_neon.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/em_default_v6.cfg to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/l3fwd_acl_scalar.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd/l3fwd_lpm_neon.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vmdq_dcb/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vmdq_dcb 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vmdq_dcb/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vmdq_dcb 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/flow_filtering/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/flow_filtering 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/flow_filtering/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/flow_filtering 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/flow_filtering/flow_blocks.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/flow_filtering 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l2fwd-event/l2fwd_event_internal_port.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l2fwd-event 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l2fwd-event/l2fwd_event.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l2fwd-event 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l2fwd-event/l2fwd_event.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l2fwd-event 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l2fwd-event/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l2fwd-event 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l2fwd-event/l2fwd_event_generic.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l2fwd-event 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l2fwd-event/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l2fwd-event 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l2fwd-event/l2fwd_common.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l2fwd-event 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l2fwd-event/l2fwd_poll.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l2fwd-event 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l2fwd-event/l2fwd_common.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l2fwd-event 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l2fwd-event/l2fwd_poll.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l2fwd-event 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/cmdline/commands.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/cmdline 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/cmdline/parse_obj_list.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/cmdline 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/cmdline/commands.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/cmdline 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/cmdline/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/cmdline 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/cmdline/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/cmdline 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/cmdline/parse_obj_list.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/cmdline 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/common/pkt_group.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/common 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/common/neon/port_group.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/common/neon 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/common/altivec/port_group.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/common/altivec 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/common/sse/port_group.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/common/sse 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ptpclient/ptpclient.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ptpclient 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ptpclient/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ptpclient 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/helloworld/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/helloworld 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/helloworld/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/helloworld 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l2fwd/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l2fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l2fwd/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l2fwd 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/rxtx_callbacks/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/rxtx_callbacks 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/rxtx_callbacks/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/rxtx_callbacks 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vm_power_manager/oob_monitor_x86.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vm_power_manager 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vm_power_manager/vm_power_cli.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vm_power_manager 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vm_power_manager/channel_manager.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vm_power_manager 00:06:14.272 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vm_power_manager/channel_monitor.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vm_power_manager 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vm_power_manager/parse.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vm_power_manager 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vm_power_manager/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vm_power_manager 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vm_power_manager/channel_monitor.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vm_power_manager 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vm_power_manager/oob_monitor_nop.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vm_power_manager 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vm_power_manager/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vm_power_manager 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vm_power_manager/parse.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vm_power_manager 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vm_power_manager/channel_manager.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vm_power_manager 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vm_power_manager/vm_power_cli.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vm_power_manager 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vm_power_manager/power_manager.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vm_power_manager 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vm_power_manager/oob_monitor.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vm_power_manager 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vm_power_manager/power_manager.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vm_power_manager 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vm_power_manager/guest_cli/parse.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vm_power_manager/guest_cli/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vm_power_manager/guest_cli/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vm_power_manager/guest_cli/parse.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vm_power_manager/guest_cli/vm_power_cli_guest.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vm_power_manager/guest_cli/vm_power_cli_guest.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vm_power_manager/guest_cli 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/qos_sched/app_thread.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/qos_sched 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/qos_sched/profile_ov.cfg to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/qos_sched 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/qos_sched/args.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/qos_sched 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/qos_sched/cfg_file.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/qos_sched 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/qos_sched/init.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/qos_sched 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/qos_sched/cmdline.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/qos_sched 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/qos_sched/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/qos_sched 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/qos_sched/cfg_file.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/qos_sched 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/qos_sched/stats.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/qos_sched 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/qos_sched/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/qos_sched 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/qos_sched/main.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/qos_sched 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/qos_sched/profile_red.cfg to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/qos_sched 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/qos_sched/profile_pie.cfg to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/qos_sched 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/qos_sched/profile.cfg to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/qos_sched 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l2fwd-cat/l2fwd-cat.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l2fwd-cat 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l2fwd-cat/cat.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l2fwd-cat 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l2fwd-cat/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l2fwd-cat 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l2fwd-cat/cat.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l2fwd-cat 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd-power/perf_core.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd-power 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd-power/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd-power 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd-power/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd-power 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd-power/main.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd-power 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd-power/perf_core.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd-power 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vdpa/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vdpa 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vdpa/commands.list to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vdpa 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vdpa/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vdpa 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vdpa/vdpa_blk_compact.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vdpa 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vhost/virtio_net.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vhost 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vhost/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vhost 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vhost/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vhost 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vhost/main.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vhost 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vhost_blk/blk_spec.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vhost_blk 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vhost_blk/vhost_blk.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vhost_blk 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vhost_blk/vhost_blk_compat.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vhost_blk 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vhost_blk/vhost_blk.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vhost_blk 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vhost_blk/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vhost_blk 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vhost_blk/blk.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vhost_blk 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/fips_validation/fips_validation_ecdsa.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/fips_validation 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/fips_validation/fips_validation_aes.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/fips_validation 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/fips_validation/fips_validation_sha.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/fips_validation 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/fips_validation/fips_validation_tdes.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/fips_validation 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/fips_validation/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/fips_validation 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/fips_validation/fips_validation.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/fips_validation 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/fips_validation/fips_dev_self_test.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/fips_validation 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/fips_validation/fips_validation_rsa.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/fips_validation 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/fips_validation/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/fips_validation 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/fips_validation/fips_dev_self_test.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/fips_validation 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/fips_validation/fips_validation_gcm.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/fips_validation 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/fips_validation/fips_validation_cmac.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/fips_validation 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/fips_validation/fips_validation_xts.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/fips_validation 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/fips_validation/fips_validation_hmac.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/fips_validation 00:06:14.273 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/fips_validation/fips_validation_ccm.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/fips_validation 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/fips_validation/fips_validation.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/fips_validation 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/bond/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/bond 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/bond/commands.list to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/bond 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/bond/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/bond 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/dma/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/dma 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/dma/dmafwd.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/dma 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/multi_process/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/multi_process 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/multi_process/hotplug_mp/commands.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/multi_process/hotplug_mp/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/multi_process/hotplug_mp/commands.list to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/multi_process/hotplug_mp/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/multi_process/hotplug_mp 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/multi_process/simple_mp/mp_commands.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/multi_process/simple_mp/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/multi_process/simple_mp/commands.list to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/multi_process/simple_mp/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/multi_process/simple_mp/mp_commands.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/multi_process/simple_mp 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/multi_process/symmetric_mp/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/multi_process/symmetric_mp 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/multi_process/symmetric_mp/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/multi_process/symmetric_mp 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/multi_process/client_server_mp/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/multi_process/client_server_mp 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/multi_process/client_server_mp/mp_server/args.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/multi_process/client_server_mp/mp_server/args.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/multi_process/client_server_mp/mp_server/init.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/multi_process/client_server_mp/mp_server/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/multi_process/client_server_mp/mp_server/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/multi_process/client_server_mp/mp_server/init.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_server 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/multi_process/client_server_mp/mp_client/client.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_client 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/multi_process/client_server_mp/mp_client/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/mp_client 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/multi_process/client_server_mp/shared/common.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/multi_process/client_server_mp/shared 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd-graph/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd-graph 00:06:14.274 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l3fwd-graph/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l3fwd-graph 00:06:14.538 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l2fwd-jobstats/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l2fwd-jobstats 00:06:14.538 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l2fwd-jobstats/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l2fwd-jobstats 00:06:14.538 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_fragmentation/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_fragmentation 00:06:14.538 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_fragmentation/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_fragmentation 00:06:14.538 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/packet_ordering/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/packet_ordering 00:06:14.538 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/packet_ordering/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/packet_ordering 00:06:14.538 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l2fwd-crypto/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l2fwd-crypto 00:06:14.538 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l2fwd-crypto/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l2fwd-crypto 00:06:14.538 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l2fwd-keepalive/shm.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:06:14.538 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l2fwd-keepalive/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:06:14.538 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l2fwd-keepalive/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:06:14.538 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l2fwd-keepalive/shm.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l2fwd-keepalive 00:06:14.538 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l2fwd-keepalive/ka-agent/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l2fwd-keepalive/ka-agent 00:06:14.538 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l2fwd-keepalive/ka-agent/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l2fwd-keepalive/ka-agent 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/skeleton/basicfwd.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/skeleton 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/skeleton/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/skeleton 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l2fwd-macsec/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l2fwd-macsec 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/l2fwd-macsec/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/l2fwd-macsec 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/service_cores/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/service_cores 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/service_cores/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/service_cores 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/distributor/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/distributor 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/distributor/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/distributor 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/eventdev_pipeline/pipeline_worker_tx.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/eventdev_pipeline/pipeline_worker_generic.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/eventdev_pipeline/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/eventdev_pipeline/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/eventdev_pipeline/pipeline_common.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/eventdev_pipeline 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/esp.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/event_helper.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/ipsec_worker.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/ipsec.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/ipip.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/ep1.cfg to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/sp4.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/ipsec.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/flow.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/ipsec_process.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/flow.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/ipsec_neon.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/rt.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/ipsec_worker.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/ipsec_lpm_neon.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/sa.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/event_helper.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/sad.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/sad.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/esp.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/ipsec-secgw.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/sp6.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/parser.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/ep0.cfg to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/parser.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/ipsec-secgw.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/test/common_defs.sh to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/test/trs_3descbc_sha1_common_defs.sh to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/test/tun_3descbc_sha1_defs.sh to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/test/trs_aescbc_sha1_defs.sh to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/test/load_env.sh to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/test/tun_aesctr_sha1_common_defs.sh to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/test/tun_null_header_reconstruct.py to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/test/run_test.sh to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/test/tun_aesgcm_defs.sh to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/test/common_defs_secgw.sh to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/test/tun_3descbc_sha1_common_defs.sh to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/test/tun_aesctr_sha1_defs.sh to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/test/trs_aesctr_sha1_defs.sh to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/test/data_rxtx.sh to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/test/bypass_defs.sh to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/test/trs_ipv6opts.py to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/test/tun_aesgcm_common_defs.sh to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/test/trs_aesgcm_defs.sh to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/test/trs_aesctr_sha1_common_defs.sh to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/test/pkttest.sh to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/test/trs_aesgcm_common_defs.sh to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/test/trs_aescbc_sha1_common_defs.sh to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/test/pkttest.py to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/test/trs_3descbc_sha1_defs.sh to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/test/linux_test.sh to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/test/tun_aescbc_sha1_defs.sh to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipsec-secgw/test/tun_aescbc_sha1_common_defs.sh to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipsec-secgw/test 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipv4_multicast/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipv4_multicast 00:06:14.539 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ipv4_multicast/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ipv4_multicast 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/server_node_efd/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/server_node_efd 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/server_node_efd/efd_node/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/server_node_efd/efd_node 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/server_node_efd/efd_node/node.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/server_node_efd/efd_node 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/server_node_efd/efd_server/args.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/server_node_efd/efd_server 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/server_node_efd/efd_server/args.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/server_node_efd/efd_server 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/server_node_efd/efd_server/init.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/server_node_efd/efd_server 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/server_node_efd/efd_server/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/server_node_efd/efd_server 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/server_node_efd/efd_server/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/server_node_efd/efd_server 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/server_node_efd/efd_server/init.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/server_node_efd/efd_server 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/server_node_efd/shared/common.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/server_node_efd/shared 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/thread.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/cli.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/action.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/tap.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/tmgr.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/swq.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/thread.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/pipeline.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/swq.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/action.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/conn.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/tap.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/link.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/cryptodev.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/mempool.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/conn.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/parser.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/link.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/cryptodev.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/common.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/parser.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/cli.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/pipeline.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/mempool.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/tmgr.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/examples/route_ecmp.cli to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/examples/rss.cli to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/examples/flow.cli to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/examples/firewall.cli to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/examples/tap.cli to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/examples/l2fwd.cli to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/examples/route.cli to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_pipeline/examples/flow_crypto.cli to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_pipeline/examples 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/bpf/t1.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/bpf 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/bpf/t3.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/bpf 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/bpf/README to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/bpf 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/bpf/dummy.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/bpf 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/bpf/t2.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/bpf 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vmdq/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vmdq 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vmdq/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vmdq 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/link_status_interrupt/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/link_status_interrupt 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/link_status_interrupt/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/link_status_interrupt 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_reassembly/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_reassembly 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ip_reassembly/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ip_reassembly 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/bbdev_app/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/bbdev_app 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/bbdev_app/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/bbdev_app 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/thread.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/cli.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/thread.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/conn.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/obj.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/conn.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/cli.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/obj.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/rss.spec to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/hash_func.cli to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.540 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/l2fwd_macswp.spec to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/mirroring.cli to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/ipsec.io to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/selector.spec to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/ethdev.io to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/mirroring.spec to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/varbit.spec to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/fib.cli to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/l2fwd.spec to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/recirculation.cli to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/recirculation.spec to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/rss.cli to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/ipsec.cli to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/varbit.cli to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/vxlan_table.txt to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/fib_routing_table.txt to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/fib.spec to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/vxlan.spec to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/fib_nexthop_table.txt to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/learner.spec to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/l2fwd_macswp.cli to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/registers.spec to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/l2fwd_pcap.cli to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/fib_nexthop_group_table.txt to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/vxlan.cli to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/registers.cli to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/vxlan_pcap.cli to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/packet.txt to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/selector.txt to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/meter.spec to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/hash_func.spec to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/pcap.io to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/ipsec.spec to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/vxlan_table.py to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/ipsec_sa.txt to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/learner.cli to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/l2fwd.cli to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/selector.cli to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/meter.cli to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/pipeline/examples/l2fwd_macswp_pcap.cli to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/pipeline/examples 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/qos_meter/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/qos_meter 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/qos_meter/rte_policer.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/qos_meter 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/qos_meter/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/qos_meter 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/qos_meter/main.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/qos_meter 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/qos_meter/rte_policer.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/qos_meter 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ethtool/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ethtool 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ethtool/ethtool-app/ethapp.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ethtool/ethtool-app/ethapp.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ethtool/ethtool-app/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ethtool/ethtool-app/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ethtool/ethtool-app 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ethtool/lib/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ethtool/lib 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ethtool/lib/rte_ethtool.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ethtool/lib 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ethtool/lib/rte_ethtool.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ethtool/lib 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ntb/commands.list to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ntb 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ntb/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ntb 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/ntb/ntb_fwd.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/ntb 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vhost_crypto/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vhost_crypto 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/vhost_crypto/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/vhost_crypto 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/timer/main.c to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/timer 00:06:14.541 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/examples/timer/Makefile to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/share/dpdk/examples/timer 00:06:14.541 Installing lib/librte_log.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.541 Installing lib/librte_log.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.541 Installing lib/librte_kvargs.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.541 Installing lib/librte_kvargs.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.541 Installing lib/librte_telemetry.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.541 Installing lib/librte_telemetry.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.541 Installing lib/librte_eal.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.541 Installing lib/librte_eal.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.541 Installing lib/librte_ring.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.541 Installing lib/librte_ring.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.541 Installing lib/librte_rcu.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.541 Installing lib/librte_rcu.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.541 Installing lib/librte_mempool.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.541 Installing lib/librte_mempool.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.541 Installing lib/librte_mbuf.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_mbuf.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_net.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_net.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_meter.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_meter.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_ethdev.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_ethdev.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_pci.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_pci.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_cmdline.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_cmdline.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_metrics.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_metrics.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_hash.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_hash.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_timer.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_timer.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_acl.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_acl.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_bbdev.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_bbdev.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_bitratestats.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_bitratestats.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_bpf.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_bpf.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_cfgfile.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_cfgfile.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_compressdev.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_compressdev.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_cryptodev.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_cryptodev.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_distributor.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_distributor.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_dmadev.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_dmadev.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_efd.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_efd.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_eventdev.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_eventdev.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_dispatcher.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_dispatcher.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_gpudev.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_gpudev.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_gro.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_gro.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_gso.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_gso.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_ip_frag.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_ip_frag.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_jobstats.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_jobstats.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_latencystats.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_latencystats.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_lpm.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_lpm.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_member.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_member.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_pcapng.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_pcapng.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_power.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_power.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_rawdev.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_rawdev.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_regexdev.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_regexdev.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_mldev.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_mldev.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_rib.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_rib.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:14.542 Installing lib/librte_reorder.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing lib/librte_reorder.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing lib/librte_sched.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing lib/librte_sched.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing lib/librte_security.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing lib/librte_security.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing lib/librte_stack.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing lib/librte_stack.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing lib/librte_vhost.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing lib/librte_vhost.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing lib/librte_ipsec.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing lib/librte_ipsec.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing lib/librte_pdcp.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing lib/librte_pdcp.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing lib/librte_fib.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing lib/librte_fib.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing lib/librte_port.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing lib/librte_port.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing lib/librte_pdump.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing lib/librte_pdump.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing lib/librte_table.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing lib/librte_table.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing lib/librte_pipeline.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing lib/librte_pipeline.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing lib/librte_graph.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing lib/librte_graph.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing lib/librte_node.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing lib/librte_node.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing drivers/librte_bus_pci.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing drivers/librte_bus_pci.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/dpdk/pmds-24.0 00:06:15.120 Installing drivers/librte_bus_vdev.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing drivers/librte_bus_vdev.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/dpdk/pmds-24.0 00:06:15.120 Installing drivers/librte_mempool_ring.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing drivers/librte_mempool_ring.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/dpdk/pmds-24.0 00:06:15.120 Installing drivers/librte_net_i40e.a to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.120 Installing drivers/librte_net_i40e.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/dpdk/pmds-24.0 00:06:15.120 Installing app/dpdk-dumpcap to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/bin 00:06:15.120 Installing app/dpdk-graph to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/bin 00:06:15.120 Installing app/dpdk-pdump to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/bin 00:06:15.120 Installing app/dpdk-proc-info to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/bin 00:06:15.120 Installing app/dpdk-test-acl to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/bin 00:06:15.120 Installing app/dpdk-test-bbdev to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/bin 00:06:15.120 Installing app/dpdk-test-cmdline to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/bin 00:06:15.120 Installing app/dpdk-test-compress-perf to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/bin 00:06:15.120 Installing app/dpdk-test-crypto-perf to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/bin 00:06:15.120 Installing app/dpdk-test-dma-perf to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/bin 00:06:15.120 Installing app/dpdk-test-eventdev to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/bin 00:06:15.120 Installing app/dpdk-test-fib to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/bin 00:06:15.120 Installing app/dpdk-test-flow-perf to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/bin 00:06:15.120 Installing app/dpdk-test-gpudev to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/bin 00:06:15.120 Installing app/dpdk-test-mldev to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/bin 00:06:15.120 Installing app/dpdk-test-pipeline to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/bin 00:06:15.120 Installing app/dpdk-testpmd to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/bin 00:06:15.120 Installing app/dpdk-test-regex to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/bin 00:06:15.120 Installing app/dpdk-test-sad to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/bin 00:06:15.120 Installing app/dpdk-test-security-perf to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/bin 00:06:15.120 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/config/rte_config.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.120 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/log/rte_log.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.120 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/kvargs/rte_kvargs.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/telemetry/rte_telemetry.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/generic/rte_atomic.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include/generic 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/generic/rte_byteorder.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include/generic 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/generic/rte_cpuflags.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include/generic 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/generic/rte_cycles.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include/generic 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/generic/rte_io.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include/generic 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/generic/rte_memcpy.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include/generic 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/generic/rte_pause.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include/generic 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/generic/rte_power_intrinsics.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include/generic 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/generic/rte_prefetch.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include/generic 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/generic/rte_rwlock.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include/generic 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/generic/rte_spinlock.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include/generic 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/generic/rte_vect.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include/generic 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/x86/include/rte_atomic.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/x86/include/rte_byteorder.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/x86/include/rte_cpuflags.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/x86/include/rte_cycles.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/x86/include/rte_io.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/x86/include/rte_memcpy.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/x86/include/rte_pause.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/x86/include/rte_power_intrinsics.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/x86/include/rte_prefetch.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/x86/include/rte_rtm.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/x86/include/rte_rwlock.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/x86/include/rte_spinlock.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/x86/include/rte_vect.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/x86/include/rte_atomic_32.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/x86/include/rte_atomic_64.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/x86/include/rte_byteorder_32.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/x86/include/rte_byteorder_64.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_alarm.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_bitmap.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_bitops.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_branch_prediction.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_bus.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_class.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_common.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_compat.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_debug.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_dev.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_devargs.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_eal.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_eal_memconfig.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_eal_trace.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_errno.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_epoll.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_fbarray.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_hexdump.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_hypervisor.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_interrupts.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_keepalive.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_launch.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_lcore.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_lock_annotations.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_malloc.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_mcslock.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_memory.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_memzone.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_pci_dev_feature_defs.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_pci_dev_features.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_per_lcore.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_pflock.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_random.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_reciprocal.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_seqcount.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_seqlock.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_service.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_service_component.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_stdatomic.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_string_fns.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_tailq.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_thread.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_ticketlock.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_time.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_trace.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_trace_point.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_trace_point_register.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_uuid.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.121 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_version.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/include/rte_vfio.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eal/linux/include/rte_os.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ring/rte_ring.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ring/rte_ring_core.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ring/rte_ring_elem.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ring/rte_ring_elem_pvt.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ring/rte_ring_c11_pvt.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ring/rte_ring_generic_pvt.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ring/rte_ring_hts.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ring/rte_ring_hts_elem_pvt.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ring/rte_ring_peek.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ring/rte_ring_peek_elem_pvt.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ring/rte_ring_peek_zc.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ring/rte_ring_rts.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ring/rte_ring_rts_elem_pvt.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/rcu/rte_rcu_qsbr.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/mempool/rte_mempool.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/mempool/rte_mempool_trace_fp.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/mbuf/rte_mbuf.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/mbuf/rte_mbuf_core.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/mbuf/rte_mbuf_ptype.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/mbuf/rte_mbuf_pool_ops.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/mbuf/rte_mbuf_dyn.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/net/rte_ip.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/net/rte_tcp.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/net/rte_udp.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/net/rte_tls.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/net/rte_dtls.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/net/rte_esp.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/net/rte_sctp.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/net/rte_icmp.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/net/rte_arp.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/net/rte_ether.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/net/rte_macsec.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/net/rte_vxlan.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/net/rte_gre.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/net/rte_gtp.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/net/rte_net.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/net/rte_net_crc.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/net/rte_mpls.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/net/rte_higig.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/net/rte_ecpri.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/net/rte_pdcp_hdr.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/net/rte_geneve.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/net/rte_l2tpv2.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/net/rte_ppp.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/net/rte_ib.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/meter/rte_meter.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ethdev/rte_cman.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ethdev/rte_ethdev.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ethdev/rte_ethdev_trace_fp.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ethdev/rte_dev_info.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ethdev/rte_flow.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ethdev/rte_flow_driver.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ethdev/rte_mtr.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ethdev/rte_mtr_driver.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ethdev/rte_tm.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ethdev/rte_tm_driver.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ethdev/rte_ethdev_core.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ethdev/rte_eth_ctrl.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/pci/rte_pci.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/cmdline/cmdline.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/cmdline/cmdline_parse.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/cmdline/cmdline_parse_num.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/cmdline/cmdline_parse_ipaddr.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/cmdline/cmdline_parse_etheraddr.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/cmdline/cmdline_parse_string.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/cmdline/cmdline_rdline.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/cmdline/cmdline_vt100.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/cmdline/cmdline_socket.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/cmdline/cmdline_cirbuf.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/cmdline/cmdline_parse_portlist.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/metrics/rte_metrics.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/metrics/rte_metrics_telemetry.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/hash/rte_fbk_hash.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/hash/rte_hash_crc.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/hash/rte_hash.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/hash/rte_jhash.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/hash/rte_thash.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/hash/rte_thash_gfni.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/hash/rte_crc_arm64.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/hash/rte_crc_generic.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/hash/rte_crc_sw.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.122 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/hash/rte_crc_x86.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/hash/rte_thash_x86_gfni.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/timer/rte_timer.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/acl/rte_acl.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/acl/rte_acl_osdep.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/bbdev/rte_bbdev.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/bbdev/rte_bbdev_pmd.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/bbdev/rte_bbdev_op.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/bitratestats/rte_bitrate.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/bpf/bpf_def.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/bpf/rte_bpf.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/bpf/rte_bpf_ethdev.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/cfgfile/rte_cfgfile.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/compressdev/rte_compressdev.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/compressdev/rte_comp.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/cryptodev/rte_cryptodev.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/cryptodev/rte_cryptodev_trace_fp.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/cryptodev/rte_crypto.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/cryptodev/rte_crypto_sym.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/cryptodev/rte_crypto_asym.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/cryptodev/rte_cryptodev_core.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/distributor/rte_distributor.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/dmadev/rte_dmadev.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/dmadev/rte_dmadev_core.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/efd/rte_efd.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eventdev/rte_event_crypto_adapter.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eventdev/rte_event_dma_adapter.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eventdev/rte_event_eth_rx_adapter.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eventdev/rte_event_eth_tx_adapter.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eventdev/rte_event_ring.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eventdev/rte_event_timer_adapter.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eventdev/rte_eventdev.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eventdev/rte_eventdev_trace_fp.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/eventdev/rte_eventdev_core.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/dispatcher/rte_dispatcher.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/gpudev/rte_gpudev.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/gro/rte_gro.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/gso/rte_gso.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ip_frag/rte_ip_frag.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/jobstats/rte_jobstats.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/latencystats/rte_latencystats.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/lpm/rte_lpm.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/lpm/rte_lpm6.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/lpm/rte_lpm_altivec.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/lpm/rte_lpm_neon.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/lpm/rte_lpm_scalar.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/lpm/rte_lpm_sse.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/lpm/rte_lpm_sve.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/member/rte_member.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/pcapng/rte_pcapng.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/power/rte_power.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/power/rte_power_guest_channel.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/power/rte_power_pmd_mgmt.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/power/rte_power_uncore.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/rawdev/rte_rawdev.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/rawdev/rte_rawdev_pmd.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/regexdev/rte_regexdev.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/regexdev/rte_regexdev_driver.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/regexdev/rte_regexdev_core.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/mldev/rte_mldev.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/mldev/rte_mldev_core.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/rib/rte_rib.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/rib/rte_rib6.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/reorder/rte_reorder.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/sched/rte_approx.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/sched/rte_red.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/sched/rte_sched.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/sched/rte_sched_common.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/sched/rte_pie.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/security/rte_security.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/security/rte_security_driver.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/stack/rte_stack.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/stack/rte_stack_std.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/stack/rte_stack_lf.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/stack/rte_stack_lf_generic.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/stack/rte_stack_lf_c11.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/stack/rte_stack_lf_stubs.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/vhost/rte_vdpa.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/vhost/rte_vhost.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/vhost/rte_vhost_async.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/vhost/rte_vhost_crypto.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ipsec/rte_ipsec.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ipsec/rte_ipsec_sa.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.123 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ipsec/rte_ipsec_sad.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/ipsec/rte_ipsec_group.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/pdcp/rte_pdcp.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/pdcp/rte_pdcp_group.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/fib/rte_fib.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/fib/rte_fib6.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/port/rte_port_ethdev.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/port/rte_port_fd.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/port/rte_port_frag.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/port/rte_port_ras.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/port/rte_port.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/port/rte_port_ring.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/port/rte_port_sched.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/port/rte_port_source_sink.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/port/rte_port_sym_crypto.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/port/rte_port_eventdev.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/port/rte_swx_port.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/port/rte_swx_port_ethdev.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/port/rte_swx_port_fd.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/port/rte_swx_port_ring.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/port/rte_swx_port_source_sink.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/pdump/rte_pdump.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/table/rte_lru.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/table/rte_swx_hash_func.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/table/rte_swx_table.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/table/rte_swx_table_em.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/table/rte_swx_table_learner.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/table/rte_swx_table_selector.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/table/rte_swx_table_wm.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/table/rte_table.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/table/rte_table_acl.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/table/rte_table_array.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/table/rte_table_hash.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/table/rte_table_hash_cuckoo.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/table/rte_table_hash_func.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/table/rte_table_lpm.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/table/rte_table_lpm_ipv6.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/table/rte_table_stub.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/table/rte_lru_arm64.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/table/rte_lru_x86.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/table/rte_table_hash_func_arm64.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/pipeline/rte_pipeline.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/pipeline/rte_port_in_action.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/pipeline/rte_table_action.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/pipeline/rte_swx_ipsec.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/pipeline/rte_swx_pipeline.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/pipeline/rte_swx_extern.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/pipeline/rte_swx_ctl.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/graph/rte_graph.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/graph/rte_graph_worker.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/graph/rte_graph_model_mcore_dispatch.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/graph/rte_graph_model_rtc.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/graph/rte_graph_worker_common.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/node/rte_node_eth_api.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/node/rte_node_ip4_api.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/node/rte_node_ip6_api.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/lib/node/rte_node_udp4_input_api.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/drivers/bus/pci/rte_bus_pci.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/drivers/bus/vdev/rte_bus_vdev.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/drivers/net/i40e/rte_pmd_i40e.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/buildtools/dpdk-cmdline-gen.py to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/bin 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/usertools/dpdk-devbind.py to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/bin 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/usertools/dpdk-pmdinfo.py to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/bin 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/usertools/dpdk-telemetry.py to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/bin 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/usertools/dpdk-hugepages.py to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/bin 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/usertools/dpdk-rss-flows.py to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/bin 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build-tmp/rte_build_config.h to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build-tmp/meson-private/libdpdk-libs.pc to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/pkgconfig 00:06:15.124 Installing /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build-tmp/meson-private/libdpdk.pc to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/pkgconfig 00:06:15.124 Installing symlink pointing to librte_log.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_log.so.24 00:06:15.124 Installing symlink pointing to librte_log.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_log.so 00:06:15.124 Installing symlink pointing to librte_kvargs.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_kvargs.so.24 00:06:15.124 Installing symlink pointing to librte_kvargs.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_kvargs.so 00:06:15.125 Installing symlink pointing to librte_telemetry.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_telemetry.so.24 00:06:15.125 Installing symlink pointing to librte_telemetry.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_telemetry.so 00:06:15.125 Installing symlink pointing to librte_eal.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_eal.so.24 00:06:15.125 Installing symlink pointing to librte_eal.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_eal.so 00:06:15.125 Installing symlink pointing to librte_ring.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_ring.so.24 00:06:15.125 Installing symlink pointing to librte_ring.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_ring.so 00:06:15.125 Installing symlink pointing to librte_rcu.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_rcu.so.24 00:06:15.125 Installing symlink pointing to librte_rcu.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_rcu.so 00:06:15.125 Installing symlink pointing to librte_mempool.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_mempool.so.24 00:06:15.125 Installing symlink pointing to librte_mempool.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_mempool.so 00:06:15.125 Installing symlink pointing to librte_mbuf.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_mbuf.so.24 00:06:15.125 Installing symlink pointing to librte_mbuf.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_mbuf.so 00:06:15.125 Installing symlink pointing to librte_net.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_net.so.24 00:06:15.125 Installing symlink pointing to librte_net.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_net.so 00:06:15.125 Installing symlink pointing to librte_meter.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_meter.so.24 00:06:15.125 Installing symlink pointing to librte_meter.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_meter.so 00:06:15.125 Installing symlink pointing to librte_ethdev.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_ethdev.so.24 00:06:15.125 Installing symlink pointing to librte_ethdev.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_ethdev.so 00:06:15.125 Installing symlink pointing to librte_pci.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_pci.so.24 00:06:15.125 Installing symlink pointing to librte_pci.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_pci.so 00:06:15.125 Installing symlink pointing to librte_cmdline.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_cmdline.so.24 00:06:15.125 Installing symlink pointing to librte_cmdline.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_cmdline.so 00:06:15.125 Installing symlink pointing to librte_metrics.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_metrics.so.24 00:06:15.125 Installing symlink pointing to librte_metrics.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_metrics.so 00:06:15.125 Installing symlink pointing to librte_hash.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_hash.so.24 00:06:15.125 Installing symlink pointing to librte_hash.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_hash.so 00:06:15.125 Installing symlink pointing to librte_timer.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_timer.so.24 00:06:15.125 Installing symlink pointing to librte_timer.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_timer.so 00:06:15.125 Installing symlink pointing to librte_acl.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_acl.so.24 00:06:15.125 Installing symlink pointing to librte_acl.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_acl.so 00:06:15.125 Installing symlink pointing to librte_bbdev.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_bbdev.so.24 00:06:15.125 Installing symlink pointing to librte_bbdev.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_bbdev.so 00:06:15.125 Installing symlink pointing to librte_bitratestats.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_bitratestats.so.24 00:06:15.125 Installing symlink pointing to librte_bitratestats.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_bitratestats.so 00:06:15.125 Installing symlink pointing to librte_bpf.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_bpf.so.24 00:06:15.125 Installing symlink pointing to librte_bpf.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_bpf.so 00:06:15.125 Installing symlink pointing to librte_cfgfile.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_cfgfile.so.24 00:06:15.125 Installing symlink pointing to librte_cfgfile.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_cfgfile.so 00:06:15.125 Installing symlink pointing to librte_compressdev.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_compressdev.so.24 00:06:15.125 Installing symlink pointing to librte_compressdev.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_compressdev.so 00:06:15.125 Installing symlink pointing to librte_cryptodev.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_cryptodev.so.24 00:06:15.125 Installing symlink pointing to librte_cryptodev.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_cryptodev.so 00:06:15.125 Installing symlink pointing to librte_distributor.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_distributor.so.24 00:06:15.125 Installing symlink pointing to librte_distributor.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_distributor.so 00:06:15.125 Installing symlink pointing to librte_dmadev.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_dmadev.so.24 00:06:15.125 Installing symlink pointing to librte_dmadev.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_dmadev.so 00:06:15.125 Installing symlink pointing to librte_efd.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_efd.so.24 00:06:15.125 Installing symlink pointing to librte_efd.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_efd.so 00:06:15.125 Installing symlink pointing to librte_eventdev.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_eventdev.so.24 00:06:15.125 Installing symlink pointing to librte_eventdev.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_eventdev.so 00:06:15.125 Installing symlink pointing to librte_dispatcher.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_dispatcher.so.24 00:06:15.125 Installing symlink pointing to librte_dispatcher.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_dispatcher.so 00:06:15.125 Installing symlink pointing to librte_gpudev.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_gpudev.so.24 00:06:15.125 Installing symlink pointing to librte_gpudev.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_gpudev.so 00:06:15.125 Installing symlink pointing to librte_gro.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_gro.so.24 00:06:15.125 Installing symlink pointing to librte_gro.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_gro.so 00:06:15.125 Installing symlink pointing to librte_gso.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_gso.so.24 00:06:15.125 Installing symlink pointing to librte_gso.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_gso.so 00:06:15.125 Installing symlink pointing to librte_ip_frag.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_ip_frag.so.24 00:06:15.125 Installing symlink pointing to librte_ip_frag.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_ip_frag.so 00:06:15.125 Installing symlink pointing to librte_jobstats.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_jobstats.so.24 00:06:15.125 Installing symlink pointing to librte_jobstats.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_jobstats.so 00:06:15.125 Installing symlink pointing to librte_latencystats.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_latencystats.so.24 00:06:15.125 Installing symlink pointing to librte_latencystats.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_latencystats.so 00:06:15.125 Installing symlink pointing to librte_lpm.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_lpm.so.24 00:06:15.125 Installing symlink pointing to librte_lpm.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_lpm.so 00:06:15.125 Installing symlink pointing to librte_member.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_member.so.24 00:06:15.125 Installing symlink pointing to librte_member.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_member.so 00:06:15.125 Installing symlink pointing to librte_pcapng.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_pcapng.so.24 00:06:15.125 Installing symlink pointing to librte_pcapng.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_pcapng.so 00:06:15.125 Installing symlink pointing to librte_power.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_power.so.24 00:06:15.125 Installing symlink pointing to librte_power.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_power.so 00:06:15.125 './librte_bus_pci.so' -> 'dpdk/pmds-24.0/librte_bus_pci.so' 00:06:15.125 './librte_bus_pci.so.24' -> 'dpdk/pmds-24.0/librte_bus_pci.so.24' 00:06:15.125 './librte_bus_pci.so.24.0' -> 'dpdk/pmds-24.0/librte_bus_pci.so.24.0' 00:06:15.125 './librte_bus_vdev.so' -> 'dpdk/pmds-24.0/librte_bus_vdev.so' 00:06:15.125 './librte_bus_vdev.so.24' -> 'dpdk/pmds-24.0/librte_bus_vdev.so.24' 00:06:15.125 './librte_bus_vdev.so.24.0' -> 'dpdk/pmds-24.0/librte_bus_vdev.so.24.0' 00:06:15.125 './librte_mempool_ring.so' -> 'dpdk/pmds-24.0/librte_mempool_ring.so' 00:06:15.125 './librte_mempool_ring.so.24' -> 'dpdk/pmds-24.0/librte_mempool_ring.so.24' 00:06:15.125 './librte_mempool_ring.so.24.0' -> 'dpdk/pmds-24.0/librte_mempool_ring.so.24.0' 00:06:15.125 './librte_net_i40e.so' -> 'dpdk/pmds-24.0/librte_net_i40e.so' 00:06:15.125 './librte_net_i40e.so.24' -> 'dpdk/pmds-24.0/librte_net_i40e.so.24' 00:06:15.125 './librte_net_i40e.so.24.0' -> 'dpdk/pmds-24.0/librte_net_i40e.so.24.0' 00:06:15.125 Installing symlink pointing to librte_rawdev.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_rawdev.so.24 00:06:15.125 Installing symlink pointing to librte_rawdev.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_rawdev.so 00:06:15.125 Installing symlink pointing to librte_regexdev.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_regexdev.so.24 00:06:15.125 Installing symlink pointing to librte_regexdev.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_regexdev.so 00:06:15.125 Installing symlink pointing to librte_mldev.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_mldev.so.24 00:06:15.125 Installing symlink pointing to librte_mldev.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_mldev.so 00:06:15.126 Installing symlink pointing to librte_rib.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_rib.so.24 00:06:15.126 Installing symlink pointing to librte_rib.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_rib.so 00:06:15.126 Installing symlink pointing to librte_reorder.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_reorder.so.24 00:06:15.126 Installing symlink pointing to librte_reorder.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_reorder.so 00:06:15.126 Installing symlink pointing to librte_sched.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_sched.so.24 00:06:15.126 Installing symlink pointing to librte_sched.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_sched.so 00:06:15.126 Installing symlink pointing to librte_security.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_security.so.24 00:06:15.126 Installing symlink pointing to librte_security.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_security.so 00:06:15.126 Installing symlink pointing to librte_stack.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_stack.so.24 00:06:15.126 Installing symlink pointing to librte_stack.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_stack.so 00:06:15.126 Installing symlink pointing to librte_vhost.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_vhost.so.24 00:06:15.126 Installing symlink pointing to librte_vhost.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_vhost.so 00:06:15.126 Installing symlink pointing to librte_ipsec.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_ipsec.so.24 00:06:15.126 Installing symlink pointing to librte_ipsec.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_ipsec.so 00:06:15.126 Installing symlink pointing to librte_pdcp.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_pdcp.so.24 00:06:15.126 Installing symlink pointing to librte_pdcp.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_pdcp.so 00:06:15.126 Installing symlink pointing to librte_fib.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_fib.so.24 00:06:15.126 Installing symlink pointing to librte_fib.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_fib.so 00:06:15.126 Installing symlink pointing to librte_port.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_port.so.24 00:06:15.126 Installing symlink pointing to librte_port.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_port.so 00:06:15.126 Installing symlink pointing to librte_pdump.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_pdump.so.24 00:06:15.126 Installing symlink pointing to librte_pdump.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_pdump.so 00:06:15.126 Installing symlink pointing to librte_table.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_table.so.24 00:06:15.126 Installing symlink pointing to librte_table.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_table.so 00:06:15.126 Installing symlink pointing to librte_pipeline.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_pipeline.so.24 00:06:15.126 Installing symlink pointing to librte_pipeline.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_pipeline.so 00:06:15.126 Installing symlink pointing to librte_graph.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_graph.so.24 00:06:15.126 Installing symlink pointing to librte_graph.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_graph.so 00:06:15.126 Installing symlink pointing to librte_node.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_node.so.24 00:06:15.126 Installing symlink pointing to librte_node.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/librte_node.so 00:06:15.126 Installing symlink pointing to librte_bus_pci.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/dpdk/pmds-24.0/librte_bus_pci.so.24 00:06:15.126 Installing symlink pointing to librte_bus_pci.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/dpdk/pmds-24.0/librte_bus_pci.so 00:06:15.126 Installing symlink pointing to librte_bus_vdev.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/dpdk/pmds-24.0/librte_bus_vdev.so.24 00:06:15.126 Installing symlink pointing to librte_bus_vdev.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/dpdk/pmds-24.0/librte_bus_vdev.so 00:06:15.126 Installing symlink pointing to librte_mempool_ring.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/dpdk/pmds-24.0/librte_mempool_ring.so.24 00:06:15.126 Installing symlink pointing to librte_mempool_ring.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/dpdk/pmds-24.0/librte_mempool_ring.so 00:06:15.126 Installing symlink pointing to librte_net_i40e.so.24.0 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/dpdk/pmds-24.0/librte_net_i40e.so.24 00:06:15.126 Installing symlink pointing to librte_net_i40e.so.24 to /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/dpdk/pmds-24.0/librte_net_i40e.so 00:06:15.126 Running custom install script '/bin/sh /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/config/../buildtools/symlink-drivers-solibs.sh lib dpdk/pmds-24.0' 00:06:15.126 17:21:56 build_native_dpdk -- common/autobuild_common.sh@220 -- $ cat 00:06:15.126 17:21:56 build_native_dpdk -- common/autobuild_common.sh@225 -- $ cd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:06:15.126 00:06:15.126 real 1m28.478s 00:06:15.126 user 18m1.959s 00:06:15.126 sys 2m11.808s 00:06:15.126 17:21:56 build_native_dpdk -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:06:15.126 17:21:56 build_native_dpdk -- common/autotest_common.sh@10 -- $ set +x 00:06:15.126 ************************************ 00:06:15.126 END TEST build_native_dpdk 00:06:15.126 ************************************ 00:06:15.387 17:21:56 -- spdk/autobuild.sh@31 -- $ case "$SPDK_TEST_AUTOBUILD" in 00:06:15.387 17:21:56 -- spdk/autobuild.sh@47 -- $ [[ 0 -eq 1 ]] 00:06:15.387 17:21:56 -- spdk/autobuild.sh@51 -- $ [[ 0 -eq 1 ]] 00:06:15.387 17:21:56 -- spdk/autobuild.sh@55 -- $ [[ -n '' ]] 00:06:15.387 17:21:56 -- spdk/autobuild.sh@57 -- $ [[ 0 -eq 1 ]] 00:06:15.387 17:21:56 -- spdk/autobuild.sh@59 -- $ [[ 0 -eq 1 ]] 00:06:15.387 17:21:56 -- spdk/autobuild.sh@62 -- $ [[ 0 -eq 1 ]] 00:06:15.387 17:21:56 -- spdk/autobuild.sh@67 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/configure --enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-coverage --with-ublk --with-vfio-user --with-dpdk=/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build --with-shared 00:06:15.387 Using /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/pkgconfig for additional libs... 00:06:15.387 DPDK libraries: /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:06:15.387 DPDK includes: //var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:06:15.647 Using default SPDK env in /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk 00:06:15.909 Using 'verbs' RDMA provider 00:06:26.465 Configuring ISA-L (logfile: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/.spdk-isal.log)...done. 00:06:36.462 Configuring ISA-L-crypto (logfile: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/.spdk-isal-crypto.log)...done. 00:06:36.462 Creating mk/config.mk...done. 00:06:36.462 Creating mk/cc.flags.mk...done. 00:06:36.462 Type 'make' to build. 00:06:36.462 17:22:17 -- spdk/autobuild.sh@70 -- $ run_test make make -j48 00:06:36.462 17:22:17 -- common/autotest_common.sh@1105 -- $ '[' 3 -le 1 ']' 00:06:36.462 17:22:17 -- common/autotest_common.sh@1111 -- $ xtrace_disable 00:06:36.462 17:22:17 -- common/autotest_common.sh@10 -- $ set +x 00:06:36.462 ************************************ 00:06:36.462 START TEST make 00:06:36.462 ************************************ 00:06:36.462 17:22:17 make -- common/autotest_common.sh@1129 -- $ make -j48 00:06:36.462 make[1]: Nothing to be done for 'all'. 00:06:38.387 The Meson build system 00:06:38.387 Version: 1.5.0 00:06:38.387 Source dir: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/libvfio-user 00:06:38.387 Build dir: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug 00:06:38.387 Build type: native build 00:06:38.387 Project name: libvfio-user 00:06:38.387 Project version: 0.0.1 00:06:38.387 C compiler for the host machine: gcc (gcc 13.3.1 "gcc (GCC) 13.3.1 20240522 (Red Hat 13.3.1-1)") 00:06:38.387 C linker for the host machine: gcc ld.bfd 2.40-14 00:06:38.387 Host machine cpu family: x86_64 00:06:38.387 Host machine cpu: x86_64 00:06:38.387 Run-time dependency threads found: YES 00:06:38.387 Library dl found: YES 00:06:38.387 Found pkg-config: YES (/usr/bin/pkg-config) 1.9.5 00:06:38.387 Run-time dependency json-c found: YES 0.17 00:06:38.387 Run-time dependency cmocka found: YES 1.1.7 00:06:38.387 Program pytest-3 found: NO 00:06:38.387 Program flake8 found: NO 00:06:38.387 Program misspell-fixer found: NO 00:06:38.387 Program restructuredtext-lint found: NO 00:06:38.387 Program valgrind found: YES (/usr/bin/valgrind) 00:06:38.387 Compiler for C supports arguments -Wno-missing-field-initializers: YES 00:06:38.387 Compiler for C supports arguments -Wmissing-declarations: YES 00:06:38.387 Compiler for C supports arguments -Wwrite-strings: YES 00:06:38.387 ../libvfio-user/test/meson.build:20: WARNING: Project targets '>= 0.53.0' but uses feature introduced in '0.57.0': exclude_suites arg in add_test_setup. 00:06:38.387 Program test-lspci.sh found: YES (/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/libvfio-user/test/test-lspci.sh) 00:06:38.387 Program test-linkage.sh found: YES (/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/libvfio-user/test/test-linkage.sh) 00:06:38.387 ../libvfio-user/test/py/meson.build:16: WARNING: Project targets '>= 0.53.0' but uses feature introduced in '0.57.0': exclude_suites arg in add_test_setup. 00:06:38.387 Build targets in project: 8 00:06:38.387 WARNING: Project specifies a minimum meson_version '>= 0.53.0' but uses features which were added in newer versions: 00:06:38.387 * 0.57.0: {'exclude_suites arg in add_test_setup'} 00:06:38.387 00:06:38.387 libvfio-user 0.0.1 00:06:38.387 00:06:38.387 User defined options 00:06:38.387 buildtype : debug 00:06:38.387 default_library: shared 00:06:38.387 libdir : /usr/local/lib 00:06:38.387 00:06:38.387 Found ninja-1.11.1.git.kitware.jobserver-1 at /usr/local/bin/ninja 00:06:38.966 ninja: Entering directory `/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug' 00:06:39.238 [1/37] Compiling C object lib/libvfio-user.so.0.0.1.p/irq.c.o 00:06:39.238 [2/37] Compiling C object samples/null.p/null.c.o 00:06:39.238 [3/37] Compiling C object lib/libvfio-user.so.0.0.1.p/tran.c.o 00:06:39.238 [4/37] Compiling C object samples/client.p/.._lib_tran.c.o 00:06:39.238 [5/37] Compiling C object samples/shadow_ioeventfd_server.p/shadow_ioeventfd_server.c.o 00:06:39.238 [6/37] Compiling C object samples/client.p/.._lib_migration.c.o 00:06:39.238 [7/37] Compiling C object samples/lspci.p/lspci.c.o 00:06:39.238 [8/37] Compiling C object lib/libvfio-user.so.0.0.1.p/dma.c.o 00:06:39.238 [9/37] Compiling C object samples/gpio-pci-idio-16.p/gpio-pci-idio-16.c.o 00:06:39.238 [10/37] Compiling C object lib/libvfio-user.so.0.0.1.p/pci.c.o 00:06:39.238 [11/37] Compiling C object test/unit_tests.p/mocks.c.o 00:06:39.238 [12/37] Compiling C object test/unit_tests.p/.._lib_tran.c.o 00:06:39.238 [13/37] Compiling C object test/unit_tests.p/.._lib_irq.c.o 00:06:39.238 [14/37] Compiling C object test/unit_tests.p/.._lib_tran_sock.c.o 00:06:39.238 [15/37] Compiling C object test/unit_tests.p/.._lib_dma.c.o 00:06:39.238 [16/37] Compiling C object test/unit_tests.p/.._lib_tran_pipe.c.o 00:06:39.238 [17/37] Compiling C object lib/libvfio-user.so.0.0.1.p/tran_sock.c.o 00:06:39.238 [18/37] Compiling C object test/unit_tests.p/.._lib_pci_caps.c.o 00:06:39.238 [19/37] Compiling C object test/unit_tests.p/.._lib_pci.c.o 00:06:39.238 [20/37] Compiling C object test/unit_tests.p/.._lib_migration.c.o 00:06:39.502 [21/37] Compiling C object lib/libvfio-user.so.0.0.1.p/migration.c.o 00:06:39.502 [22/37] Compiling C object test/unit_tests.p/unit-tests.c.o 00:06:39.502 [23/37] Compiling C object samples/server.p/server.c.o 00:06:39.502 [24/37] Compiling C object lib/libvfio-user.so.0.0.1.p/pci_caps.c.o 00:06:39.502 [25/37] Compiling C object samples/client.p/.._lib_tran_sock.c.o 00:06:39.502 [26/37] Compiling C object samples/client.p/client.c.o 00:06:39.502 [27/37] Linking target samples/client 00:06:39.502 [28/37] Compiling C object lib/libvfio-user.so.0.0.1.p/libvfio-user.c.o 00:06:39.502 [29/37] Linking target lib/libvfio-user.so.0.0.1 00:06:39.502 [30/37] Compiling C object test/unit_tests.p/.._lib_libvfio-user.c.o 00:06:39.767 [31/37] Linking target test/unit_tests 00:06:39.767 [32/37] Generating symbol file lib/libvfio-user.so.0.0.1.p/libvfio-user.so.0.0.1.symbols 00:06:39.767 [33/37] Linking target samples/server 00:06:39.767 [34/37] Linking target samples/gpio-pci-idio-16 00:06:39.767 [35/37] Linking target samples/shadow_ioeventfd_server 00:06:39.767 [36/37] Linking target samples/lspci 00:06:39.767 [37/37] Linking target samples/null 00:06:39.767 INFO: autodetecting backend as ninja 00:06:39.767 INFO: calculating backend command to run: /usr/local/bin/ninja -C /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug 00:06:40.033 DESTDIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user meson install --quiet -C /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug 00:06:40.982 ninja: Entering directory `/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug' 00:06:40.982 ninja: no work to do. 00:07:19.715 CC lib/ut_mock/mock.o 00:07:19.715 CC lib/ut/ut.o 00:07:19.715 CC lib/log/log.o 00:07:19.715 CC lib/log/log_flags.o 00:07:19.715 CC lib/log/log_deprecated.o 00:07:19.715 LIB libspdk_ut.a 00:07:19.715 LIB libspdk_ut_mock.a 00:07:19.715 LIB libspdk_log.a 00:07:19.715 SO libspdk_ut_mock.so.6.0 00:07:19.715 SO libspdk_ut.so.2.0 00:07:19.715 SO libspdk_log.so.7.1 00:07:19.715 SYMLINK libspdk_ut.so 00:07:19.715 SYMLINK libspdk_ut_mock.so 00:07:19.716 SYMLINK libspdk_log.so 00:07:19.716 CC lib/dma/dma.o 00:07:19.716 CXX lib/trace_parser/trace.o 00:07:19.716 CC lib/ioat/ioat.o 00:07:19.716 CC lib/util/base64.o 00:07:19.716 CC lib/util/bit_array.o 00:07:19.716 CC lib/util/cpuset.o 00:07:19.716 CC lib/util/crc16.o 00:07:19.716 CC lib/util/crc32.o 00:07:19.716 CC lib/util/crc32c.o 00:07:19.716 CC lib/util/crc32_ieee.o 00:07:19.716 CC lib/util/crc64.o 00:07:19.716 CC lib/util/dif.o 00:07:19.716 CC lib/util/fd.o 00:07:19.716 CC lib/util/fd_group.o 00:07:19.716 CC lib/util/file.o 00:07:19.716 CC lib/util/hexlify.o 00:07:19.716 CC lib/util/iov.o 00:07:19.716 CC lib/util/math.o 00:07:19.716 CC lib/util/net.o 00:07:19.716 CC lib/util/strerror_tls.o 00:07:19.716 CC lib/util/pipe.o 00:07:19.716 CC lib/util/string.o 00:07:19.716 CC lib/util/uuid.o 00:07:19.716 CC lib/util/xor.o 00:07:19.716 CC lib/util/zipf.o 00:07:19.716 CC lib/util/md5.o 00:07:19.716 CC lib/vfio_user/host/vfio_user_pci.o 00:07:19.716 CC lib/vfio_user/host/vfio_user.o 00:07:19.716 LIB libspdk_dma.a 00:07:19.716 SO libspdk_dma.so.5.0 00:07:19.716 SYMLINK libspdk_dma.so 00:07:19.716 LIB libspdk_ioat.a 00:07:19.716 SO libspdk_ioat.so.7.0 00:07:19.716 LIB libspdk_vfio_user.a 00:07:19.716 SYMLINK libspdk_ioat.so 00:07:19.716 SO libspdk_vfio_user.so.5.0 00:07:19.716 SYMLINK libspdk_vfio_user.so 00:07:19.716 LIB libspdk_util.a 00:07:19.716 SO libspdk_util.so.10.1 00:07:19.716 SYMLINK libspdk_util.so 00:07:19.716 CC lib/conf/conf.o 00:07:19.716 CC lib/rdma_utils/rdma_utils.o 00:07:19.716 CC lib/idxd/idxd.o 00:07:19.716 CC lib/json/json_parse.o 00:07:19.716 CC lib/vmd/vmd.o 00:07:19.716 CC lib/json/json_util.o 00:07:19.716 CC lib/idxd/idxd_user.o 00:07:19.716 CC lib/vmd/led.o 00:07:19.716 CC lib/env_dpdk/env.o 00:07:19.716 CC lib/json/json_write.o 00:07:19.716 CC lib/idxd/idxd_kernel.o 00:07:19.716 CC lib/env_dpdk/memory.o 00:07:19.716 CC lib/env_dpdk/pci.o 00:07:19.716 CC lib/env_dpdk/init.o 00:07:19.716 CC lib/env_dpdk/threads.o 00:07:19.716 CC lib/env_dpdk/pci_ioat.o 00:07:19.716 CC lib/env_dpdk/pci_virtio.o 00:07:19.716 CC lib/env_dpdk/pci_vmd.o 00:07:19.716 CC lib/env_dpdk/pci_idxd.o 00:07:19.716 CC lib/env_dpdk/pci_event.o 00:07:19.716 CC lib/env_dpdk/sigbus_handler.o 00:07:19.716 CC lib/env_dpdk/pci_dpdk.o 00:07:19.716 CC lib/env_dpdk/pci_dpdk_2211.o 00:07:19.716 CC lib/env_dpdk/pci_dpdk_2207.o 00:07:19.716 LIB libspdk_conf.a 00:07:19.716 LIB libspdk_json.a 00:07:19.716 LIB libspdk_rdma_utils.a 00:07:19.716 SO libspdk_conf.so.6.0 00:07:19.716 SO libspdk_json.so.6.0 00:07:19.716 SO libspdk_rdma_utils.so.1.0 00:07:19.716 SYMLINK libspdk_conf.so 00:07:19.716 SYMLINK libspdk_json.so 00:07:19.716 SYMLINK libspdk_rdma_utils.so 00:07:19.716 CC lib/jsonrpc/jsonrpc_server.o 00:07:19.716 CC lib/jsonrpc/jsonrpc_server_tcp.o 00:07:19.716 CC lib/jsonrpc/jsonrpc_client.o 00:07:19.716 CC lib/rdma_provider/common.o 00:07:19.716 CC lib/jsonrpc/jsonrpc_client_tcp.o 00:07:19.716 CC lib/rdma_provider/rdma_provider_verbs.o 00:07:19.716 LIB libspdk_idxd.a 00:07:19.716 SO libspdk_idxd.so.12.1 00:07:19.716 LIB libspdk_vmd.a 00:07:19.716 SO libspdk_vmd.so.6.0 00:07:19.716 SYMLINK libspdk_idxd.so 00:07:19.716 SYMLINK libspdk_vmd.so 00:07:19.716 LIB libspdk_rdma_provider.a 00:07:19.716 SO libspdk_rdma_provider.so.7.0 00:07:19.716 LIB libspdk_jsonrpc.a 00:07:19.716 LIB libspdk_trace_parser.a 00:07:19.716 SO libspdk_trace_parser.so.6.0 00:07:19.716 SO libspdk_jsonrpc.so.6.0 00:07:19.716 SYMLINK libspdk_rdma_provider.so 00:07:19.716 SYMLINK libspdk_jsonrpc.so 00:07:19.716 SYMLINK libspdk_trace_parser.so 00:07:19.716 CC lib/rpc/rpc.o 00:07:19.716 LIB libspdk_rpc.a 00:07:19.716 SO libspdk_rpc.so.6.0 00:07:19.716 SYMLINK libspdk_rpc.so 00:07:19.716 CC lib/trace/trace.o 00:07:19.716 CC lib/notify/notify.o 00:07:19.716 CC lib/keyring/keyring.o 00:07:19.716 CC lib/trace/trace_flags.o 00:07:19.716 CC lib/notify/notify_rpc.o 00:07:19.716 CC lib/keyring/keyring_rpc.o 00:07:19.716 CC lib/trace/trace_rpc.o 00:07:19.975 LIB libspdk_notify.a 00:07:19.975 SO libspdk_notify.so.6.0 00:07:19.975 SYMLINK libspdk_notify.so 00:07:19.975 LIB libspdk_keyring.a 00:07:20.234 LIB libspdk_trace.a 00:07:20.234 SO libspdk_keyring.so.2.0 00:07:20.234 SO libspdk_trace.so.11.0 00:07:20.234 SYMLINK libspdk_keyring.so 00:07:20.234 SYMLINK libspdk_trace.so 00:07:20.234 CC lib/thread/thread.o 00:07:20.234 CC lib/thread/iobuf.o 00:07:20.234 CC lib/sock/sock.o 00:07:20.234 CC lib/sock/sock_rpc.o 00:07:20.234 LIB libspdk_env_dpdk.a 00:07:20.493 SO libspdk_env_dpdk.so.15.1 00:07:20.493 SYMLINK libspdk_env_dpdk.so 00:07:20.753 LIB libspdk_sock.a 00:07:20.753 SO libspdk_sock.so.10.0 00:07:20.753 SYMLINK libspdk_sock.so 00:07:21.013 CC lib/nvme/nvme_ctrlr_cmd.o 00:07:21.013 CC lib/nvme/nvme_ctrlr.o 00:07:21.013 CC lib/nvme/nvme_fabric.o 00:07:21.013 CC lib/nvme/nvme_ns_cmd.o 00:07:21.013 CC lib/nvme/nvme_ns.o 00:07:21.013 CC lib/nvme/nvme_pcie_common.o 00:07:21.013 CC lib/nvme/nvme_pcie.o 00:07:21.013 CC lib/nvme/nvme_qpair.o 00:07:21.013 CC lib/nvme/nvme.o 00:07:21.013 CC lib/nvme/nvme_quirks.o 00:07:21.013 CC lib/nvme/nvme_transport.o 00:07:21.013 CC lib/nvme/nvme_discovery.o 00:07:21.013 CC lib/nvme/nvme_ctrlr_ocssd_cmd.o 00:07:21.013 CC lib/nvme/nvme_ns_ocssd_cmd.o 00:07:21.013 CC lib/nvme/nvme_tcp.o 00:07:21.013 CC lib/nvme/nvme_opal.o 00:07:21.013 CC lib/nvme/nvme_io_msg.o 00:07:21.013 CC lib/nvme/nvme_poll_group.o 00:07:21.013 CC lib/nvme/nvme_zns.o 00:07:21.013 CC lib/nvme/nvme_stubs.o 00:07:21.013 CC lib/nvme/nvme_auth.o 00:07:21.013 CC lib/nvme/nvme_cuse.o 00:07:21.013 CC lib/nvme/nvme_vfio_user.o 00:07:21.013 CC lib/nvme/nvme_rdma.o 00:07:21.954 LIB libspdk_thread.a 00:07:21.954 SO libspdk_thread.so.11.0 00:07:21.954 SYMLINK libspdk_thread.so 00:07:22.213 CC lib/accel/accel.o 00:07:22.213 CC lib/accel/accel_rpc.o 00:07:22.213 CC lib/accel/accel_sw.o 00:07:22.213 CC lib/fsdev/fsdev.o 00:07:22.213 CC lib/fsdev/fsdev_io.o 00:07:22.213 CC lib/vfu_tgt/tgt_endpoint.o 00:07:22.213 CC lib/init/json_config.o 00:07:22.213 CC lib/virtio/virtio.o 00:07:22.213 CC lib/vfu_tgt/tgt_rpc.o 00:07:22.213 CC lib/blob/blobstore.o 00:07:22.213 CC lib/fsdev/fsdev_rpc.o 00:07:22.213 CC lib/virtio/virtio_vhost_user.o 00:07:22.213 CC lib/init/subsystem.o 00:07:22.213 CC lib/virtio/virtio_vfio_user.o 00:07:22.213 CC lib/blob/request.o 00:07:22.213 CC lib/blob/zeroes.o 00:07:22.213 CC lib/init/subsystem_rpc.o 00:07:22.213 CC lib/virtio/virtio_pci.o 00:07:22.213 CC lib/blob/blob_bs_dev.o 00:07:22.213 CC lib/init/rpc.o 00:07:22.472 LIB libspdk_init.a 00:07:22.472 SO libspdk_init.so.6.0 00:07:22.472 LIB libspdk_virtio.a 00:07:22.472 SYMLINK libspdk_init.so 00:07:22.472 LIB libspdk_vfu_tgt.a 00:07:22.730 SO libspdk_virtio.so.7.0 00:07:22.730 SO libspdk_vfu_tgt.so.3.0 00:07:22.730 SYMLINK libspdk_virtio.so 00:07:22.730 SYMLINK libspdk_vfu_tgt.so 00:07:22.730 CC lib/event/app.o 00:07:22.730 CC lib/event/reactor.o 00:07:22.730 CC lib/event/log_rpc.o 00:07:22.730 CC lib/event/app_rpc.o 00:07:22.730 CC lib/event/scheduler_static.o 00:07:22.987 LIB libspdk_fsdev.a 00:07:22.987 SO libspdk_fsdev.so.2.0 00:07:22.987 SYMLINK libspdk_fsdev.so 00:07:23.244 CC lib/fuse_dispatcher/fuse_dispatcher.o 00:07:23.244 LIB libspdk_event.a 00:07:23.244 SO libspdk_event.so.14.0 00:07:23.244 SYMLINK libspdk_event.so 00:07:23.501 LIB libspdk_accel.a 00:07:23.501 SO libspdk_accel.so.16.0 00:07:23.501 LIB libspdk_nvme.a 00:07:23.501 SYMLINK libspdk_accel.so 00:07:23.501 SO libspdk_nvme.so.15.0 00:07:23.797 CC lib/bdev/bdev.o 00:07:23.797 CC lib/bdev/bdev_rpc.o 00:07:23.797 CC lib/bdev/bdev_zone.o 00:07:23.797 CC lib/bdev/part.o 00:07:23.797 CC lib/bdev/scsi_nvme.o 00:07:23.797 LIB libspdk_fuse_dispatcher.a 00:07:23.797 SO libspdk_fuse_dispatcher.so.1.0 00:07:23.797 SYMLINK libspdk_nvme.so 00:07:23.797 SYMLINK libspdk_fuse_dispatcher.so 00:07:25.700 LIB libspdk_blob.a 00:07:25.700 SO libspdk_blob.so.12.0 00:07:25.700 SYMLINK libspdk_blob.so 00:07:25.700 CC lib/blobfs/blobfs.o 00:07:25.700 CC lib/blobfs/tree.o 00:07:25.700 CC lib/lvol/lvol.o 00:07:26.268 LIB libspdk_bdev.a 00:07:26.268 SO libspdk_bdev.so.17.0 00:07:26.532 LIB libspdk_blobfs.a 00:07:26.532 SYMLINK libspdk_bdev.so 00:07:26.532 SO libspdk_blobfs.so.11.0 00:07:26.532 SYMLINK libspdk_blobfs.so 00:07:26.532 LIB libspdk_lvol.a 00:07:26.532 SO libspdk_lvol.so.11.0 00:07:26.532 CC lib/nbd/nbd.o 00:07:26.532 CC lib/ublk/ublk.o 00:07:26.532 CC lib/nbd/nbd_rpc.o 00:07:26.532 CC lib/ublk/ublk_rpc.o 00:07:26.532 CC lib/scsi/dev.o 00:07:26.532 CC lib/nvmf/ctrlr.o 00:07:26.532 CC lib/scsi/lun.o 00:07:26.532 CC lib/nvmf/ctrlr_discovery.o 00:07:26.532 CC lib/scsi/port.o 00:07:26.532 CC lib/ftl/ftl_core.o 00:07:26.532 CC lib/nvmf/ctrlr_bdev.o 00:07:26.532 CC lib/scsi/scsi.o 00:07:26.532 CC lib/ftl/ftl_init.o 00:07:26.532 CC lib/nvmf/subsystem.o 00:07:26.532 CC lib/ftl/ftl_layout.o 00:07:26.532 CC lib/scsi/scsi_bdev.o 00:07:26.532 CC lib/ftl/ftl_debug.o 00:07:26.532 CC lib/scsi/scsi_pr.o 00:07:26.532 CC lib/nvmf/nvmf.o 00:07:26.532 CC lib/ftl/ftl_io.o 00:07:26.532 CC lib/nvmf/nvmf_rpc.o 00:07:26.532 CC lib/scsi/scsi_rpc.o 00:07:26.532 CC lib/scsi/task.o 00:07:26.532 CC lib/ftl/ftl_sb.o 00:07:26.532 CC lib/nvmf/transport.o 00:07:26.532 CC lib/nvmf/tcp.o 00:07:26.532 CC lib/ftl/ftl_l2p.o 00:07:26.532 CC lib/ftl/ftl_l2p_flat.o 00:07:26.532 CC lib/nvmf/stubs.o 00:07:26.532 CC lib/ftl/ftl_nv_cache.o 00:07:26.532 CC lib/nvmf/mdns_server.o 00:07:26.532 CC lib/nvmf/vfio_user.o 00:07:26.532 CC lib/ftl/ftl_band.o 00:07:26.532 CC lib/ftl/ftl_band_ops.o 00:07:26.532 CC lib/nvmf/rdma.o 00:07:26.532 CC lib/ftl/ftl_writer.o 00:07:26.532 CC lib/nvmf/auth.o 00:07:26.532 CC lib/ftl/ftl_rq.o 00:07:26.532 CC lib/ftl/ftl_reloc.o 00:07:26.532 CC lib/ftl/ftl_l2p_cache.o 00:07:26.532 CC lib/ftl/ftl_p2l.o 00:07:26.532 CC lib/ftl/ftl_p2l_log.o 00:07:26.532 CC lib/ftl/mngt/ftl_mngt.o 00:07:26.532 CC lib/ftl/mngt/ftl_mngt_bdev.o 00:07:26.532 CC lib/ftl/mngt/ftl_mngt_shutdown.o 00:07:26.532 CC lib/ftl/mngt/ftl_mngt_startup.o 00:07:26.532 CC lib/ftl/mngt/ftl_mngt_md.o 00:07:26.532 SYMLINK libspdk_lvol.so 00:07:26.532 CC lib/ftl/mngt/ftl_mngt_misc.o 00:07:27.114 CC lib/ftl/mngt/ftl_mngt_ioch.o 00:07:27.114 CC lib/ftl/mngt/ftl_mngt_l2p.o 00:07:27.114 CC lib/ftl/mngt/ftl_mngt_band.o 00:07:27.114 CC lib/ftl/mngt/ftl_mngt_self_test.o 00:07:27.114 CC lib/ftl/mngt/ftl_mngt_p2l.o 00:07:27.114 CC lib/ftl/mngt/ftl_mngt_recovery.o 00:07:27.114 CC lib/ftl/mngt/ftl_mngt_upgrade.o 00:07:27.114 CC lib/ftl/utils/ftl_conf.o 00:07:27.114 CC lib/ftl/utils/ftl_md.o 00:07:27.114 CC lib/ftl/utils/ftl_mempool.o 00:07:27.114 CC lib/ftl/utils/ftl_bitmap.o 00:07:27.114 CC lib/ftl/utils/ftl_property.o 00:07:27.114 CC lib/ftl/utils/ftl_layout_tracker_bdev.o 00:07:27.114 CC lib/ftl/upgrade/ftl_layout_upgrade.o 00:07:27.114 CC lib/ftl/upgrade/ftl_sb_upgrade.o 00:07:27.114 CC lib/ftl/upgrade/ftl_p2l_upgrade.o 00:07:27.114 CC lib/ftl/upgrade/ftl_band_upgrade.o 00:07:27.114 CC lib/ftl/upgrade/ftl_chunk_upgrade.o 00:07:27.114 CC lib/ftl/upgrade/ftl_trim_upgrade.o 00:07:27.377 CC lib/ftl/upgrade/ftl_sb_v3.o 00:07:27.377 CC lib/ftl/upgrade/ftl_sb_v5.o 00:07:27.377 CC lib/ftl/nvc/ftl_nvc_dev.o 00:07:27.377 CC lib/ftl/nvc/ftl_nvc_bdev_vss.o 00:07:27.377 CC lib/ftl/nvc/ftl_nvc_bdev_non_vss.o 00:07:27.377 CC lib/ftl/nvc/ftl_nvc_bdev_common.o 00:07:27.377 CC lib/ftl/base/ftl_base_dev.o 00:07:27.377 CC lib/ftl/base/ftl_base_bdev.o 00:07:27.377 CC lib/ftl/ftl_trace.o 00:07:27.637 LIB libspdk_nbd.a 00:07:27.637 SO libspdk_nbd.so.7.0 00:07:27.637 LIB libspdk_scsi.a 00:07:27.637 SYMLINK libspdk_nbd.so 00:07:27.637 SO libspdk_scsi.so.9.0 00:07:27.637 LIB libspdk_ublk.a 00:07:27.637 SO libspdk_ublk.so.3.0 00:07:27.637 SYMLINK libspdk_scsi.so 00:07:27.897 SYMLINK libspdk_ublk.so 00:07:27.897 CC lib/iscsi/conn.o 00:07:27.897 CC lib/vhost/vhost.o 00:07:27.897 CC lib/iscsi/init_grp.o 00:07:27.897 CC lib/vhost/vhost_rpc.o 00:07:27.897 CC lib/iscsi/iscsi.o 00:07:27.897 CC lib/vhost/vhost_scsi.o 00:07:27.897 CC lib/iscsi/param.o 00:07:27.897 CC lib/vhost/vhost_blk.o 00:07:27.897 CC lib/iscsi/portal_grp.o 00:07:27.897 CC lib/vhost/rte_vhost_user.o 00:07:27.897 CC lib/iscsi/tgt_node.o 00:07:27.897 CC lib/iscsi/iscsi_rpc.o 00:07:27.897 CC lib/iscsi/iscsi_subsystem.o 00:07:27.897 CC lib/iscsi/task.o 00:07:28.156 LIB libspdk_ftl.a 00:07:28.156 SO libspdk_ftl.so.9.0 00:07:28.415 SYMLINK libspdk_ftl.so 00:07:29.352 LIB libspdk_vhost.a 00:07:29.352 SO libspdk_vhost.so.8.0 00:07:29.352 SYMLINK libspdk_vhost.so 00:07:29.352 LIB libspdk_nvmf.a 00:07:29.352 SO libspdk_nvmf.so.20.0 00:07:29.352 LIB libspdk_iscsi.a 00:07:29.352 SO libspdk_iscsi.so.8.0 00:07:29.611 SYMLINK libspdk_nvmf.so 00:07:29.611 SYMLINK libspdk_iscsi.so 00:07:29.871 CC module/env_dpdk/env_dpdk_rpc.o 00:07:29.871 CC module/vfu_device/vfu_virtio.o 00:07:29.871 CC module/vfu_device/vfu_virtio_blk.o 00:07:29.871 CC module/vfu_device/vfu_virtio_scsi.o 00:07:29.871 CC module/vfu_device/vfu_virtio_rpc.o 00:07:29.871 CC module/vfu_device/vfu_virtio_fs.o 00:07:29.871 CC module/sock/posix/posix.o 00:07:29.871 CC module/scheduler/gscheduler/gscheduler.o 00:07:29.871 CC module/scheduler/dynamic/scheduler_dynamic.o 00:07:29.871 CC module/keyring/file/keyring.o 00:07:29.871 CC module/keyring/linux/keyring.o 00:07:29.871 CC module/keyring/linux/keyring_rpc.o 00:07:29.871 CC module/keyring/file/keyring_rpc.o 00:07:29.871 CC module/accel/dsa/accel_dsa.o 00:07:29.871 CC module/scheduler/dpdk_governor/dpdk_governor.o 00:07:29.871 CC module/accel/iaa/accel_iaa.o 00:07:29.871 CC module/accel/dsa/accel_dsa_rpc.o 00:07:29.871 CC module/accel/iaa/accel_iaa_rpc.o 00:07:29.871 CC module/accel/ioat/accel_ioat.o 00:07:29.871 CC module/fsdev/aio/fsdev_aio.o 00:07:29.871 CC module/blob/bdev/blob_bdev.o 00:07:29.871 CC module/fsdev/aio/fsdev_aio_rpc.o 00:07:29.871 CC module/accel/ioat/accel_ioat_rpc.o 00:07:29.871 CC module/fsdev/aio/linux_aio_mgr.o 00:07:29.871 CC module/accel/error/accel_error.o 00:07:29.871 CC module/accel/error/accel_error_rpc.o 00:07:29.871 LIB libspdk_env_dpdk_rpc.a 00:07:30.131 SO libspdk_env_dpdk_rpc.so.6.0 00:07:30.131 SYMLINK libspdk_env_dpdk_rpc.so 00:07:30.131 LIB libspdk_keyring_file.a 00:07:30.131 LIB libspdk_scheduler_dpdk_governor.a 00:07:30.131 LIB libspdk_scheduler_gscheduler.a 00:07:30.131 SO libspdk_keyring_file.so.2.0 00:07:30.131 SO libspdk_scheduler_dpdk_governor.so.4.0 00:07:30.131 SO libspdk_scheduler_gscheduler.so.4.0 00:07:30.131 LIB libspdk_keyring_linux.a 00:07:30.131 LIB libspdk_scheduler_dynamic.a 00:07:30.131 SYMLINK libspdk_scheduler_gscheduler.so 00:07:30.131 SO libspdk_keyring_linux.so.1.0 00:07:30.131 SYMLINK libspdk_scheduler_dpdk_governor.so 00:07:30.131 SYMLINK libspdk_keyring_file.so 00:07:30.131 SO libspdk_scheduler_dynamic.so.4.0 00:07:30.131 LIB libspdk_accel_ioat.a 00:07:30.131 SYMLINK libspdk_scheduler_dynamic.so 00:07:30.131 SYMLINK libspdk_keyring_linux.so 00:07:30.131 SO libspdk_accel_ioat.so.6.0 00:07:30.131 LIB libspdk_accel_dsa.a 00:07:30.131 LIB libspdk_blob_bdev.a 00:07:30.131 LIB libspdk_accel_error.a 00:07:30.131 LIB libspdk_accel_iaa.a 00:07:30.390 SO libspdk_blob_bdev.so.12.0 00:07:30.390 SO libspdk_accel_dsa.so.5.0 00:07:30.390 SO libspdk_accel_iaa.so.3.0 00:07:30.390 SO libspdk_accel_error.so.2.0 00:07:30.390 SYMLINK libspdk_accel_ioat.so 00:07:30.390 SYMLINK libspdk_blob_bdev.so 00:07:30.390 SYMLINK libspdk_accel_dsa.so 00:07:30.390 SYMLINK libspdk_accel_iaa.so 00:07:30.390 SYMLINK libspdk_accel_error.so 00:07:30.652 LIB libspdk_vfu_device.a 00:07:30.652 SO libspdk_vfu_device.so.3.0 00:07:30.652 CC module/blobfs/bdev/blobfs_bdev.o 00:07:30.652 CC module/blobfs/bdev/blobfs_bdev_rpc.o 00:07:30.652 CC module/bdev/gpt/gpt.o 00:07:30.652 CC module/bdev/gpt/vbdev_gpt.o 00:07:30.652 CC module/bdev/delay/vbdev_delay.o 00:07:30.652 CC module/bdev/delay/vbdev_delay_rpc.o 00:07:30.652 CC module/bdev/lvol/vbdev_lvol.o 00:07:30.652 CC module/bdev/malloc/bdev_malloc.o 00:07:30.652 CC module/bdev/lvol/vbdev_lvol_rpc.o 00:07:30.652 CC module/bdev/malloc/bdev_malloc_rpc.o 00:07:30.652 CC module/bdev/error/vbdev_error.o 00:07:30.652 CC module/bdev/error/vbdev_error_rpc.o 00:07:30.652 CC module/bdev/raid/bdev_raid_rpc.o 00:07:30.652 CC module/bdev/raid/bdev_raid.o 00:07:30.652 CC module/bdev/null/bdev_null.o 00:07:30.652 CC module/bdev/null/bdev_null_rpc.o 00:07:30.652 CC module/bdev/raid/bdev_raid_sb.o 00:07:30.652 CC module/bdev/raid/raid0.o 00:07:30.652 CC module/bdev/aio/bdev_aio.o 00:07:30.652 CC module/bdev/iscsi/bdev_iscsi.o 00:07:30.652 CC module/bdev/aio/bdev_aio_rpc.o 00:07:30.652 CC module/bdev/split/vbdev_split.o 00:07:30.652 CC module/bdev/raid/raid1.o 00:07:30.652 CC module/bdev/iscsi/bdev_iscsi_rpc.o 00:07:30.652 CC module/bdev/raid/concat.o 00:07:30.652 CC module/bdev/virtio/bdev_virtio_scsi.o 00:07:30.652 CC module/bdev/split/vbdev_split_rpc.o 00:07:30.652 CC module/bdev/nvme/bdev_nvme.o 00:07:30.652 CC module/bdev/virtio/bdev_virtio_blk.o 00:07:30.652 CC module/bdev/ftl/bdev_ftl.o 00:07:30.652 CC module/bdev/nvme/bdev_nvme_rpc.o 00:07:30.652 CC module/bdev/virtio/bdev_virtio_rpc.o 00:07:30.652 CC module/bdev/nvme/nvme_rpc.o 00:07:30.652 CC module/bdev/ftl/bdev_ftl_rpc.o 00:07:30.652 CC module/bdev/nvme/bdev_mdns_client.o 00:07:30.652 CC module/bdev/nvme/vbdev_opal.o 00:07:30.652 CC module/bdev/nvme/vbdev_opal_rpc.o 00:07:30.652 CC module/bdev/passthru/vbdev_passthru.o 00:07:30.652 CC module/bdev/nvme/bdev_nvme_cuse_rpc.o 00:07:30.652 CC module/bdev/passthru/vbdev_passthru_rpc.o 00:07:30.652 CC module/bdev/zone_block/vbdev_zone_block.o 00:07:30.652 CC module/bdev/zone_block/vbdev_zone_block_rpc.o 00:07:30.652 SYMLINK libspdk_vfu_device.so 00:07:30.652 LIB libspdk_fsdev_aio.a 00:07:30.652 SO libspdk_fsdev_aio.so.1.0 00:07:30.912 LIB libspdk_sock_posix.a 00:07:30.912 SO libspdk_sock_posix.so.6.0 00:07:30.912 SYMLINK libspdk_fsdev_aio.so 00:07:30.912 LIB libspdk_blobfs_bdev.a 00:07:30.912 SYMLINK libspdk_sock_posix.so 00:07:30.912 SO libspdk_blobfs_bdev.so.6.0 00:07:31.171 LIB libspdk_bdev_split.a 00:07:31.171 SYMLINK libspdk_blobfs_bdev.so 00:07:31.171 SO libspdk_bdev_split.so.6.0 00:07:31.171 LIB libspdk_bdev_gpt.a 00:07:31.171 LIB libspdk_bdev_ftl.a 00:07:31.171 LIB libspdk_bdev_passthru.a 00:07:31.171 LIB libspdk_bdev_null.a 00:07:31.171 SO libspdk_bdev_gpt.so.6.0 00:07:31.171 SO libspdk_bdev_ftl.so.6.0 00:07:31.171 LIB libspdk_bdev_error.a 00:07:31.171 SO libspdk_bdev_passthru.so.6.0 00:07:31.171 SYMLINK libspdk_bdev_split.so 00:07:31.171 SO libspdk_bdev_error.so.6.0 00:07:31.171 SO libspdk_bdev_null.so.6.0 00:07:31.171 SYMLINK libspdk_bdev_gpt.so 00:07:31.171 SYMLINK libspdk_bdev_ftl.so 00:07:31.171 SYMLINK libspdk_bdev_passthru.so 00:07:31.171 LIB libspdk_bdev_aio.a 00:07:31.171 SYMLINK libspdk_bdev_null.so 00:07:31.171 SYMLINK libspdk_bdev_error.so 00:07:31.171 LIB libspdk_bdev_zone_block.a 00:07:31.171 SO libspdk_bdev_aio.so.6.0 00:07:31.171 SO libspdk_bdev_zone_block.so.6.0 00:07:31.171 LIB libspdk_bdev_delay.a 00:07:31.171 LIB libspdk_bdev_malloc.a 00:07:31.171 SO libspdk_bdev_delay.so.6.0 00:07:31.171 SO libspdk_bdev_malloc.so.6.0 00:07:31.171 LIB libspdk_bdev_iscsi.a 00:07:31.171 SYMLINK libspdk_bdev_aio.so 00:07:31.171 SYMLINK libspdk_bdev_zone_block.so 00:07:31.171 SO libspdk_bdev_iscsi.so.6.0 00:07:31.430 SYMLINK libspdk_bdev_delay.so 00:07:31.430 SYMLINK libspdk_bdev_malloc.so 00:07:31.430 SYMLINK libspdk_bdev_iscsi.so 00:07:31.430 LIB libspdk_bdev_lvol.a 00:07:31.430 LIB libspdk_bdev_virtio.a 00:07:31.430 SO libspdk_bdev_lvol.so.6.0 00:07:31.430 SO libspdk_bdev_virtio.so.6.0 00:07:31.430 SYMLINK libspdk_bdev_lvol.so 00:07:31.430 SYMLINK libspdk_bdev_virtio.so 00:07:31.690 LIB libspdk_bdev_raid.a 00:07:31.690 SO libspdk_bdev_raid.so.6.0 00:07:31.950 SYMLINK libspdk_bdev_raid.so 00:07:33.337 LIB libspdk_bdev_nvme.a 00:07:33.337 SO libspdk_bdev_nvme.so.7.1 00:07:33.596 SYMLINK libspdk_bdev_nvme.so 00:07:33.855 CC module/event/subsystems/iobuf/iobuf.o 00:07:33.855 CC module/event/subsystems/vhost_blk/vhost_blk.o 00:07:33.855 CC module/event/subsystems/iobuf/iobuf_rpc.o 00:07:33.855 CC module/event/subsystems/vfu_tgt/vfu_tgt.o 00:07:33.855 CC module/event/subsystems/sock/sock.o 00:07:33.855 CC module/event/subsystems/keyring/keyring.o 00:07:33.855 CC module/event/subsystems/scheduler/scheduler.o 00:07:33.855 CC module/event/subsystems/fsdev/fsdev.o 00:07:33.855 CC module/event/subsystems/vmd/vmd.o 00:07:33.855 CC module/event/subsystems/vmd/vmd_rpc.o 00:07:34.116 LIB libspdk_event_keyring.a 00:07:34.116 LIB libspdk_event_vhost_blk.a 00:07:34.116 LIB libspdk_event_scheduler.a 00:07:34.116 LIB libspdk_event_fsdev.a 00:07:34.116 LIB libspdk_event_sock.a 00:07:34.116 LIB libspdk_event_vmd.a 00:07:34.116 LIB libspdk_event_vfu_tgt.a 00:07:34.116 SO libspdk_event_keyring.so.1.0 00:07:34.116 LIB libspdk_event_iobuf.a 00:07:34.116 SO libspdk_event_vhost_blk.so.3.0 00:07:34.116 SO libspdk_event_scheduler.so.4.0 00:07:34.116 SO libspdk_event_fsdev.so.1.0 00:07:34.116 SO libspdk_event_sock.so.5.0 00:07:34.116 SO libspdk_event_vfu_tgt.so.3.0 00:07:34.116 SO libspdk_event_vmd.so.6.0 00:07:34.116 SO libspdk_event_iobuf.so.3.0 00:07:34.116 SYMLINK libspdk_event_vhost_blk.so 00:07:34.116 SYMLINK libspdk_event_keyring.so 00:07:34.116 SYMLINK libspdk_event_scheduler.so 00:07:34.116 SYMLINK libspdk_event_fsdev.so 00:07:34.116 SYMLINK libspdk_event_sock.so 00:07:34.116 SYMLINK libspdk_event_vfu_tgt.so 00:07:34.116 SYMLINK libspdk_event_vmd.so 00:07:34.116 SYMLINK libspdk_event_iobuf.so 00:07:34.377 CC module/event/subsystems/accel/accel.o 00:07:34.377 LIB libspdk_event_accel.a 00:07:34.377 SO libspdk_event_accel.so.6.0 00:07:34.637 SYMLINK libspdk_event_accel.so 00:07:34.637 CC module/event/subsystems/bdev/bdev.o 00:07:34.897 LIB libspdk_event_bdev.a 00:07:34.897 SO libspdk_event_bdev.so.6.0 00:07:34.897 SYMLINK libspdk_event_bdev.so 00:07:35.156 CC module/event/subsystems/scsi/scsi.o 00:07:35.156 CC module/event/subsystems/nbd/nbd.o 00:07:35.156 CC module/event/subsystems/ublk/ublk.o 00:07:35.156 CC module/event/subsystems/nvmf/nvmf_rpc.o 00:07:35.156 CC module/event/subsystems/nvmf/nvmf_tgt.o 00:07:35.156 LIB libspdk_event_ublk.a 00:07:35.156 LIB libspdk_event_nbd.a 00:07:35.415 LIB libspdk_event_scsi.a 00:07:35.415 SO libspdk_event_ublk.so.3.0 00:07:35.415 SO libspdk_event_nbd.so.6.0 00:07:35.415 SO libspdk_event_scsi.so.6.0 00:07:35.415 SYMLINK libspdk_event_ublk.so 00:07:35.415 SYMLINK libspdk_event_nbd.so 00:07:35.415 SYMLINK libspdk_event_scsi.so 00:07:35.415 LIB libspdk_event_nvmf.a 00:07:35.415 SO libspdk_event_nvmf.so.6.0 00:07:35.415 SYMLINK libspdk_event_nvmf.so 00:07:35.415 CC module/event/subsystems/iscsi/iscsi.o 00:07:35.415 CC module/event/subsystems/vhost_scsi/vhost_scsi.o 00:07:35.674 LIB libspdk_event_vhost_scsi.a 00:07:35.674 LIB libspdk_event_iscsi.a 00:07:35.674 SO libspdk_event_vhost_scsi.so.3.0 00:07:35.674 SO libspdk_event_iscsi.so.6.0 00:07:35.674 SYMLINK libspdk_event_vhost_scsi.so 00:07:35.674 SYMLINK libspdk_event_iscsi.so 00:07:35.936 SO libspdk.so.6.0 00:07:35.936 SYMLINK libspdk.so 00:07:35.936 CXX app/trace/trace.o 00:07:35.936 CC test/rpc_client/rpc_client_test.o 00:07:35.936 CC app/trace_record/trace_record.o 00:07:35.936 TEST_HEADER include/spdk/accel.h 00:07:35.936 TEST_HEADER include/spdk/accel_module.h 00:07:36.201 CC app/spdk_nvme_perf/perf.o 00:07:36.201 TEST_HEADER include/spdk/assert.h 00:07:36.201 TEST_HEADER include/spdk/barrier.h 00:07:36.201 TEST_HEADER include/spdk/base64.h 00:07:36.201 TEST_HEADER include/spdk/bdev.h 00:07:36.201 CC app/spdk_nvme_discover/discovery_aer.o 00:07:36.201 TEST_HEADER include/spdk/bdev_module.h 00:07:36.201 CC app/spdk_top/spdk_top.o 00:07:36.201 TEST_HEADER include/spdk/bdev_zone.h 00:07:36.202 TEST_HEADER include/spdk/bit_array.h 00:07:36.202 TEST_HEADER include/spdk/bit_pool.h 00:07:36.202 CC app/spdk_nvme_identify/identify.o 00:07:36.202 TEST_HEADER include/spdk/blob_bdev.h 00:07:36.202 CC app/spdk_lspci/spdk_lspci.o 00:07:36.202 TEST_HEADER include/spdk/blobfs_bdev.h 00:07:36.202 TEST_HEADER include/spdk/blobfs.h 00:07:36.202 TEST_HEADER include/spdk/blob.h 00:07:36.202 TEST_HEADER include/spdk/conf.h 00:07:36.202 TEST_HEADER include/spdk/config.h 00:07:36.202 TEST_HEADER include/spdk/cpuset.h 00:07:36.202 TEST_HEADER include/spdk/crc16.h 00:07:36.202 TEST_HEADER include/spdk/crc32.h 00:07:36.202 TEST_HEADER include/spdk/dif.h 00:07:36.202 TEST_HEADER include/spdk/crc64.h 00:07:36.202 TEST_HEADER include/spdk/dma.h 00:07:36.202 TEST_HEADER include/spdk/env_dpdk.h 00:07:36.202 TEST_HEADER include/spdk/endian.h 00:07:36.202 TEST_HEADER include/spdk/env.h 00:07:36.202 TEST_HEADER include/spdk/event.h 00:07:36.202 TEST_HEADER include/spdk/fd_group.h 00:07:36.202 TEST_HEADER include/spdk/fd.h 00:07:36.202 TEST_HEADER include/spdk/file.h 00:07:36.202 TEST_HEADER include/spdk/fsdev_module.h 00:07:36.202 TEST_HEADER include/spdk/fsdev.h 00:07:36.202 TEST_HEADER include/spdk/ftl.h 00:07:36.202 TEST_HEADER include/spdk/fuse_dispatcher.h 00:07:36.202 TEST_HEADER include/spdk/gpt_spec.h 00:07:36.202 TEST_HEADER include/spdk/hexlify.h 00:07:36.202 TEST_HEADER include/spdk/histogram_data.h 00:07:36.202 TEST_HEADER include/spdk/idxd.h 00:07:36.202 TEST_HEADER include/spdk/idxd_spec.h 00:07:36.202 TEST_HEADER include/spdk/init.h 00:07:36.202 TEST_HEADER include/spdk/ioat_spec.h 00:07:36.202 TEST_HEADER include/spdk/ioat.h 00:07:36.202 TEST_HEADER include/spdk/iscsi_spec.h 00:07:36.202 TEST_HEADER include/spdk/json.h 00:07:36.202 TEST_HEADER include/spdk/jsonrpc.h 00:07:36.202 TEST_HEADER include/spdk/keyring_module.h 00:07:36.202 TEST_HEADER include/spdk/keyring.h 00:07:36.202 TEST_HEADER include/spdk/likely.h 00:07:36.202 TEST_HEADER include/spdk/log.h 00:07:36.202 TEST_HEADER include/spdk/lvol.h 00:07:36.202 TEST_HEADER include/spdk/md5.h 00:07:36.202 TEST_HEADER include/spdk/memory.h 00:07:36.202 TEST_HEADER include/spdk/mmio.h 00:07:36.202 TEST_HEADER include/spdk/net.h 00:07:36.202 TEST_HEADER include/spdk/nbd.h 00:07:36.202 TEST_HEADER include/spdk/notify.h 00:07:36.202 TEST_HEADER include/spdk/nvme.h 00:07:36.202 TEST_HEADER include/spdk/nvme_ocssd.h 00:07:36.202 TEST_HEADER include/spdk/nvme_intel.h 00:07:36.202 TEST_HEADER include/spdk/nvme_ocssd_spec.h 00:07:36.202 TEST_HEADER include/spdk/nvme_spec.h 00:07:36.202 TEST_HEADER include/spdk/nvme_zns.h 00:07:36.202 TEST_HEADER include/spdk/nvmf_cmd.h 00:07:36.202 TEST_HEADER include/spdk/nvmf_fc_spec.h 00:07:36.202 TEST_HEADER include/spdk/nvmf.h 00:07:36.202 TEST_HEADER include/spdk/nvmf_spec.h 00:07:36.202 TEST_HEADER include/spdk/nvmf_transport.h 00:07:36.202 TEST_HEADER include/spdk/opal.h 00:07:36.202 TEST_HEADER include/spdk/opal_spec.h 00:07:36.202 TEST_HEADER include/spdk/pci_ids.h 00:07:36.202 TEST_HEADER include/spdk/pipe.h 00:07:36.202 TEST_HEADER include/spdk/queue.h 00:07:36.202 TEST_HEADER include/spdk/reduce.h 00:07:36.202 TEST_HEADER include/spdk/scheduler.h 00:07:36.202 TEST_HEADER include/spdk/rpc.h 00:07:36.202 TEST_HEADER include/spdk/scsi.h 00:07:36.202 TEST_HEADER include/spdk/scsi_spec.h 00:07:36.202 TEST_HEADER include/spdk/sock.h 00:07:36.202 TEST_HEADER include/spdk/stdinc.h 00:07:36.202 TEST_HEADER include/spdk/string.h 00:07:36.202 TEST_HEADER include/spdk/thread.h 00:07:36.202 TEST_HEADER include/spdk/trace.h 00:07:36.202 TEST_HEADER include/spdk/trace_parser.h 00:07:36.202 TEST_HEADER include/spdk/tree.h 00:07:36.202 TEST_HEADER include/spdk/ublk.h 00:07:36.202 TEST_HEADER include/spdk/util.h 00:07:36.202 TEST_HEADER include/spdk/version.h 00:07:36.202 TEST_HEADER include/spdk/uuid.h 00:07:36.202 TEST_HEADER include/spdk/vfio_user_pci.h 00:07:36.202 TEST_HEADER include/spdk/vfio_user_spec.h 00:07:36.202 CC examples/interrupt_tgt/interrupt_tgt.o 00:07:36.202 TEST_HEADER include/spdk/vhost.h 00:07:36.202 TEST_HEADER include/spdk/vmd.h 00:07:36.202 TEST_HEADER include/spdk/xor.h 00:07:36.202 TEST_HEADER include/spdk/zipf.h 00:07:36.202 CXX test/cpp_headers/accel.o 00:07:36.202 CXX test/cpp_headers/accel_module.o 00:07:36.202 CXX test/cpp_headers/assert.o 00:07:36.202 CXX test/cpp_headers/barrier.o 00:07:36.202 CXX test/cpp_headers/base64.o 00:07:36.202 CXX test/cpp_headers/bdev.o 00:07:36.202 CXX test/cpp_headers/bdev_module.o 00:07:36.202 CXX test/cpp_headers/bdev_zone.o 00:07:36.202 CXX test/cpp_headers/bit_array.o 00:07:36.202 CXX test/cpp_headers/bit_pool.o 00:07:36.202 CXX test/cpp_headers/blob_bdev.o 00:07:36.202 CXX test/cpp_headers/blobfs_bdev.o 00:07:36.202 CXX test/cpp_headers/blobfs.o 00:07:36.202 CXX test/cpp_headers/blob.o 00:07:36.202 CXX test/cpp_headers/conf.o 00:07:36.202 CXX test/cpp_headers/config.o 00:07:36.202 CXX test/cpp_headers/cpuset.o 00:07:36.202 CC app/spdk_dd/spdk_dd.o 00:07:36.202 CXX test/cpp_headers/crc16.o 00:07:36.202 CC app/iscsi_tgt/iscsi_tgt.o 00:07:36.202 CC app/nvmf_tgt/nvmf_main.o 00:07:36.202 CXX test/cpp_headers/crc32.o 00:07:36.202 CC app/spdk_tgt/spdk_tgt.o 00:07:36.202 CC test/env/vtophys/vtophys.o 00:07:36.202 CC examples/ioat/verify/verify.o 00:07:36.202 CC examples/util/zipf/zipf.o 00:07:36.202 CC test/app/jsoncat/jsoncat.o 00:07:36.202 CC test/env/memory/memory_ut.o 00:07:36.202 CC test/app/histogram_perf/histogram_perf.o 00:07:36.202 CC examples/ioat/perf/perf.o 00:07:36.202 CC test/env/env_dpdk_post_init/env_dpdk_post_init.o 00:07:36.202 CC test/thread/poller_perf/poller_perf.o 00:07:36.202 CC test/env/pci/pci_ut.o 00:07:36.202 CC app/fio/nvme/fio_plugin.o 00:07:36.202 CC test/app/stub/stub.o 00:07:36.202 CC test/dma/test_dma/test_dma.o 00:07:36.202 CC test/app/bdev_svc/bdev_svc.o 00:07:36.202 CC app/fio/bdev/fio_plugin.o 00:07:36.465 CC test/env/mem_callbacks/mem_callbacks.o 00:07:36.465 CC test/app/fuzz/nvme_fuzz/nvme_fuzz.o 00:07:36.465 LINK spdk_lspci 00:07:36.465 LINK rpc_client_test 00:07:36.465 LINK spdk_nvme_discover 00:07:36.465 LINK jsoncat 00:07:36.465 LINK vtophys 00:07:36.465 LINK histogram_perf 00:07:36.465 CXX test/cpp_headers/crc64.o 00:07:36.465 LINK poller_perf 00:07:36.465 LINK interrupt_tgt 00:07:36.465 CXX test/cpp_headers/dif.o 00:07:36.465 LINK env_dpdk_post_init 00:07:36.465 CXX test/cpp_headers/dma.o 00:07:36.465 CXX test/cpp_headers/endian.o 00:07:36.732 LINK zipf 00:07:36.732 CXX test/cpp_headers/env_dpdk.o 00:07:36.732 CXX test/cpp_headers/env.o 00:07:36.732 CXX test/cpp_headers/event.o 00:07:36.732 CXX test/cpp_headers/fd_group.o 00:07:36.732 CXX test/cpp_headers/fd.o 00:07:36.732 LINK spdk_trace_record 00:07:36.732 LINK nvmf_tgt 00:07:36.732 CXX test/cpp_headers/file.o 00:07:36.732 CXX test/cpp_headers/fsdev.o 00:07:36.732 LINK iscsi_tgt 00:07:36.732 CXX test/cpp_headers/fsdev_module.o 00:07:36.732 LINK stub 00:07:36.732 CXX test/cpp_headers/ftl.o 00:07:36.732 CXX test/cpp_headers/fuse_dispatcher.o 00:07:36.732 CXX test/cpp_headers/gpt_spec.o 00:07:36.732 LINK verify 00:07:36.732 CXX test/cpp_headers/hexlify.o 00:07:36.732 CXX test/cpp_headers/histogram_data.o 00:07:36.732 LINK bdev_svc 00:07:36.732 LINK spdk_tgt 00:07:36.732 LINK ioat_perf 00:07:36.732 CC test/app/fuzz/vhost_fuzz/vhost_fuzz_rpc.o 00:07:36.732 CC test/app/fuzz/iscsi_fuzz/iscsi_fuzz.o 00:07:36.732 CC test/app/fuzz/vhost_fuzz/vhost_fuzz.o 00:07:36.732 CXX test/cpp_headers/idxd.o 00:07:37.000 CXX test/cpp_headers/idxd_spec.o 00:07:37.000 CXX test/cpp_headers/init.o 00:07:37.000 CXX test/cpp_headers/ioat.o 00:07:37.000 CXX test/cpp_headers/ioat_spec.o 00:07:37.000 CXX test/cpp_headers/iscsi_spec.o 00:07:37.000 CXX test/cpp_headers/json.o 00:07:37.000 CXX test/cpp_headers/jsonrpc.o 00:07:37.000 CXX test/cpp_headers/keyring.o 00:07:37.000 LINK spdk_dd 00:07:37.000 CXX test/cpp_headers/keyring_module.o 00:07:37.000 LINK spdk_trace 00:07:37.000 CXX test/cpp_headers/likely.o 00:07:37.000 CXX test/cpp_headers/log.o 00:07:37.000 CXX test/cpp_headers/lvol.o 00:07:37.000 CXX test/cpp_headers/md5.o 00:07:37.000 CXX test/cpp_headers/memory.o 00:07:37.000 CXX test/cpp_headers/mmio.o 00:07:37.000 CXX test/cpp_headers/nbd.o 00:07:37.000 CXX test/cpp_headers/net.o 00:07:37.000 CXX test/cpp_headers/notify.o 00:07:37.000 CXX test/cpp_headers/nvme.o 00:07:37.000 LINK pci_ut 00:07:37.000 CXX test/cpp_headers/nvme_intel.o 00:07:37.000 CXX test/cpp_headers/nvme_ocssd.o 00:07:37.000 CXX test/cpp_headers/nvme_ocssd_spec.o 00:07:37.000 CXX test/cpp_headers/nvme_spec.o 00:07:37.000 CXX test/cpp_headers/nvme_zns.o 00:07:37.000 CXX test/cpp_headers/nvmf_cmd.o 00:07:37.000 CXX test/cpp_headers/nvmf_fc_spec.o 00:07:37.270 CXX test/cpp_headers/nvmf.o 00:07:37.270 LINK nvme_fuzz 00:07:37.270 CXX test/cpp_headers/nvmf_spec.o 00:07:37.270 CC test/event/event_perf/event_perf.o 00:07:37.270 CXX test/cpp_headers/nvmf_transport.o 00:07:37.270 CC test/event/reactor/reactor.o 00:07:37.270 CXX test/cpp_headers/opal.o 00:07:37.270 CXX test/cpp_headers/opal_spec.o 00:07:37.270 CC test/event/reactor_perf/reactor_perf.o 00:07:37.270 LINK test_dma 00:07:37.270 CC examples/sock/hello_world/hello_sock.o 00:07:37.270 CXX test/cpp_headers/pci_ids.o 00:07:37.270 CXX test/cpp_headers/pipe.o 00:07:37.270 CC examples/vmd/lsvmd/lsvmd.o 00:07:37.270 CC examples/idxd/perf/perf.o 00:07:37.270 CC test/event/app_repeat/app_repeat.o 00:07:37.270 CXX test/cpp_headers/queue.o 00:07:37.270 CXX test/cpp_headers/reduce.o 00:07:37.270 CC examples/vmd/led/led.o 00:07:37.534 CC examples/thread/thread/thread_ex.o 00:07:37.534 CXX test/cpp_headers/rpc.o 00:07:37.534 CXX test/cpp_headers/scheduler.o 00:07:37.534 CXX test/cpp_headers/scsi.o 00:07:37.534 CXX test/cpp_headers/scsi_spec.o 00:07:37.534 CXX test/cpp_headers/sock.o 00:07:37.534 CXX test/cpp_headers/stdinc.o 00:07:37.534 CXX test/cpp_headers/string.o 00:07:37.534 CXX test/cpp_headers/thread.o 00:07:37.534 CC test/event/scheduler/scheduler.o 00:07:37.534 CXX test/cpp_headers/trace.o 00:07:37.534 LINK spdk_bdev 00:07:37.534 CXX test/cpp_headers/trace_parser.o 00:07:37.534 CXX test/cpp_headers/tree.o 00:07:37.534 CXX test/cpp_headers/ublk.o 00:07:37.534 CXX test/cpp_headers/util.o 00:07:37.534 CXX test/cpp_headers/uuid.o 00:07:37.534 LINK vhost_fuzz 00:07:37.534 CXX test/cpp_headers/version.o 00:07:37.534 CXX test/cpp_headers/vfio_user_pci.o 00:07:37.534 CXX test/cpp_headers/vfio_user_spec.o 00:07:37.534 CXX test/cpp_headers/vhost.o 00:07:37.534 CXX test/cpp_headers/vmd.o 00:07:37.534 LINK reactor 00:07:37.534 LINK event_perf 00:07:37.534 CXX test/cpp_headers/xor.o 00:07:37.534 LINK mem_callbacks 00:07:37.534 CC app/vhost/vhost.o 00:07:37.534 CXX test/cpp_headers/zipf.o 00:07:37.534 LINK reactor_perf 00:07:37.534 LINK spdk_nvme 00:07:37.534 LINK lsvmd 00:07:37.795 LINK spdk_nvme_perf 00:07:37.795 LINK app_repeat 00:07:37.795 LINK led 00:07:37.795 LINK spdk_nvme_identify 00:07:37.795 LINK spdk_top 00:07:37.795 LINK hello_sock 00:07:38.054 LINK thread 00:07:38.054 LINK scheduler 00:07:38.055 CC test/nvme/e2edp/nvme_dp.o 00:07:38.055 CC test/nvme/sgl/sgl.o 00:07:38.055 CC test/nvme/cuse/cuse.o 00:07:38.055 CC test/nvme/reset/reset.o 00:07:38.055 CC test/nvme/fused_ordering/fused_ordering.o 00:07:38.055 CC test/nvme/overhead/overhead.o 00:07:38.055 CC test/nvme/startup/startup.o 00:07:38.055 CC test/nvme/fdp/fdp.o 00:07:38.055 CC test/nvme/err_injection/err_injection.o 00:07:38.055 CC test/nvme/compliance/nvme_compliance.o 00:07:38.055 CC test/nvme/aer/aer.o 00:07:38.055 CC test/nvme/simple_copy/simple_copy.o 00:07:38.055 CC test/nvme/connect_stress/connect_stress.o 00:07:38.055 CC test/nvme/boot_partition/boot_partition.o 00:07:38.055 CC test/nvme/doorbell_aers/doorbell_aers.o 00:07:38.055 CC test/nvme/reserve/reserve.o 00:07:38.055 LINK vhost 00:07:38.055 CC test/accel/dif/dif.o 00:07:38.055 LINK idxd_perf 00:07:38.055 CC test/blobfs/mkfs/mkfs.o 00:07:38.055 CC test/lvol/esnap/esnap.o 00:07:38.314 LINK fused_ordering 00:07:38.314 LINK err_injection 00:07:38.314 LINK boot_partition 00:07:38.314 LINK reserve 00:07:38.314 CC examples/nvme/cmb_copy/cmb_copy.o 00:07:38.314 CC examples/nvme/hello_world/hello_world.o 00:07:38.314 CC examples/nvme/arbitration/arbitration.o 00:07:38.314 CC examples/nvme/nvme_manage/nvme_manage.o 00:07:38.314 CC examples/nvme/pmr_persistence/pmr_persistence.o 00:07:38.314 CC examples/nvme/reconnect/reconnect.o 00:07:38.314 CC examples/nvme/hotplug/hotplug.o 00:07:38.314 CC examples/nvme/abort/abort.o 00:07:38.314 LINK startup 00:07:38.314 LINK sgl 00:07:38.314 LINK aer 00:07:38.315 LINK nvme_dp 00:07:38.315 LINK connect_stress 00:07:38.315 LINK doorbell_aers 00:07:38.315 LINK mkfs 00:07:38.315 LINK overhead 00:07:38.315 LINK memory_ut 00:07:38.315 CC examples/accel/perf/accel_perf.o 00:07:38.315 LINK simple_copy 00:07:38.315 LINK nvme_compliance 00:07:38.574 LINK fdp 00:07:38.574 CC examples/blob/cli/blobcli.o 00:07:38.574 CC examples/fsdev/hello_world/hello_fsdev.o 00:07:38.574 CC examples/blob/hello_world/hello_blob.o 00:07:38.574 LINK reset 00:07:38.574 LINK pmr_persistence 00:07:38.574 LINK cmb_copy 00:07:38.574 LINK hello_world 00:07:38.833 LINK arbitration 00:07:38.833 LINK hotplug 00:07:38.833 LINK abort 00:07:38.833 LINK hello_blob 00:07:38.833 LINK dif 00:07:38.833 LINK hello_fsdev 00:07:38.833 LINK reconnect 00:07:38.833 LINK nvme_manage 00:07:39.092 LINK blobcli 00:07:39.092 LINK accel_perf 00:07:39.351 CC test/bdev/bdevio/bdevio.o 00:07:39.351 LINK iscsi_fuzz 00:07:39.351 CC examples/bdev/hello_world/hello_bdev.o 00:07:39.351 CC examples/bdev/bdevperf/bdevperf.o 00:07:39.609 LINK cuse 00:07:39.609 LINK hello_bdev 00:07:39.609 LINK bdevio 00:07:40.175 LINK bdevperf 00:07:40.742 CC examples/nvmf/nvmf/nvmf.o 00:07:41.000 LINK nvmf 00:07:43.528 LINK esnap 00:07:43.786 00:07:43.786 real 1m7.680s 00:07:43.786 user 9m3.784s 00:07:43.786 sys 1m58.630s 00:07:43.786 17:23:25 make -- common/autotest_common.sh@1130 -- $ xtrace_disable 00:07:43.786 17:23:25 make -- common/autotest_common.sh@10 -- $ set +x 00:07:43.786 ************************************ 00:07:43.786 END TEST make 00:07:43.786 ************************************ 00:07:43.786 17:23:25 -- spdk/autobuild.sh@1 -- $ stop_monitor_resources 00:07:43.786 17:23:25 -- pm/common@29 -- $ signal_monitor_resources TERM 00:07:43.786 17:23:25 -- pm/common@40 -- $ local monitor pid pids signal=TERM 00:07:43.786 17:23:25 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:07:43.786 17:23:25 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-cpu-load.pid ]] 00:07:43.786 17:23:25 -- pm/common@44 -- $ pid=16276 00:07:43.786 17:23:25 -- pm/common@50 -- $ kill -TERM 16276 00:07:43.786 17:23:25 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:07:43.786 17:23:25 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-vmstat.pid ]] 00:07:43.786 17:23:25 -- pm/common@44 -- $ pid=16278 00:07:43.786 17:23:25 -- pm/common@50 -- $ kill -TERM 16278 00:07:43.786 17:23:25 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:07:43.786 17:23:25 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-cpu-temp.pid ]] 00:07:43.786 17:23:25 -- pm/common@44 -- $ pid=16280 00:07:43.786 17:23:25 -- pm/common@50 -- $ kill -TERM 16280 00:07:43.786 17:23:25 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:07:43.786 17:23:25 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-bmc-pm.pid ]] 00:07:43.786 17:23:25 -- pm/common@44 -- $ pid=16309 00:07:43.786 17:23:25 -- pm/common@50 -- $ sudo -E kill -TERM 16309 00:07:43.786 17:23:25 -- spdk/autorun.sh@26 -- $ (( SPDK_TEST_UNITTEST == 1 || SPDK_RUN_FUNCTIONAL_TEST == 1 )) 00:07:43.786 17:23:25 -- spdk/autorun.sh@27 -- $ sudo -E /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/autotest.sh /var/jenkins/workspace/nvmf-tcp-phy-autotest/autorun-spdk.conf 00:07:43.786 17:23:25 -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:07:43.786 17:23:25 -- common/autotest_common.sh@1711 -- # lcov --version 00:07:43.786 17:23:25 -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:07:44.045 17:23:25 -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:07:44.045 17:23:25 -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:44.045 17:23:25 -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:44.045 17:23:25 -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:44.045 17:23:25 -- scripts/common.sh@336 -- # IFS=.-: 00:07:44.045 17:23:25 -- scripts/common.sh@336 -- # read -ra ver1 00:07:44.045 17:23:25 -- scripts/common.sh@337 -- # IFS=.-: 00:07:44.045 17:23:25 -- scripts/common.sh@337 -- # read -ra ver2 00:07:44.045 17:23:25 -- scripts/common.sh@338 -- # local 'op=<' 00:07:44.045 17:23:25 -- scripts/common.sh@340 -- # ver1_l=2 00:07:44.045 17:23:25 -- scripts/common.sh@341 -- # ver2_l=1 00:07:44.045 17:23:25 -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:44.045 17:23:25 -- scripts/common.sh@344 -- # case "$op" in 00:07:44.045 17:23:25 -- scripts/common.sh@345 -- # : 1 00:07:44.045 17:23:25 -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:44.045 17:23:25 -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:44.045 17:23:25 -- scripts/common.sh@365 -- # decimal 1 00:07:44.045 17:23:25 -- scripts/common.sh@353 -- # local d=1 00:07:44.045 17:23:25 -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:44.045 17:23:25 -- scripts/common.sh@355 -- # echo 1 00:07:44.045 17:23:25 -- scripts/common.sh@365 -- # ver1[v]=1 00:07:44.045 17:23:25 -- scripts/common.sh@366 -- # decimal 2 00:07:44.045 17:23:25 -- scripts/common.sh@353 -- # local d=2 00:07:44.045 17:23:25 -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:44.045 17:23:25 -- scripts/common.sh@355 -- # echo 2 00:07:44.045 17:23:25 -- scripts/common.sh@366 -- # ver2[v]=2 00:07:44.045 17:23:25 -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:44.045 17:23:25 -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:44.045 17:23:25 -- scripts/common.sh@368 -- # return 0 00:07:44.045 17:23:25 -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:44.045 17:23:25 -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:07:44.045 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:44.045 --rc genhtml_branch_coverage=1 00:07:44.045 --rc genhtml_function_coverage=1 00:07:44.045 --rc genhtml_legend=1 00:07:44.045 --rc geninfo_all_blocks=1 00:07:44.045 --rc geninfo_unexecuted_blocks=1 00:07:44.045 00:07:44.045 ' 00:07:44.046 17:23:25 -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:07:44.046 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:44.046 --rc genhtml_branch_coverage=1 00:07:44.046 --rc genhtml_function_coverage=1 00:07:44.046 --rc genhtml_legend=1 00:07:44.046 --rc geninfo_all_blocks=1 00:07:44.046 --rc geninfo_unexecuted_blocks=1 00:07:44.046 00:07:44.046 ' 00:07:44.046 17:23:25 -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:07:44.046 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:44.046 --rc genhtml_branch_coverage=1 00:07:44.046 --rc genhtml_function_coverage=1 00:07:44.046 --rc genhtml_legend=1 00:07:44.046 --rc geninfo_all_blocks=1 00:07:44.046 --rc geninfo_unexecuted_blocks=1 00:07:44.046 00:07:44.046 ' 00:07:44.046 17:23:25 -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:07:44.046 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:44.046 --rc genhtml_branch_coverage=1 00:07:44.046 --rc genhtml_function_coverage=1 00:07:44.046 --rc genhtml_legend=1 00:07:44.046 --rc geninfo_all_blocks=1 00:07:44.046 --rc geninfo_unexecuted_blocks=1 00:07:44.046 00:07:44.046 ' 00:07:44.046 17:23:25 -- spdk/autotest.sh@25 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:07:44.046 17:23:25 -- nvmf/common.sh@7 -- # uname -s 00:07:44.046 17:23:25 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:07:44.046 17:23:25 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:07:44.046 17:23:25 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:07:44.046 17:23:25 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:07:44.046 17:23:25 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:07:44.046 17:23:25 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:07:44.046 17:23:25 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:07:44.046 17:23:25 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:07:44.046 17:23:25 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:07:44.046 17:23:25 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:07:44.046 17:23:25 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:07:44.046 17:23:25 -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:07:44.046 17:23:25 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:07:44.046 17:23:25 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:07:44.046 17:23:25 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:07:44.046 17:23:25 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:07:44.046 17:23:25 -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:07:44.046 17:23:25 -- scripts/common.sh@15 -- # shopt -s extglob 00:07:44.046 17:23:25 -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:07:44.046 17:23:25 -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:07:44.046 17:23:25 -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:07:44.046 17:23:25 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:44.046 17:23:25 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:44.046 17:23:25 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:44.046 17:23:25 -- paths/export.sh@5 -- # export PATH 00:07:44.046 17:23:25 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:44.046 17:23:25 -- nvmf/common.sh@51 -- # : 0 00:07:44.046 17:23:25 -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:07:44.046 17:23:25 -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:07:44.046 17:23:25 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:07:44.046 17:23:25 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:07:44.046 17:23:25 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:07:44.046 17:23:25 -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:07:44.046 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:07:44.046 17:23:25 -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:07:44.046 17:23:25 -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:07:44.046 17:23:25 -- nvmf/common.sh@55 -- # have_pci_nics=0 00:07:44.046 17:23:25 -- spdk/autotest.sh@27 -- # '[' 0 -ne 0 ']' 00:07:44.046 17:23:25 -- spdk/autotest.sh@32 -- # uname -s 00:07:44.046 17:23:25 -- spdk/autotest.sh@32 -- # '[' Linux = Linux ']' 00:07:44.046 17:23:25 -- spdk/autotest.sh@33 -- # old_core_pattern='|/usr/lib/systemd/systemd-coredump %P %u %g %s %t %c %h' 00:07:44.046 17:23:25 -- spdk/autotest.sh@34 -- # mkdir -p /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/coredumps 00:07:44.046 17:23:25 -- spdk/autotest.sh@39 -- # echo '|/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/core-collector.sh %P %s %t' 00:07:44.046 17:23:25 -- spdk/autotest.sh@40 -- # echo /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/coredumps 00:07:44.046 17:23:25 -- spdk/autotest.sh@44 -- # modprobe nbd 00:07:44.046 17:23:25 -- spdk/autotest.sh@46 -- # type -P udevadm 00:07:44.046 17:23:25 -- spdk/autotest.sh@46 -- # udevadm=/usr/sbin/udevadm 00:07:44.046 17:23:25 -- spdk/autotest.sh@48 -- # udevadm_pid=97934 00:07:44.046 17:23:25 -- spdk/autotest.sh@47 -- # /usr/sbin/udevadm monitor --property 00:07:44.046 17:23:25 -- spdk/autotest.sh@53 -- # start_monitor_resources 00:07:44.046 17:23:25 -- pm/common@17 -- # local monitor 00:07:44.046 17:23:25 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:07:44.046 17:23:25 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:07:44.046 17:23:25 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:07:44.046 17:23:25 -- pm/common@21 -- # date +%s 00:07:44.046 17:23:25 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:07:44.046 17:23:25 -- pm/common@21 -- # date +%s 00:07:44.046 17:23:25 -- pm/common@25 -- # sleep 1 00:07:44.046 17:23:25 -- pm/common@21 -- # date +%s 00:07:44.046 17:23:25 -- pm/common@21 -- # date +%s 00:07:44.046 17:23:25 -- pm/common@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-load -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autotest.sh.1733502205 00:07:44.046 17:23:25 -- pm/common@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-vmstat -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autotest.sh.1733502205 00:07:44.046 17:23:25 -- pm/common@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-temp -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autotest.sh.1733502205 00:07:44.046 17:23:25 -- pm/common@21 -- # sudo -E /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-bmc-pm -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autotest.sh.1733502205 00:07:44.046 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autotest.sh.1733502205_collect-vmstat.pm.log 00:07:44.046 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autotest.sh.1733502205_collect-cpu-load.pm.log 00:07:44.046 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autotest.sh.1733502205_collect-cpu-temp.pm.log 00:07:44.046 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autotest.sh.1733502205_collect-bmc-pm.bmc.pm.log 00:07:44.982 17:23:26 -- spdk/autotest.sh@55 -- # trap 'autotest_cleanup || :; exit 1' SIGINT SIGTERM EXIT 00:07:44.982 17:23:26 -- spdk/autotest.sh@57 -- # timing_enter autotest 00:07:44.982 17:23:26 -- common/autotest_common.sh@726 -- # xtrace_disable 00:07:44.982 17:23:26 -- common/autotest_common.sh@10 -- # set +x 00:07:44.982 17:23:26 -- spdk/autotest.sh@59 -- # create_test_list 00:07:44.982 17:23:26 -- common/autotest_common.sh@752 -- # xtrace_disable 00:07:44.982 17:23:26 -- common/autotest_common.sh@10 -- # set +x 00:07:44.982 17:23:26 -- spdk/autotest.sh@61 -- # dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/autotest.sh 00:07:44.982 17:23:26 -- spdk/autotest.sh@61 -- # readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:07:44.982 17:23:26 -- spdk/autotest.sh@61 -- # src=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:07:44.982 17:23:26 -- spdk/autotest.sh@62 -- # out=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output 00:07:44.982 17:23:26 -- spdk/autotest.sh@63 -- # cd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:07:44.982 17:23:26 -- spdk/autotest.sh@65 -- # freebsd_update_contigmem_mod 00:07:44.982 17:23:26 -- common/autotest_common.sh@1457 -- # uname 00:07:44.982 17:23:26 -- common/autotest_common.sh@1457 -- # '[' Linux = FreeBSD ']' 00:07:44.982 17:23:26 -- spdk/autotest.sh@66 -- # freebsd_set_maxsock_buf 00:07:44.982 17:23:26 -- common/autotest_common.sh@1477 -- # uname 00:07:44.982 17:23:26 -- common/autotest_common.sh@1477 -- # [[ Linux = FreeBSD ]] 00:07:44.982 17:23:26 -- spdk/autotest.sh@68 -- # [[ y == y ]] 00:07:44.982 17:23:26 -- spdk/autotest.sh@70 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 --version 00:07:45.241 lcov: LCOV version 1.15 00:07:45.241 17:23:26 -- spdk/autotest.sh@72 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -i -t Baseline -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_base.info 00:08:07.181 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/nvme/nvme_stubs.gcno:no functions found 00:08:07.181 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/nvme/nvme_stubs.gcno 00:08:25.296 17:24:04 -- spdk/autotest.sh@76 -- # timing_enter pre_cleanup 00:08:25.296 17:24:04 -- common/autotest_common.sh@726 -- # xtrace_disable 00:08:25.296 17:24:04 -- common/autotest_common.sh@10 -- # set +x 00:08:25.296 17:24:04 -- spdk/autotest.sh@78 -- # rm -f 00:08:25.297 17:24:04 -- spdk/autotest.sh@81 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:08:25.297 0000:88:00.0 (8086 0a54): Already using the nvme driver 00:08:25.297 0000:00:04.7 (8086 0e27): Already using the ioatdma driver 00:08:25.297 0000:00:04.6 (8086 0e26): Already using the ioatdma driver 00:08:25.297 0000:00:04.5 (8086 0e25): Already using the ioatdma driver 00:08:25.297 0000:00:04.4 (8086 0e24): Already using the ioatdma driver 00:08:25.297 0000:00:04.3 (8086 0e23): Already using the ioatdma driver 00:08:25.297 0000:00:04.2 (8086 0e22): Already using the ioatdma driver 00:08:25.297 0000:00:04.1 (8086 0e21): Already using the ioatdma driver 00:08:25.297 0000:00:04.0 (8086 0e20): Already using the ioatdma driver 00:08:25.297 0000:80:04.7 (8086 0e27): Already using the ioatdma driver 00:08:25.297 0000:80:04.6 (8086 0e26): Already using the ioatdma driver 00:08:25.297 0000:80:04.5 (8086 0e25): Already using the ioatdma driver 00:08:25.297 0000:80:04.4 (8086 0e24): Already using the ioatdma driver 00:08:25.297 0000:80:04.3 (8086 0e23): Already using the ioatdma driver 00:08:25.297 0000:80:04.2 (8086 0e22): Already using the ioatdma driver 00:08:25.297 0000:80:04.1 (8086 0e21): Already using the ioatdma driver 00:08:25.297 0000:80:04.0 (8086 0e20): Already using the ioatdma driver 00:08:25.297 17:24:05 -- spdk/autotest.sh@83 -- # get_zoned_devs 00:08:25.297 17:24:05 -- common/autotest_common.sh@1657 -- # zoned_devs=() 00:08:25.297 17:24:05 -- common/autotest_common.sh@1657 -- # local -gA zoned_devs 00:08:25.297 17:24:05 -- common/autotest_common.sh@1658 -- # zoned_ctrls=() 00:08:25.297 17:24:05 -- common/autotest_common.sh@1658 -- # local -A zoned_ctrls 00:08:25.297 17:24:05 -- common/autotest_common.sh@1659 -- # local nvme bdf ns 00:08:25.297 17:24:05 -- common/autotest_common.sh@1668 -- # for nvme in /sys/class/nvme/nvme* 00:08:25.297 17:24:05 -- common/autotest_common.sh@1669 -- # bdf=0000:88:00.0 00:08:25.297 17:24:05 -- common/autotest_common.sh@1670 -- # for ns in "$nvme/"nvme*n* 00:08:25.297 17:24:05 -- common/autotest_common.sh@1671 -- # is_block_zoned nvme0n1 00:08:25.297 17:24:05 -- common/autotest_common.sh@1650 -- # local device=nvme0n1 00:08:25.297 17:24:05 -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:08:25.297 17:24:05 -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:08:25.297 17:24:05 -- spdk/autotest.sh@85 -- # (( 0 > 0 )) 00:08:25.297 17:24:05 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:08:25.297 17:24:05 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:08:25.297 17:24:05 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme0n1 00:08:25.297 17:24:05 -- scripts/common.sh@381 -- # local block=/dev/nvme0n1 pt 00:08:25.297 17:24:05 -- scripts/common.sh@390 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdk-gpt.py /dev/nvme0n1 00:08:25.297 No valid GPT data, bailing 00:08:25.297 17:24:05 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:08:25.297 17:24:05 -- scripts/common.sh@394 -- # pt= 00:08:25.297 17:24:05 -- scripts/common.sh@395 -- # return 1 00:08:25.297 17:24:05 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme0n1 bs=1M count=1 00:08:25.297 1+0 records in 00:08:25.297 1+0 records out 00:08:25.297 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00234456 s, 447 MB/s 00:08:25.297 17:24:05 -- spdk/autotest.sh@105 -- # sync 00:08:25.297 17:24:05 -- spdk/autotest.sh@107 -- # xtrace_disable_per_cmd reap_spdk_processes 00:08:25.297 17:24:05 -- common/autotest_common.sh@22 -- # eval 'reap_spdk_processes 12> /dev/null' 00:08:25.297 17:24:05 -- common/autotest_common.sh@22 -- # reap_spdk_processes 00:08:26.236 17:24:07 -- spdk/autotest.sh@111 -- # uname -s 00:08:26.236 17:24:07 -- spdk/autotest.sh@111 -- # [[ Linux == Linux ]] 00:08:26.236 17:24:07 -- spdk/autotest.sh@111 -- # [[ 0 -eq 1 ]] 00:08:26.236 17:24:07 -- spdk/autotest.sh@115 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh status 00:08:27.615 Hugepages 00:08:27.615 node hugesize free / total 00:08:27.615 node0 1048576kB 0 / 0 00:08:27.615 node0 2048kB 0 / 0 00:08:27.615 node1 1048576kB 0 / 0 00:08:27.615 node1 2048kB 0 / 0 00:08:27.615 00:08:27.615 Type BDF Vendor Device NUMA Driver Device Block devices 00:08:27.615 I/OAT 0000:00:04.0 8086 0e20 0 ioatdma - - 00:08:27.615 I/OAT 0000:00:04.1 8086 0e21 0 ioatdma - - 00:08:27.615 I/OAT 0000:00:04.2 8086 0e22 0 ioatdma - - 00:08:27.615 I/OAT 0000:00:04.3 8086 0e23 0 ioatdma - - 00:08:27.615 I/OAT 0000:00:04.4 8086 0e24 0 ioatdma - - 00:08:27.615 I/OAT 0000:00:04.5 8086 0e25 0 ioatdma - - 00:08:27.615 I/OAT 0000:00:04.6 8086 0e26 0 ioatdma - - 00:08:27.615 I/OAT 0000:00:04.7 8086 0e27 0 ioatdma - - 00:08:27.615 I/OAT 0000:80:04.0 8086 0e20 1 ioatdma - - 00:08:27.615 I/OAT 0000:80:04.1 8086 0e21 1 ioatdma - - 00:08:27.615 I/OAT 0000:80:04.2 8086 0e22 1 ioatdma - - 00:08:27.615 I/OAT 0000:80:04.3 8086 0e23 1 ioatdma - - 00:08:27.615 I/OAT 0000:80:04.4 8086 0e24 1 ioatdma - - 00:08:27.615 I/OAT 0000:80:04.5 8086 0e25 1 ioatdma - - 00:08:27.615 I/OAT 0000:80:04.6 8086 0e26 1 ioatdma - - 00:08:27.615 I/OAT 0000:80:04.7 8086 0e27 1 ioatdma - - 00:08:27.615 NVMe 0000:88:00.0 8086 0a54 1 nvme nvme0 nvme0n1 00:08:27.615 17:24:09 -- spdk/autotest.sh@117 -- # uname -s 00:08:27.615 17:24:09 -- spdk/autotest.sh@117 -- # [[ Linux == Linux ]] 00:08:27.615 17:24:09 -- spdk/autotest.sh@119 -- # nvme_namespace_revert 00:08:27.615 17:24:09 -- common/autotest_common.sh@1516 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:08:28.995 0000:00:04.7 (8086 0e27): ioatdma -> vfio-pci 00:08:28.995 0000:00:04.6 (8086 0e26): ioatdma -> vfio-pci 00:08:28.995 0000:00:04.5 (8086 0e25): ioatdma -> vfio-pci 00:08:28.995 0000:00:04.4 (8086 0e24): ioatdma -> vfio-pci 00:08:28.995 0000:00:04.3 (8086 0e23): ioatdma -> vfio-pci 00:08:28.995 0000:00:04.2 (8086 0e22): ioatdma -> vfio-pci 00:08:28.995 0000:00:04.1 (8086 0e21): ioatdma -> vfio-pci 00:08:28.995 0000:00:04.0 (8086 0e20): ioatdma -> vfio-pci 00:08:28.995 0000:80:04.7 (8086 0e27): ioatdma -> vfio-pci 00:08:28.995 0000:80:04.6 (8086 0e26): ioatdma -> vfio-pci 00:08:28.995 0000:80:04.5 (8086 0e25): ioatdma -> vfio-pci 00:08:28.995 0000:80:04.4 (8086 0e24): ioatdma -> vfio-pci 00:08:28.995 0000:80:04.3 (8086 0e23): ioatdma -> vfio-pci 00:08:28.995 0000:80:04.2 (8086 0e22): ioatdma -> vfio-pci 00:08:28.995 0000:80:04.1 (8086 0e21): ioatdma -> vfio-pci 00:08:28.995 0000:80:04.0 (8086 0e20): ioatdma -> vfio-pci 00:08:29.939 0000:88:00.0 (8086 0a54): nvme -> vfio-pci 00:08:29.939 17:24:11 -- common/autotest_common.sh@1517 -- # sleep 1 00:08:30.882 17:24:12 -- common/autotest_common.sh@1518 -- # bdfs=() 00:08:30.882 17:24:12 -- common/autotest_common.sh@1518 -- # local bdfs 00:08:30.882 17:24:12 -- common/autotest_common.sh@1520 -- # bdfs=($(get_nvme_bdfs)) 00:08:30.882 17:24:12 -- common/autotest_common.sh@1520 -- # get_nvme_bdfs 00:08:30.882 17:24:12 -- common/autotest_common.sh@1498 -- # bdfs=() 00:08:30.882 17:24:12 -- common/autotest_common.sh@1498 -- # local bdfs 00:08:30.882 17:24:12 -- common/autotest_common.sh@1499 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:08:30.882 17:24:12 -- common/autotest_common.sh@1499 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh 00:08:30.882 17:24:12 -- common/autotest_common.sh@1499 -- # jq -r '.config[].params.traddr' 00:08:31.140 17:24:12 -- common/autotest_common.sh@1500 -- # (( 1 == 0 )) 00:08:31.140 17:24:12 -- common/autotest_common.sh@1504 -- # printf '%s\n' 0000:88:00.0 00:08:31.140 17:24:12 -- common/autotest_common.sh@1522 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:08:32.080 Waiting for block devices as requested 00:08:32.080 0000:88:00.0 (8086 0a54): vfio-pci -> nvme 00:08:32.339 0000:00:04.7 (8086 0e27): vfio-pci -> ioatdma 00:08:32.339 0000:00:04.6 (8086 0e26): vfio-pci -> ioatdma 00:08:32.597 0000:00:04.5 (8086 0e25): vfio-pci -> ioatdma 00:08:32.597 0000:00:04.4 (8086 0e24): vfio-pci -> ioatdma 00:08:32.597 0000:00:04.3 (8086 0e23): vfio-pci -> ioatdma 00:08:32.597 0000:00:04.2 (8086 0e22): vfio-pci -> ioatdma 00:08:32.857 0000:00:04.1 (8086 0e21): vfio-pci -> ioatdma 00:08:32.857 0000:00:04.0 (8086 0e20): vfio-pci -> ioatdma 00:08:32.857 0000:80:04.7 (8086 0e27): vfio-pci -> ioatdma 00:08:32.857 0000:80:04.6 (8086 0e26): vfio-pci -> ioatdma 00:08:33.115 0000:80:04.5 (8086 0e25): vfio-pci -> ioatdma 00:08:33.115 0000:80:04.4 (8086 0e24): vfio-pci -> ioatdma 00:08:33.115 0000:80:04.3 (8086 0e23): vfio-pci -> ioatdma 00:08:33.372 0000:80:04.2 (8086 0e22): vfio-pci -> ioatdma 00:08:33.372 0000:80:04.1 (8086 0e21): vfio-pci -> ioatdma 00:08:33.372 0000:80:04.0 (8086 0e20): vfio-pci -> ioatdma 00:08:33.630 17:24:15 -- common/autotest_common.sh@1524 -- # for bdf in "${bdfs[@]}" 00:08:33.630 17:24:15 -- common/autotest_common.sh@1525 -- # get_nvme_ctrlr_from_bdf 0000:88:00.0 00:08:33.630 17:24:15 -- common/autotest_common.sh@1487 -- # readlink -f /sys/class/nvme/nvme0 00:08:33.630 17:24:15 -- common/autotest_common.sh@1487 -- # grep 0000:88:00.0/nvme/nvme 00:08:33.630 17:24:15 -- common/autotest_common.sh@1487 -- # bdf_sysfs_path=/sys/devices/pci0000:80/0000:80:03.0/0000:88:00.0/nvme/nvme0 00:08:33.631 17:24:15 -- common/autotest_common.sh@1488 -- # [[ -z /sys/devices/pci0000:80/0000:80:03.0/0000:88:00.0/nvme/nvme0 ]] 00:08:33.631 17:24:15 -- common/autotest_common.sh@1492 -- # basename /sys/devices/pci0000:80/0000:80:03.0/0000:88:00.0/nvme/nvme0 00:08:33.631 17:24:15 -- common/autotest_common.sh@1492 -- # printf '%s\n' nvme0 00:08:33.631 17:24:15 -- common/autotest_common.sh@1525 -- # nvme_ctrlr=/dev/nvme0 00:08:33.631 17:24:15 -- common/autotest_common.sh@1526 -- # [[ -z /dev/nvme0 ]] 00:08:33.631 17:24:15 -- common/autotest_common.sh@1531 -- # nvme id-ctrl /dev/nvme0 00:08:33.631 17:24:15 -- common/autotest_common.sh@1531 -- # grep oacs 00:08:33.631 17:24:15 -- common/autotest_common.sh@1531 -- # cut -d: -f2 00:08:33.631 17:24:15 -- common/autotest_common.sh@1531 -- # oacs=' 0xf' 00:08:33.631 17:24:15 -- common/autotest_common.sh@1532 -- # oacs_ns_manage=8 00:08:33.631 17:24:15 -- common/autotest_common.sh@1534 -- # [[ 8 -ne 0 ]] 00:08:33.631 17:24:15 -- common/autotest_common.sh@1540 -- # nvme id-ctrl /dev/nvme0 00:08:33.631 17:24:15 -- common/autotest_common.sh@1540 -- # grep unvmcap 00:08:33.631 17:24:15 -- common/autotest_common.sh@1540 -- # cut -d: -f2 00:08:33.631 17:24:15 -- common/autotest_common.sh@1540 -- # unvmcap=' 0' 00:08:33.631 17:24:15 -- common/autotest_common.sh@1541 -- # [[ 0 -eq 0 ]] 00:08:33.631 17:24:15 -- common/autotest_common.sh@1543 -- # continue 00:08:33.631 17:24:15 -- spdk/autotest.sh@122 -- # timing_exit pre_cleanup 00:08:33.631 17:24:15 -- common/autotest_common.sh@732 -- # xtrace_disable 00:08:33.631 17:24:15 -- common/autotest_common.sh@10 -- # set +x 00:08:33.631 17:24:15 -- spdk/autotest.sh@125 -- # timing_enter afterboot 00:08:33.631 17:24:15 -- common/autotest_common.sh@726 -- # xtrace_disable 00:08:33.631 17:24:15 -- common/autotest_common.sh@10 -- # set +x 00:08:33.631 17:24:15 -- spdk/autotest.sh@126 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:08:35.006 0000:00:04.7 (8086 0e27): ioatdma -> vfio-pci 00:08:35.006 0000:00:04.6 (8086 0e26): ioatdma -> vfio-pci 00:08:35.006 0000:00:04.5 (8086 0e25): ioatdma -> vfio-pci 00:08:35.006 0000:00:04.4 (8086 0e24): ioatdma -> vfio-pci 00:08:35.006 0000:00:04.3 (8086 0e23): ioatdma -> vfio-pci 00:08:35.006 0000:00:04.2 (8086 0e22): ioatdma -> vfio-pci 00:08:35.006 0000:00:04.1 (8086 0e21): ioatdma -> vfio-pci 00:08:35.006 0000:00:04.0 (8086 0e20): ioatdma -> vfio-pci 00:08:35.006 0000:80:04.7 (8086 0e27): ioatdma -> vfio-pci 00:08:35.006 0000:80:04.6 (8086 0e26): ioatdma -> vfio-pci 00:08:35.006 0000:80:04.5 (8086 0e25): ioatdma -> vfio-pci 00:08:35.006 0000:80:04.4 (8086 0e24): ioatdma -> vfio-pci 00:08:35.006 0000:80:04.3 (8086 0e23): ioatdma -> vfio-pci 00:08:35.006 0000:80:04.2 (8086 0e22): ioatdma -> vfio-pci 00:08:35.006 0000:80:04.1 (8086 0e21): ioatdma -> vfio-pci 00:08:35.006 0000:80:04.0 (8086 0e20): ioatdma -> vfio-pci 00:08:35.945 0000:88:00.0 (8086 0a54): nvme -> vfio-pci 00:08:35.945 17:24:17 -- spdk/autotest.sh@127 -- # timing_exit afterboot 00:08:35.945 17:24:17 -- common/autotest_common.sh@732 -- # xtrace_disable 00:08:35.945 17:24:17 -- common/autotest_common.sh@10 -- # set +x 00:08:35.945 17:24:17 -- spdk/autotest.sh@131 -- # opal_revert_cleanup 00:08:35.945 17:24:17 -- common/autotest_common.sh@1578 -- # mapfile -t bdfs 00:08:35.945 17:24:17 -- common/autotest_common.sh@1578 -- # get_nvme_bdfs_by_id 0x0a54 00:08:35.945 17:24:17 -- common/autotest_common.sh@1563 -- # bdfs=() 00:08:35.945 17:24:17 -- common/autotest_common.sh@1563 -- # _bdfs=() 00:08:35.945 17:24:17 -- common/autotest_common.sh@1563 -- # local bdfs _bdfs 00:08:35.945 17:24:17 -- common/autotest_common.sh@1564 -- # _bdfs=($(get_nvme_bdfs)) 00:08:35.945 17:24:17 -- common/autotest_common.sh@1564 -- # get_nvme_bdfs 00:08:35.945 17:24:17 -- common/autotest_common.sh@1498 -- # bdfs=() 00:08:35.945 17:24:17 -- common/autotest_common.sh@1498 -- # local bdfs 00:08:35.945 17:24:17 -- common/autotest_common.sh@1499 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:08:35.945 17:24:17 -- common/autotest_common.sh@1499 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh 00:08:35.945 17:24:17 -- common/autotest_common.sh@1499 -- # jq -r '.config[].params.traddr' 00:08:35.945 17:24:17 -- common/autotest_common.sh@1500 -- # (( 1 == 0 )) 00:08:35.945 17:24:17 -- common/autotest_common.sh@1504 -- # printf '%s\n' 0000:88:00.0 00:08:35.945 17:24:17 -- common/autotest_common.sh@1565 -- # for bdf in "${_bdfs[@]}" 00:08:35.945 17:24:17 -- common/autotest_common.sh@1566 -- # cat /sys/bus/pci/devices/0000:88:00.0/device 00:08:35.945 17:24:17 -- common/autotest_common.sh@1566 -- # device=0x0a54 00:08:35.945 17:24:17 -- common/autotest_common.sh@1567 -- # [[ 0x0a54 == \0\x\0\a\5\4 ]] 00:08:35.945 17:24:17 -- common/autotest_common.sh@1568 -- # bdfs+=($bdf) 00:08:35.945 17:24:17 -- common/autotest_common.sh@1572 -- # (( 1 > 0 )) 00:08:35.945 17:24:17 -- common/autotest_common.sh@1573 -- # printf '%s\n' 0000:88:00.0 00:08:35.945 17:24:17 -- common/autotest_common.sh@1579 -- # [[ -z 0000:88:00.0 ]] 00:08:35.945 17:24:17 -- common/autotest_common.sh@1584 -- # spdk_tgt_pid=109273 00:08:35.945 17:24:17 -- common/autotest_common.sh@1583 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:08:35.945 17:24:17 -- common/autotest_common.sh@1585 -- # waitforlisten 109273 00:08:35.945 17:24:17 -- common/autotest_common.sh@835 -- # '[' -z 109273 ']' 00:08:35.945 17:24:17 -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:35.945 17:24:17 -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:35.945 17:24:17 -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:35.945 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:35.945 17:24:17 -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:35.945 17:24:17 -- common/autotest_common.sh@10 -- # set +x 00:08:36.205 [2024-12-06 17:24:17.850856] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:08:36.205 [2024-12-06 17:24:17.850946] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid109273 ] 00:08:36.205 [2024-12-06 17:24:17.918869] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:36.205 [2024-12-06 17:24:17.961873] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:36.463 17:24:18 -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:36.463 17:24:18 -- common/autotest_common.sh@868 -- # return 0 00:08:36.463 17:24:18 -- common/autotest_common.sh@1587 -- # bdf_id=0 00:08:36.463 17:24:18 -- common/autotest_common.sh@1588 -- # for bdf in "${bdfs[@]}" 00:08:36.463 17:24:18 -- common/autotest_common.sh@1589 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_nvme_attach_controller -b nvme0 -t pcie -a 0000:88:00.0 00:08:39.754 nvme0n1 00:08:39.754 17:24:21 -- common/autotest_common.sh@1591 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_nvme_opal_revert -b nvme0 -p test 00:08:39.754 [2024-12-06 17:24:21.554548] nvme_opal.c:2063:spdk_opal_cmd_revert_tper: *ERROR*: Error on starting admin SP session with error 18 00:08:39.754 [2024-12-06 17:24:21.554593] vbdev_opal_rpc.c: 134:rpc_bdev_nvme_opal_revert: *ERROR*: Revert TPer failure: 18 00:08:39.754 request: 00:08:39.754 { 00:08:39.754 "nvme_ctrlr_name": "nvme0", 00:08:39.754 "password": "test", 00:08:39.754 "method": "bdev_nvme_opal_revert", 00:08:39.754 "req_id": 1 00:08:39.754 } 00:08:39.754 Got JSON-RPC error response 00:08:39.754 response: 00:08:39.754 { 00:08:39.754 "code": -32603, 00:08:39.754 "message": "Internal error" 00:08:39.754 } 00:08:39.754 17:24:21 -- common/autotest_common.sh@1591 -- # true 00:08:39.754 17:24:21 -- common/autotest_common.sh@1592 -- # (( ++bdf_id )) 00:08:39.754 17:24:21 -- common/autotest_common.sh@1595 -- # killprocess 109273 00:08:39.754 17:24:21 -- common/autotest_common.sh@954 -- # '[' -z 109273 ']' 00:08:39.754 17:24:21 -- common/autotest_common.sh@958 -- # kill -0 109273 00:08:39.754 17:24:21 -- common/autotest_common.sh@959 -- # uname 00:08:39.754 17:24:21 -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:39.754 17:24:21 -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 109273 00:08:40.013 17:24:21 -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:40.013 17:24:21 -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:40.013 17:24:21 -- common/autotest_common.sh@972 -- # echo 'killing process with pid 109273' 00:08:40.013 killing process with pid 109273 00:08:40.013 17:24:21 -- common/autotest_common.sh@973 -- # kill 109273 00:08:40.013 17:24:21 -- common/autotest_common.sh@978 -- # wait 109273 00:08:41.915 17:24:23 -- spdk/autotest.sh@137 -- # '[' 0 -eq 1 ']' 00:08:41.915 17:24:23 -- spdk/autotest.sh@141 -- # '[' 1 -eq 1 ']' 00:08:41.915 17:24:23 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:08:41.915 17:24:23 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:08:41.915 17:24:23 -- spdk/autotest.sh@149 -- # timing_enter lib 00:08:41.915 17:24:23 -- common/autotest_common.sh@726 -- # xtrace_disable 00:08:41.915 17:24:23 -- common/autotest_common.sh@10 -- # set +x 00:08:41.915 17:24:23 -- spdk/autotest.sh@151 -- # [[ 0 -eq 1 ]] 00:08:41.915 17:24:23 -- spdk/autotest.sh@155 -- # run_test env /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/env.sh 00:08:41.915 17:24:23 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:08:41.915 17:24:23 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:41.915 17:24:23 -- common/autotest_common.sh@10 -- # set +x 00:08:41.915 ************************************ 00:08:41.915 START TEST env 00:08:41.915 ************************************ 00:08:41.915 17:24:23 env -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/env.sh 00:08:41.915 * Looking for test storage... 00:08:41.915 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env 00:08:41.915 17:24:23 env -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:08:41.915 17:24:23 env -- common/autotest_common.sh@1711 -- # lcov --version 00:08:41.915 17:24:23 env -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:08:41.915 17:24:23 env -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:08:41.915 17:24:23 env -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:41.915 17:24:23 env -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:41.915 17:24:23 env -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:41.915 17:24:23 env -- scripts/common.sh@336 -- # IFS=.-: 00:08:41.915 17:24:23 env -- scripts/common.sh@336 -- # read -ra ver1 00:08:41.915 17:24:23 env -- scripts/common.sh@337 -- # IFS=.-: 00:08:41.915 17:24:23 env -- scripts/common.sh@337 -- # read -ra ver2 00:08:41.915 17:24:23 env -- scripts/common.sh@338 -- # local 'op=<' 00:08:41.915 17:24:23 env -- scripts/common.sh@340 -- # ver1_l=2 00:08:41.915 17:24:23 env -- scripts/common.sh@341 -- # ver2_l=1 00:08:41.915 17:24:23 env -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:41.915 17:24:23 env -- scripts/common.sh@344 -- # case "$op" in 00:08:41.915 17:24:23 env -- scripts/common.sh@345 -- # : 1 00:08:41.915 17:24:23 env -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:41.915 17:24:23 env -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:41.915 17:24:23 env -- scripts/common.sh@365 -- # decimal 1 00:08:41.915 17:24:23 env -- scripts/common.sh@353 -- # local d=1 00:08:41.915 17:24:23 env -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:41.915 17:24:23 env -- scripts/common.sh@355 -- # echo 1 00:08:41.915 17:24:23 env -- scripts/common.sh@365 -- # ver1[v]=1 00:08:41.915 17:24:23 env -- scripts/common.sh@366 -- # decimal 2 00:08:41.915 17:24:23 env -- scripts/common.sh@353 -- # local d=2 00:08:41.915 17:24:23 env -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:41.915 17:24:23 env -- scripts/common.sh@355 -- # echo 2 00:08:41.915 17:24:23 env -- scripts/common.sh@366 -- # ver2[v]=2 00:08:41.915 17:24:23 env -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:41.915 17:24:23 env -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:41.915 17:24:23 env -- scripts/common.sh@368 -- # return 0 00:08:41.915 17:24:23 env -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:41.915 17:24:23 env -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:08:41.916 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:41.916 --rc genhtml_branch_coverage=1 00:08:41.916 --rc genhtml_function_coverage=1 00:08:41.916 --rc genhtml_legend=1 00:08:41.916 --rc geninfo_all_blocks=1 00:08:41.916 --rc geninfo_unexecuted_blocks=1 00:08:41.916 00:08:41.916 ' 00:08:41.916 17:24:23 env -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:08:41.916 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:41.916 --rc genhtml_branch_coverage=1 00:08:41.916 --rc genhtml_function_coverage=1 00:08:41.916 --rc genhtml_legend=1 00:08:41.916 --rc geninfo_all_blocks=1 00:08:41.916 --rc geninfo_unexecuted_blocks=1 00:08:41.916 00:08:41.916 ' 00:08:41.916 17:24:23 env -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:08:41.916 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:41.916 --rc genhtml_branch_coverage=1 00:08:41.916 --rc genhtml_function_coverage=1 00:08:41.916 --rc genhtml_legend=1 00:08:41.916 --rc geninfo_all_blocks=1 00:08:41.916 --rc geninfo_unexecuted_blocks=1 00:08:41.916 00:08:41.916 ' 00:08:41.916 17:24:23 env -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:08:41.916 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:41.916 --rc genhtml_branch_coverage=1 00:08:41.916 --rc genhtml_function_coverage=1 00:08:41.916 --rc genhtml_legend=1 00:08:41.916 --rc geninfo_all_blocks=1 00:08:41.916 --rc geninfo_unexecuted_blocks=1 00:08:41.916 00:08:41.916 ' 00:08:41.916 17:24:23 env -- env/env.sh@10 -- # run_test env_memory /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/memory/memory_ut 00:08:41.916 17:24:23 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:08:41.916 17:24:23 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:41.916 17:24:23 env -- common/autotest_common.sh@10 -- # set +x 00:08:41.916 ************************************ 00:08:41.916 START TEST env_memory 00:08:41.916 ************************************ 00:08:41.916 17:24:23 env.env_memory -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/memory/memory_ut 00:08:41.916 00:08:41.916 00:08:41.916 CUnit - A unit testing framework for C - Version 2.1-3 00:08:41.916 http://cunit.sourceforge.net/ 00:08:41.916 00:08:41.916 00:08:41.916 Suite: memory 00:08:41.916 Test: alloc and free memory map ...[2024-12-06 17:24:23.583612] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 283:spdk_mem_map_alloc: *ERROR*: Initial mem_map notify failed 00:08:41.916 passed 00:08:41.916 Test: mem map translation ...[2024-12-06 17:24:23.604935] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=2097152 len=1234 00:08:41.916 [2024-12-06 17:24:23.604971] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=1234 len=2097152 00:08:41.916 [2024-12-06 17:24:23.605029] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 589:spdk_mem_map_set_translation: *ERROR*: invalid usermode virtual address 281474976710656 00:08:41.916 [2024-12-06 17:24:23.605041] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 605:spdk_mem_map_set_translation: *ERROR*: could not get 0xffffffe00000 map 00:08:41.916 passed 00:08:41.916 Test: mem map registration ...[2024-12-06 17:24:23.647826] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=200000 len=1234 00:08:41.916 [2024-12-06 17:24:23.647844] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=4d2 len=2097152 00:08:41.916 passed 00:08:41.916 Test: mem map adjacent registrations ...passed 00:08:41.916 00:08:41.916 Run Summary: Type Total Ran Passed Failed Inactive 00:08:41.916 suites 1 1 n/a 0 0 00:08:41.916 tests 4 4 4 0 0 00:08:41.916 asserts 152 152 152 0 n/a 00:08:41.916 00:08:41.916 Elapsed time = 0.146 seconds 00:08:41.916 00:08:41.916 real 0m0.157s 00:08:41.916 user 0m0.148s 00:08:41.916 sys 0m0.008s 00:08:41.916 17:24:23 env.env_memory -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:41.916 17:24:23 env.env_memory -- common/autotest_common.sh@10 -- # set +x 00:08:41.916 ************************************ 00:08:41.916 END TEST env_memory 00:08:41.916 ************************************ 00:08:41.916 17:24:23 env -- env/env.sh@11 -- # run_test env_vtophys /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/vtophys/vtophys 00:08:41.916 17:24:23 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:08:41.916 17:24:23 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:41.916 17:24:23 env -- common/autotest_common.sh@10 -- # set +x 00:08:41.916 ************************************ 00:08:41.916 START TEST env_vtophys 00:08:41.916 ************************************ 00:08:41.916 17:24:23 env.env_vtophys -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/vtophys/vtophys 00:08:42.176 EAL: lib.eal log level changed from notice to debug 00:08:42.176 EAL: Detected lcore 0 as core 0 on socket 0 00:08:42.176 EAL: Detected lcore 1 as core 1 on socket 0 00:08:42.176 EAL: Detected lcore 2 as core 2 on socket 0 00:08:42.176 EAL: Detected lcore 3 as core 3 on socket 0 00:08:42.176 EAL: Detected lcore 4 as core 4 on socket 0 00:08:42.176 EAL: Detected lcore 5 as core 5 on socket 0 00:08:42.176 EAL: Detected lcore 6 as core 8 on socket 0 00:08:42.176 EAL: Detected lcore 7 as core 9 on socket 0 00:08:42.176 EAL: Detected lcore 8 as core 10 on socket 0 00:08:42.176 EAL: Detected lcore 9 as core 11 on socket 0 00:08:42.176 EAL: Detected lcore 10 as core 12 on socket 0 00:08:42.176 EAL: Detected lcore 11 as core 13 on socket 0 00:08:42.176 EAL: Detected lcore 12 as core 0 on socket 1 00:08:42.176 EAL: Detected lcore 13 as core 1 on socket 1 00:08:42.176 EAL: Detected lcore 14 as core 2 on socket 1 00:08:42.176 EAL: Detected lcore 15 as core 3 on socket 1 00:08:42.176 EAL: Detected lcore 16 as core 4 on socket 1 00:08:42.176 EAL: Detected lcore 17 as core 5 on socket 1 00:08:42.176 EAL: Detected lcore 18 as core 8 on socket 1 00:08:42.176 EAL: Detected lcore 19 as core 9 on socket 1 00:08:42.176 EAL: Detected lcore 20 as core 10 on socket 1 00:08:42.176 EAL: Detected lcore 21 as core 11 on socket 1 00:08:42.176 EAL: Detected lcore 22 as core 12 on socket 1 00:08:42.176 EAL: Detected lcore 23 as core 13 on socket 1 00:08:42.176 EAL: Detected lcore 24 as core 0 on socket 0 00:08:42.176 EAL: Detected lcore 25 as core 1 on socket 0 00:08:42.176 EAL: Detected lcore 26 as core 2 on socket 0 00:08:42.176 EAL: Detected lcore 27 as core 3 on socket 0 00:08:42.176 EAL: Detected lcore 28 as core 4 on socket 0 00:08:42.176 EAL: Detected lcore 29 as core 5 on socket 0 00:08:42.176 EAL: Detected lcore 30 as core 8 on socket 0 00:08:42.176 EAL: Detected lcore 31 as core 9 on socket 0 00:08:42.176 EAL: Detected lcore 32 as core 10 on socket 0 00:08:42.176 EAL: Detected lcore 33 as core 11 on socket 0 00:08:42.176 EAL: Detected lcore 34 as core 12 on socket 0 00:08:42.176 EAL: Detected lcore 35 as core 13 on socket 0 00:08:42.176 EAL: Detected lcore 36 as core 0 on socket 1 00:08:42.176 EAL: Detected lcore 37 as core 1 on socket 1 00:08:42.176 EAL: Detected lcore 38 as core 2 on socket 1 00:08:42.176 EAL: Detected lcore 39 as core 3 on socket 1 00:08:42.176 EAL: Detected lcore 40 as core 4 on socket 1 00:08:42.176 EAL: Detected lcore 41 as core 5 on socket 1 00:08:42.176 EAL: Detected lcore 42 as core 8 on socket 1 00:08:42.176 EAL: Detected lcore 43 as core 9 on socket 1 00:08:42.176 EAL: Detected lcore 44 as core 10 on socket 1 00:08:42.176 EAL: Detected lcore 45 as core 11 on socket 1 00:08:42.176 EAL: Detected lcore 46 as core 12 on socket 1 00:08:42.176 EAL: Detected lcore 47 as core 13 on socket 1 00:08:42.176 EAL: Maximum logical cores by configuration: 128 00:08:42.176 EAL: Detected CPU lcores: 48 00:08:42.176 EAL: Detected NUMA nodes: 2 00:08:42.176 EAL: Checking presence of .so 'librte_eal.so.24.0' 00:08:42.176 EAL: Detected shared linkage of DPDK 00:08:42.176 EAL: open shared lib /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/dpdk/pmds-24.0/librte_bus_pci.so.24.0 00:08:42.176 EAL: open shared lib /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/dpdk/pmds-24.0/librte_bus_vdev.so.24.0 00:08:42.176 EAL: Registered [vdev] bus. 00:08:42.176 EAL: bus.vdev log level changed from disabled to notice 00:08:42.176 EAL: open shared lib /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/dpdk/pmds-24.0/librte_mempool_ring.so.24.0 00:08:42.176 EAL: open shared lib /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/dpdk/pmds-24.0/librte_net_i40e.so.24.0 00:08:42.176 EAL: pmd.net.i40e.init log level changed from disabled to notice 00:08:42.176 EAL: pmd.net.i40e.driver log level changed from disabled to notice 00:08:42.176 EAL: open shared lib /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/dpdk/pmds-24.0/librte_bus_pci.so 00:08:42.176 EAL: open shared lib /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/dpdk/pmds-24.0/librte_bus_vdev.so 00:08:42.176 EAL: open shared lib /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/dpdk/pmds-24.0/librte_mempool_ring.so 00:08:42.176 EAL: open shared lib /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib/dpdk/pmds-24.0/librte_net_i40e.so 00:08:42.176 EAL: No shared files mode enabled, IPC will be disabled 00:08:42.176 EAL: No shared files mode enabled, IPC is disabled 00:08:42.176 EAL: Bus pci wants IOVA as 'DC' 00:08:42.176 EAL: Bus vdev wants IOVA as 'DC' 00:08:42.176 EAL: Buses did not request a specific IOVA mode. 00:08:42.176 EAL: IOMMU is available, selecting IOVA as VA mode. 00:08:42.176 EAL: Selected IOVA mode 'VA' 00:08:42.176 EAL: Probing VFIO support... 00:08:42.176 EAL: IOMMU type 1 (Type 1) is supported 00:08:42.176 EAL: IOMMU type 7 (sPAPR) is not supported 00:08:42.176 EAL: IOMMU type 8 (No-IOMMU) is not supported 00:08:42.176 EAL: VFIO support initialized 00:08:42.176 EAL: Ask a virtual area of 0x2e000 bytes 00:08:42.176 EAL: Virtual area found at 0x200000000000 (size = 0x2e000) 00:08:42.176 EAL: Setting up physically contiguous memory... 00:08:42.176 EAL: Setting maximum number of open files to 524288 00:08:42.176 EAL: Detected memory type: socket_id:0 hugepage_sz:2097152 00:08:42.176 EAL: Detected memory type: socket_id:1 hugepage_sz:2097152 00:08:42.176 EAL: Creating 4 segment lists: n_segs:8192 socket_id:0 hugepage_sz:2097152 00:08:42.176 EAL: Ask a virtual area of 0x61000 bytes 00:08:42.176 EAL: Virtual area found at 0x20000002e000 (size = 0x61000) 00:08:42.176 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:08:42.176 EAL: Ask a virtual area of 0x400000000 bytes 00:08:42.176 EAL: Virtual area found at 0x200000200000 (size = 0x400000000) 00:08:42.176 EAL: VA reserved for memseg list at 0x200000200000, size 400000000 00:08:42.176 EAL: Ask a virtual area of 0x61000 bytes 00:08:42.176 EAL: Virtual area found at 0x200400200000 (size = 0x61000) 00:08:42.176 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:08:42.176 EAL: Ask a virtual area of 0x400000000 bytes 00:08:42.176 EAL: Virtual area found at 0x200400400000 (size = 0x400000000) 00:08:42.176 EAL: VA reserved for memseg list at 0x200400400000, size 400000000 00:08:42.176 EAL: Ask a virtual area of 0x61000 bytes 00:08:42.176 EAL: Virtual area found at 0x200800400000 (size = 0x61000) 00:08:42.176 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:08:42.176 EAL: Ask a virtual area of 0x400000000 bytes 00:08:42.176 EAL: Virtual area found at 0x200800600000 (size = 0x400000000) 00:08:42.177 EAL: VA reserved for memseg list at 0x200800600000, size 400000000 00:08:42.177 EAL: Ask a virtual area of 0x61000 bytes 00:08:42.177 EAL: Virtual area found at 0x200c00600000 (size = 0x61000) 00:08:42.177 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:08:42.177 EAL: Ask a virtual area of 0x400000000 bytes 00:08:42.177 EAL: Virtual area found at 0x200c00800000 (size = 0x400000000) 00:08:42.177 EAL: VA reserved for memseg list at 0x200c00800000, size 400000000 00:08:42.177 EAL: Creating 4 segment lists: n_segs:8192 socket_id:1 hugepage_sz:2097152 00:08:42.177 EAL: Ask a virtual area of 0x61000 bytes 00:08:42.177 EAL: Virtual area found at 0x201000800000 (size = 0x61000) 00:08:42.177 EAL: Memseg list allocated at socket 1, page size 0x800kB 00:08:42.177 EAL: Ask a virtual area of 0x400000000 bytes 00:08:42.177 EAL: Virtual area found at 0x201000a00000 (size = 0x400000000) 00:08:42.177 EAL: VA reserved for memseg list at 0x201000a00000, size 400000000 00:08:42.177 EAL: Ask a virtual area of 0x61000 bytes 00:08:42.177 EAL: Virtual area found at 0x201400a00000 (size = 0x61000) 00:08:42.177 EAL: Memseg list allocated at socket 1, page size 0x800kB 00:08:42.177 EAL: Ask a virtual area of 0x400000000 bytes 00:08:42.177 EAL: Virtual area found at 0x201400c00000 (size = 0x400000000) 00:08:42.177 EAL: VA reserved for memseg list at 0x201400c00000, size 400000000 00:08:42.177 EAL: Ask a virtual area of 0x61000 bytes 00:08:42.177 EAL: Virtual area found at 0x201800c00000 (size = 0x61000) 00:08:42.177 EAL: Memseg list allocated at socket 1, page size 0x800kB 00:08:42.177 EAL: Ask a virtual area of 0x400000000 bytes 00:08:42.177 EAL: Virtual area found at 0x201800e00000 (size = 0x400000000) 00:08:42.177 EAL: VA reserved for memseg list at 0x201800e00000, size 400000000 00:08:42.177 EAL: Ask a virtual area of 0x61000 bytes 00:08:42.177 EAL: Virtual area found at 0x201c00e00000 (size = 0x61000) 00:08:42.177 EAL: Memseg list allocated at socket 1, page size 0x800kB 00:08:42.177 EAL: Ask a virtual area of 0x400000000 bytes 00:08:42.177 EAL: Virtual area found at 0x201c01000000 (size = 0x400000000) 00:08:42.177 EAL: VA reserved for memseg list at 0x201c01000000, size 400000000 00:08:42.177 EAL: Hugepages will be freed exactly as allocated. 00:08:42.177 EAL: No shared files mode enabled, IPC is disabled 00:08:42.177 EAL: No shared files mode enabled, IPC is disabled 00:08:42.177 EAL: TSC frequency is ~2700000 KHz 00:08:42.177 EAL: Main lcore 0 is ready (tid=7f2bb4cc6a00;cpuset=[0]) 00:08:42.177 EAL: Trying to obtain current memory policy. 00:08:42.177 EAL: Setting policy MPOL_PREFERRED for socket 0 00:08:42.177 EAL: Restoring previous memory policy: 0 00:08:42.177 EAL: request: mp_malloc_sync 00:08:42.177 EAL: No shared files mode enabled, IPC is disabled 00:08:42.177 EAL: Heap on socket 0 was expanded by 2MB 00:08:42.177 EAL: No shared files mode enabled, IPC is disabled 00:08:42.177 EAL: No shared files mode enabled, IPC is disabled 00:08:42.177 EAL: No PCI address specified using 'addr=' in: bus=pci 00:08:42.177 EAL: Mem event callback 'spdk:(nil)' registered 00:08:42.177 00:08:42.177 00:08:42.177 CUnit - A unit testing framework for C - Version 2.1-3 00:08:42.177 http://cunit.sourceforge.net/ 00:08:42.177 00:08:42.177 00:08:42.177 Suite: components_suite 00:08:42.177 Test: vtophys_malloc_test ...passed 00:08:42.177 Test: vtophys_spdk_malloc_test ...EAL: Trying to obtain current memory policy. 00:08:42.177 EAL: Setting policy MPOL_PREFERRED for socket 0 00:08:42.177 EAL: Restoring previous memory policy: 4 00:08:42.177 EAL: Calling mem event callback 'spdk:(nil)' 00:08:42.177 EAL: request: mp_malloc_sync 00:08:42.177 EAL: No shared files mode enabled, IPC is disabled 00:08:42.177 EAL: Heap on socket 0 was expanded by 4MB 00:08:42.177 EAL: Calling mem event callback 'spdk:(nil)' 00:08:42.177 EAL: request: mp_malloc_sync 00:08:42.177 EAL: No shared files mode enabled, IPC is disabled 00:08:42.177 EAL: Heap on socket 0 was shrunk by 4MB 00:08:42.177 EAL: Trying to obtain current memory policy. 00:08:42.177 EAL: Setting policy MPOL_PREFERRED for socket 0 00:08:42.177 EAL: Restoring previous memory policy: 4 00:08:42.177 EAL: Calling mem event callback 'spdk:(nil)' 00:08:42.177 EAL: request: mp_malloc_sync 00:08:42.177 EAL: No shared files mode enabled, IPC is disabled 00:08:42.177 EAL: Heap on socket 0 was expanded by 6MB 00:08:42.177 EAL: Calling mem event callback 'spdk:(nil)' 00:08:42.177 EAL: request: mp_malloc_sync 00:08:42.177 EAL: No shared files mode enabled, IPC is disabled 00:08:42.177 EAL: Heap on socket 0 was shrunk by 6MB 00:08:42.177 EAL: Trying to obtain current memory policy. 00:08:42.177 EAL: Setting policy MPOL_PREFERRED for socket 0 00:08:42.177 EAL: Restoring previous memory policy: 4 00:08:42.177 EAL: Calling mem event callback 'spdk:(nil)' 00:08:42.177 EAL: request: mp_malloc_sync 00:08:42.177 EAL: No shared files mode enabled, IPC is disabled 00:08:42.177 EAL: Heap on socket 0 was expanded by 10MB 00:08:42.177 EAL: Calling mem event callback 'spdk:(nil)' 00:08:42.177 EAL: request: mp_malloc_sync 00:08:42.177 EAL: No shared files mode enabled, IPC is disabled 00:08:42.177 EAL: Heap on socket 0 was shrunk by 10MB 00:08:42.177 EAL: Trying to obtain current memory policy. 00:08:42.177 EAL: Setting policy MPOL_PREFERRED for socket 0 00:08:42.177 EAL: Restoring previous memory policy: 4 00:08:42.177 EAL: Calling mem event callback 'spdk:(nil)' 00:08:42.177 EAL: request: mp_malloc_sync 00:08:42.177 EAL: No shared files mode enabled, IPC is disabled 00:08:42.177 EAL: Heap on socket 0 was expanded by 18MB 00:08:42.177 EAL: Calling mem event callback 'spdk:(nil)' 00:08:42.177 EAL: request: mp_malloc_sync 00:08:42.177 EAL: No shared files mode enabled, IPC is disabled 00:08:42.177 EAL: Heap on socket 0 was shrunk by 18MB 00:08:42.177 EAL: Trying to obtain current memory policy. 00:08:42.177 EAL: Setting policy MPOL_PREFERRED for socket 0 00:08:42.177 EAL: Restoring previous memory policy: 4 00:08:42.177 EAL: Calling mem event callback 'spdk:(nil)' 00:08:42.177 EAL: request: mp_malloc_sync 00:08:42.177 EAL: No shared files mode enabled, IPC is disabled 00:08:42.177 EAL: Heap on socket 0 was expanded by 34MB 00:08:42.177 EAL: Calling mem event callback 'spdk:(nil)' 00:08:42.177 EAL: request: mp_malloc_sync 00:08:42.177 EAL: No shared files mode enabled, IPC is disabled 00:08:42.177 EAL: Heap on socket 0 was shrunk by 34MB 00:08:42.177 EAL: Trying to obtain current memory policy. 00:08:42.177 EAL: Setting policy MPOL_PREFERRED for socket 0 00:08:42.177 EAL: Restoring previous memory policy: 4 00:08:42.177 EAL: Calling mem event callback 'spdk:(nil)' 00:08:42.177 EAL: request: mp_malloc_sync 00:08:42.177 EAL: No shared files mode enabled, IPC is disabled 00:08:42.177 EAL: Heap on socket 0 was expanded by 66MB 00:08:42.177 EAL: Calling mem event callback 'spdk:(nil)' 00:08:42.177 EAL: request: mp_malloc_sync 00:08:42.177 EAL: No shared files mode enabled, IPC is disabled 00:08:42.177 EAL: Heap on socket 0 was shrunk by 66MB 00:08:42.177 EAL: Trying to obtain current memory policy. 00:08:42.177 EAL: Setting policy MPOL_PREFERRED for socket 0 00:08:42.177 EAL: Restoring previous memory policy: 4 00:08:42.177 EAL: Calling mem event callback 'spdk:(nil)' 00:08:42.177 EAL: request: mp_malloc_sync 00:08:42.177 EAL: No shared files mode enabled, IPC is disabled 00:08:42.177 EAL: Heap on socket 0 was expanded by 130MB 00:08:42.177 EAL: Calling mem event callback 'spdk:(nil)' 00:08:42.177 EAL: request: mp_malloc_sync 00:08:42.177 EAL: No shared files mode enabled, IPC is disabled 00:08:42.177 EAL: Heap on socket 0 was shrunk by 130MB 00:08:42.177 EAL: Trying to obtain current memory policy. 00:08:42.177 EAL: Setting policy MPOL_PREFERRED for socket 0 00:08:42.437 EAL: Restoring previous memory policy: 4 00:08:42.437 EAL: Calling mem event callback 'spdk:(nil)' 00:08:42.437 EAL: request: mp_malloc_sync 00:08:42.437 EAL: No shared files mode enabled, IPC is disabled 00:08:42.437 EAL: Heap on socket 0 was expanded by 258MB 00:08:42.437 EAL: Calling mem event callback 'spdk:(nil)' 00:08:42.437 EAL: request: mp_malloc_sync 00:08:42.437 EAL: No shared files mode enabled, IPC is disabled 00:08:42.437 EAL: Heap on socket 0 was shrunk by 258MB 00:08:42.437 EAL: Trying to obtain current memory policy. 00:08:42.437 EAL: Setting policy MPOL_PREFERRED for socket 0 00:08:42.697 EAL: Restoring previous memory policy: 4 00:08:42.697 EAL: Calling mem event callback 'spdk:(nil)' 00:08:42.697 EAL: request: mp_malloc_sync 00:08:42.697 EAL: No shared files mode enabled, IPC is disabled 00:08:42.697 EAL: Heap on socket 0 was expanded by 514MB 00:08:42.697 EAL: Calling mem event callback 'spdk:(nil)' 00:08:42.697 EAL: request: mp_malloc_sync 00:08:42.697 EAL: No shared files mode enabled, IPC is disabled 00:08:42.697 EAL: Heap on socket 0 was shrunk by 514MB 00:08:42.697 EAL: Trying to obtain current memory policy. 00:08:42.697 EAL: Setting policy MPOL_PREFERRED for socket 0 00:08:42.955 EAL: Restoring previous memory policy: 4 00:08:42.955 EAL: Calling mem event callback 'spdk:(nil)' 00:08:42.955 EAL: request: mp_malloc_sync 00:08:42.955 EAL: No shared files mode enabled, IPC is disabled 00:08:42.955 EAL: Heap on socket 0 was expanded by 1026MB 00:08:43.215 EAL: Calling mem event callback 'spdk:(nil)' 00:08:43.475 EAL: request: mp_malloc_sync 00:08:43.475 EAL: No shared files mode enabled, IPC is disabled 00:08:43.475 EAL: Heap on socket 0 was shrunk by 1026MB 00:08:43.475 passed 00:08:43.475 00:08:43.475 Run Summary: Type Total Ran Passed Failed Inactive 00:08:43.475 suites 1 1 n/a 0 0 00:08:43.475 tests 2 2 2 0 0 00:08:43.475 asserts 497 497 497 0 n/a 00:08:43.475 00:08:43.475 Elapsed time = 1.334 seconds 00:08:43.475 EAL: Calling mem event callback 'spdk:(nil)' 00:08:43.475 EAL: request: mp_malloc_sync 00:08:43.475 EAL: No shared files mode enabled, IPC is disabled 00:08:43.475 EAL: Heap on socket 0 was shrunk by 2MB 00:08:43.475 EAL: No shared files mode enabled, IPC is disabled 00:08:43.475 EAL: No shared files mode enabled, IPC is disabled 00:08:43.475 EAL: No shared files mode enabled, IPC is disabled 00:08:43.475 00:08:43.475 real 0m1.449s 00:08:43.475 user 0m0.856s 00:08:43.475 sys 0m0.563s 00:08:43.475 17:24:25 env.env_vtophys -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:43.475 17:24:25 env.env_vtophys -- common/autotest_common.sh@10 -- # set +x 00:08:43.475 ************************************ 00:08:43.475 END TEST env_vtophys 00:08:43.475 ************************************ 00:08:43.475 17:24:25 env -- env/env.sh@12 -- # run_test env_pci /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/pci/pci_ut 00:08:43.475 17:24:25 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:08:43.475 17:24:25 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:43.475 17:24:25 env -- common/autotest_common.sh@10 -- # set +x 00:08:43.475 ************************************ 00:08:43.475 START TEST env_pci 00:08:43.475 ************************************ 00:08:43.475 17:24:25 env.env_pci -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/pci/pci_ut 00:08:43.475 00:08:43.475 00:08:43.475 CUnit - A unit testing framework for C - Version 2.1-3 00:08:43.475 http://cunit.sourceforge.net/ 00:08:43.475 00:08:43.475 00:08:43.475 Suite: pci 00:08:43.475 Test: pci_hook ...[2024-12-06 17:24:25.261241] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/pci.c:1117:spdk_pci_device_claim: *ERROR*: Cannot create lock on device /var/tmp/spdk_pci_lock_10000:00:01.0, probably process 110168 has claimed it 00:08:43.475 EAL: Cannot find device (10000:00:01.0) 00:08:43.475 EAL: Failed to attach device on primary process 00:08:43.475 passed 00:08:43.475 00:08:43.475 Run Summary: Type Total Ran Passed Failed Inactive 00:08:43.475 suites 1 1 n/a 0 0 00:08:43.475 tests 1 1 1 0 0 00:08:43.475 asserts 25 25 25 0 n/a 00:08:43.475 00:08:43.475 Elapsed time = 0.021 seconds 00:08:43.475 00:08:43.475 real 0m0.035s 00:08:43.475 user 0m0.011s 00:08:43.475 sys 0m0.024s 00:08:43.475 17:24:25 env.env_pci -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:43.475 17:24:25 env.env_pci -- common/autotest_common.sh@10 -- # set +x 00:08:43.475 ************************************ 00:08:43.475 END TEST env_pci 00:08:43.475 ************************************ 00:08:43.475 17:24:25 env -- env/env.sh@14 -- # argv='-c 0x1 ' 00:08:43.475 17:24:25 env -- env/env.sh@15 -- # uname 00:08:43.475 17:24:25 env -- env/env.sh@15 -- # '[' Linux = Linux ']' 00:08:43.475 17:24:25 env -- env/env.sh@22 -- # argv+=--base-virtaddr=0x200000000000 00:08:43.475 17:24:25 env -- env/env.sh@24 -- # run_test env_dpdk_post_init /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:08:43.475 17:24:25 env -- common/autotest_common.sh@1105 -- # '[' 5 -le 1 ']' 00:08:43.475 17:24:25 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:43.475 17:24:25 env -- common/autotest_common.sh@10 -- # set +x 00:08:43.735 ************************************ 00:08:43.735 START TEST env_dpdk_post_init 00:08:43.735 ************************************ 00:08:43.735 17:24:25 env.env_dpdk_post_init -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:08:43.735 EAL: Detected CPU lcores: 48 00:08:43.735 EAL: Detected NUMA nodes: 2 00:08:43.735 EAL: Detected shared linkage of DPDK 00:08:43.735 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:08:43.735 EAL: Selected IOVA mode 'VA' 00:08:43.735 EAL: VFIO support initialized 00:08:43.735 TELEMETRY: No legacy callbacks, legacy socket not created 00:08:43.735 EAL: Using IOMMU type 1 (Type 1) 00:08:43.735 EAL: Probe PCI driver: spdk_ioat (8086:0e20) device: 0000:00:04.0 (socket 0) 00:08:43.735 EAL: Probe PCI driver: spdk_ioat (8086:0e21) device: 0000:00:04.1 (socket 0) 00:08:43.735 EAL: Probe PCI driver: spdk_ioat (8086:0e22) device: 0000:00:04.2 (socket 0) 00:08:43.735 EAL: Probe PCI driver: spdk_ioat (8086:0e23) device: 0000:00:04.3 (socket 0) 00:08:43.735 EAL: Probe PCI driver: spdk_ioat (8086:0e24) device: 0000:00:04.4 (socket 0) 00:08:43.735 EAL: Probe PCI driver: spdk_ioat (8086:0e25) device: 0000:00:04.5 (socket 0) 00:08:43.735 EAL: Probe PCI driver: spdk_ioat (8086:0e26) device: 0000:00:04.6 (socket 0) 00:08:43.735 EAL: Probe PCI driver: spdk_ioat (8086:0e27) device: 0000:00:04.7 (socket 0) 00:08:43.735 EAL: Probe PCI driver: spdk_ioat (8086:0e20) device: 0000:80:04.0 (socket 1) 00:08:43.735 EAL: Probe PCI driver: spdk_ioat (8086:0e21) device: 0000:80:04.1 (socket 1) 00:08:43.995 EAL: Probe PCI driver: spdk_ioat (8086:0e22) device: 0000:80:04.2 (socket 1) 00:08:43.995 EAL: Probe PCI driver: spdk_ioat (8086:0e23) device: 0000:80:04.3 (socket 1) 00:08:43.995 EAL: Probe PCI driver: spdk_ioat (8086:0e24) device: 0000:80:04.4 (socket 1) 00:08:43.995 EAL: Probe PCI driver: spdk_ioat (8086:0e25) device: 0000:80:04.5 (socket 1) 00:08:43.995 EAL: Probe PCI driver: spdk_ioat (8086:0e26) device: 0000:80:04.6 (socket 1) 00:08:43.995 EAL: Probe PCI driver: spdk_ioat (8086:0e27) device: 0000:80:04.7 (socket 1) 00:08:44.567 EAL: Probe PCI driver: spdk_nvme (8086:0a54) device: 0000:88:00.0 (socket 1) 00:08:47.855 EAL: Releasing PCI mapped resource for 0000:88:00.0 00:08:47.855 EAL: Calling pci_unmap_resource for 0000:88:00.0 at 0x202001040000 00:08:48.114 Starting DPDK initialization... 00:08:48.114 Starting SPDK post initialization... 00:08:48.114 SPDK NVMe probe 00:08:48.114 Attaching to 0000:88:00.0 00:08:48.114 Attached to 0000:88:00.0 00:08:48.114 Cleaning up... 00:08:48.114 00:08:48.114 real 0m4.409s 00:08:48.114 user 0m3.265s 00:08:48.114 sys 0m0.201s 00:08:48.114 17:24:29 env.env_dpdk_post_init -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:48.114 17:24:29 env.env_dpdk_post_init -- common/autotest_common.sh@10 -- # set +x 00:08:48.114 ************************************ 00:08:48.114 END TEST env_dpdk_post_init 00:08:48.114 ************************************ 00:08:48.114 17:24:29 env -- env/env.sh@26 -- # uname 00:08:48.114 17:24:29 env -- env/env.sh@26 -- # '[' Linux = Linux ']' 00:08:48.114 17:24:29 env -- env/env.sh@29 -- # run_test env_mem_callbacks /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/mem_callbacks/mem_callbacks 00:08:48.114 17:24:29 env -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:08:48.114 17:24:29 env -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:48.114 17:24:29 env -- common/autotest_common.sh@10 -- # set +x 00:08:48.114 ************************************ 00:08:48.114 START TEST env_mem_callbacks 00:08:48.114 ************************************ 00:08:48.114 17:24:29 env.env_mem_callbacks -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/mem_callbacks/mem_callbacks 00:08:48.114 EAL: Detected CPU lcores: 48 00:08:48.114 EAL: Detected NUMA nodes: 2 00:08:48.114 EAL: Detected shared linkage of DPDK 00:08:48.114 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:08:48.114 EAL: Selected IOVA mode 'VA' 00:08:48.114 EAL: VFIO support initialized 00:08:48.114 TELEMETRY: No legacy callbacks, legacy socket not created 00:08:48.114 00:08:48.114 00:08:48.114 CUnit - A unit testing framework for C - Version 2.1-3 00:08:48.114 http://cunit.sourceforge.net/ 00:08:48.114 00:08:48.114 00:08:48.114 Suite: memory 00:08:48.114 Test: test ... 00:08:48.114 register 0x200000200000 2097152 00:08:48.114 malloc 3145728 00:08:48.114 register 0x200000400000 4194304 00:08:48.114 buf 0x200000500000 len 3145728 PASSED 00:08:48.114 malloc 64 00:08:48.114 buf 0x2000004fff40 len 64 PASSED 00:08:48.114 malloc 4194304 00:08:48.114 register 0x200000800000 6291456 00:08:48.114 buf 0x200000a00000 len 4194304 PASSED 00:08:48.114 free 0x200000500000 3145728 00:08:48.114 free 0x2000004fff40 64 00:08:48.114 unregister 0x200000400000 4194304 PASSED 00:08:48.114 free 0x200000a00000 4194304 00:08:48.114 unregister 0x200000800000 6291456 PASSED 00:08:48.114 malloc 8388608 00:08:48.114 register 0x200000400000 10485760 00:08:48.114 buf 0x200000600000 len 8388608 PASSED 00:08:48.114 free 0x200000600000 8388608 00:08:48.114 unregister 0x200000400000 10485760 PASSED 00:08:48.114 passed 00:08:48.114 00:08:48.114 Run Summary: Type Total Ran Passed Failed Inactive 00:08:48.114 suites 1 1 n/a 0 0 00:08:48.114 tests 1 1 1 0 0 00:08:48.114 asserts 15 15 15 0 n/a 00:08:48.114 00:08:48.114 Elapsed time = 0.005 seconds 00:08:48.114 00:08:48.114 real 0m0.049s 00:08:48.114 user 0m0.016s 00:08:48.114 sys 0m0.033s 00:08:48.114 17:24:29 env.env_mem_callbacks -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:48.114 17:24:29 env.env_mem_callbacks -- common/autotest_common.sh@10 -- # set +x 00:08:48.114 ************************************ 00:08:48.114 END TEST env_mem_callbacks 00:08:48.114 ************************************ 00:08:48.114 00:08:48.114 real 0m6.504s 00:08:48.114 user 0m4.483s 00:08:48.114 sys 0m1.069s 00:08:48.114 17:24:29 env -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:48.114 17:24:29 env -- common/autotest_common.sh@10 -- # set +x 00:08:48.114 ************************************ 00:08:48.114 END TEST env 00:08:48.114 ************************************ 00:08:48.114 17:24:29 -- spdk/autotest.sh@156 -- # run_test rpc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/rpc.sh 00:08:48.114 17:24:29 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:08:48.114 17:24:29 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:48.114 17:24:29 -- common/autotest_common.sh@10 -- # set +x 00:08:48.114 ************************************ 00:08:48.114 START TEST rpc 00:08:48.114 ************************************ 00:08:48.114 17:24:29 rpc -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/rpc.sh 00:08:48.373 * Looking for test storage... 00:08:48.373 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc 00:08:48.373 17:24:29 rpc -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:08:48.373 17:24:29 rpc -- common/autotest_common.sh@1711 -- # lcov --version 00:08:48.373 17:24:29 rpc -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:08:48.373 17:24:30 rpc -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:08:48.373 17:24:30 rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:48.373 17:24:30 rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:48.373 17:24:30 rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:48.373 17:24:30 rpc -- scripts/common.sh@336 -- # IFS=.-: 00:08:48.373 17:24:30 rpc -- scripts/common.sh@336 -- # read -ra ver1 00:08:48.373 17:24:30 rpc -- scripts/common.sh@337 -- # IFS=.-: 00:08:48.373 17:24:30 rpc -- scripts/common.sh@337 -- # read -ra ver2 00:08:48.373 17:24:30 rpc -- scripts/common.sh@338 -- # local 'op=<' 00:08:48.373 17:24:30 rpc -- scripts/common.sh@340 -- # ver1_l=2 00:08:48.374 17:24:30 rpc -- scripts/common.sh@341 -- # ver2_l=1 00:08:48.374 17:24:30 rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:48.374 17:24:30 rpc -- scripts/common.sh@344 -- # case "$op" in 00:08:48.374 17:24:30 rpc -- scripts/common.sh@345 -- # : 1 00:08:48.374 17:24:30 rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:48.374 17:24:30 rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:48.374 17:24:30 rpc -- scripts/common.sh@365 -- # decimal 1 00:08:48.374 17:24:30 rpc -- scripts/common.sh@353 -- # local d=1 00:08:48.374 17:24:30 rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:48.374 17:24:30 rpc -- scripts/common.sh@355 -- # echo 1 00:08:48.374 17:24:30 rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:08:48.374 17:24:30 rpc -- scripts/common.sh@366 -- # decimal 2 00:08:48.374 17:24:30 rpc -- scripts/common.sh@353 -- # local d=2 00:08:48.374 17:24:30 rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:48.374 17:24:30 rpc -- scripts/common.sh@355 -- # echo 2 00:08:48.374 17:24:30 rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:08:48.374 17:24:30 rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:48.374 17:24:30 rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:48.374 17:24:30 rpc -- scripts/common.sh@368 -- # return 0 00:08:48.374 17:24:30 rpc -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:48.374 17:24:30 rpc -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:08:48.374 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:48.374 --rc genhtml_branch_coverage=1 00:08:48.374 --rc genhtml_function_coverage=1 00:08:48.374 --rc genhtml_legend=1 00:08:48.374 --rc geninfo_all_blocks=1 00:08:48.374 --rc geninfo_unexecuted_blocks=1 00:08:48.374 00:08:48.374 ' 00:08:48.374 17:24:30 rpc -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:08:48.374 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:48.374 --rc genhtml_branch_coverage=1 00:08:48.374 --rc genhtml_function_coverage=1 00:08:48.374 --rc genhtml_legend=1 00:08:48.374 --rc geninfo_all_blocks=1 00:08:48.374 --rc geninfo_unexecuted_blocks=1 00:08:48.374 00:08:48.374 ' 00:08:48.374 17:24:30 rpc -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:08:48.374 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:48.374 --rc genhtml_branch_coverage=1 00:08:48.374 --rc genhtml_function_coverage=1 00:08:48.374 --rc genhtml_legend=1 00:08:48.374 --rc geninfo_all_blocks=1 00:08:48.374 --rc geninfo_unexecuted_blocks=1 00:08:48.374 00:08:48.374 ' 00:08:48.374 17:24:30 rpc -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:08:48.374 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:48.374 --rc genhtml_branch_coverage=1 00:08:48.374 --rc genhtml_function_coverage=1 00:08:48.374 --rc genhtml_legend=1 00:08:48.374 --rc geninfo_all_blocks=1 00:08:48.374 --rc geninfo_unexecuted_blocks=1 00:08:48.374 00:08:48.374 ' 00:08:48.374 17:24:30 rpc -- rpc/rpc.sh@65 -- # spdk_pid=110959 00:08:48.374 17:24:30 rpc -- rpc/rpc.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -e bdev 00:08:48.374 17:24:30 rpc -- rpc/rpc.sh@66 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:08:48.374 17:24:30 rpc -- rpc/rpc.sh@67 -- # waitforlisten 110959 00:08:48.374 17:24:30 rpc -- common/autotest_common.sh@835 -- # '[' -z 110959 ']' 00:08:48.374 17:24:30 rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:48.374 17:24:30 rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:48.374 17:24:30 rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:48.374 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:48.374 17:24:30 rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:48.374 17:24:30 rpc -- common/autotest_common.sh@10 -- # set +x 00:08:48.374 [2024-12-06 17:24:30.120682] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:08:48.374 [2024-12-06 17:24:30.120796] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid110959 ] 00:08:48.374 [2024-12-06 17:24:30.188466] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:48.633 [2024-12-06 17:24:30.236550] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask bdev specified. 00:08:48.633 [2024-12-06 17:24:30.236606] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s spdk_tgt -p 110959' to capture a snapshot of events at runtime. 00:08:48.633 [2024-12-06 17:24:30.236635] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:08:48.633 [2024-12-06 17:24:30.236646] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:08:48.633 [2024-12-06 17:24:30.236656] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/spdk_tgt_trace.pid110959 for offline analysis/debug. 00:08:48.633 [2024-12-06 17:24:30.237261] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:48.893 17:24:30 rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:48.893 17:24:30 rpc -- common/autotest_common.sh@868 -- # return 0 00:08:48.893 17:24:30 rpc -- rpc/rpc.sh@69 -- # export PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc 00:08:48.893 17:24:30 rpc -- rpc/rpc.sh@69 -- # PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc 00:08:48.893 17:24:30 rpc -- rpc/rpc.sh@72 -- # rpc=rpc_cmd 00:08:48.893 17:24:30 rpc -- rpc/rpc.sh@73 -- # run_test rpc_integrity rpc_integrity 00:08:48.893 17:24:30 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:08:48.893 17:24:30 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:48.893 17:24:30 rpc -- common/autotest_common.sh@10 -- # set +x 00:08:48.893 ************************************ 00:08:48.893 START TEST rpc_integrity 00:08:48.893 ************************************ 00:08:48.893 17:24:30 rpc.rpc_integrity -- common/autotest_common.sh@1129 -- # rpc_integrity 00:08:48.893 17:24:30 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:08:48.893 17:24:30 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.893 17:24:30 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:08:48.893 17:24:30 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.893 17:24:30 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:08:48.893 17:24:30 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # jq length 00:08:48.893 17:24:30 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:08:48.893 17:24:30 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:08:48.893 17:24:30 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.893 17:24:30 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:08:48.893 17:24:30 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.893 17:24:30 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc0 00:08:48.893 17:24:30 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:08:48.893 17:24:30 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.893 17:24:30 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:08:48.893 17:24:30 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.893 17:24:30 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:08:48.893 { 00:08:48.893 "name": "Malloc0", 00:08:48.893 "aliases": [ 00:08:48.893 "5f43285e-b279-4f15-8322-ff41079eba1d" 00:08:48.893 ], 00:08:48.893 "product_name": "Malloc disk", 00:08:48.893 "block_size": 512, 00:08:48.893 "num_blocks": 16384, 00:08:48.893 "uuid": "5f43285e-b279-4f15-8322-ff41079eba1d", 00:08:48.893 "assigned_rate_limits": { 00:08:48.893 "rw_ios_per_sec": 0, 00:08:48.893 "rw_mbytes_per_sec": 0, 00:08:48.893 "r_mbytes_per_sec": 0, 00:08:48.893 "w_mbytes_per_sec": 0 00:08:48.893 }, 00:08:48.893 "claimed": false, 00:08:48.893 "zoned": false, 00:08:48.893 "supported_io_types": { 00:08:48.893 "read": true, 00:08:48.893 "write": true, 00:08:48.893 "unmap": true, 00:08:48.893 "flush": true, 00:08:48.893 "reset": true, 00:08:48.893 "nvme_admin": false, 00:08:48.893 "nvme_io": false, 00:08:48.893 "nvme_io_md": false, 00:08:48.893 "write_zeroes": true, 00:08:48.893 "zcopy": true, 00:08:48.893 "get_zone_info": false, 00:08:48.893 "zone_management": false, 00:08:48.893 "zone_append": false, 00:08:48.893 "compare": false, 00:08:48.893 "compare_and_write": false, 00:08:48.893 "abort": true, 00:08:48.893 "seek_hole": false, 00:08:48.893 "seek_data": false, 00:08:48.893 "copy": true, 00:08:48.893 "nvme_iov_md": false 00:08:48.893 }, 00:08:48.893 "memory_domains": [ 00:08:48.893 { 00:08:48.893 "dma_device_id": "system", 00:08:48.893 "dma_device_type": 1 00:08:48.893 }, 00:08:48.893 { 00:08:48.893 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:48.893 "dma_device_type": 2 00:08:48.893 } 00:08:48.893 ], 00:08:48.893 "driver_specific": {} 00:08:48.893 } 00:08:48.893 ]' 00:08:48.893 17:24:30 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # jq length 00:08:48.893 17:24:30 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:08:48.893 17:24:30 rpc.rpc_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc0 -p Passthru0 00:08:48.893 17:24:30 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.893 17:24:30 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:08:48.893 [2024-12-06 17:24:30.648553] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on Malloc0 00:08:48.893 [2024-12-06 17:24:30.648591] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:48.893 [2024-12-06 17:24:30.648629] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x2261520 00:08:48.893 [2024-12-06 17:24:30.648642] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:48.893 [2024-12-06 17:24:30.650066] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:48.893 [2024-12-06 17:24:30.650088] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:08:48.893 Passthru0 00:08:48.893 17:24:30 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.893 17:24:30 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:08:48.893 17:24:30 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.893 17:24:30 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:08:48.893 17:24:30 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.893 17:24:30 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:08:48.893 { 00:08:48.893 "name": "Malloc0", 00:08:48.893 "aliases": [ 00:08:48.893 "5f43285e-b279-4f15-8322-ff41079eba1d" 00:08:48.893 ], 00:08:48.893 "product_name": "Malloc disk", 00:08:48.893 "block_size": 512, 00:08:48.893 "num_blocks": 16384, 00:08:48.893 "uuid": "5f43285e-b279-4f15-8322-ff41079eba1d", 00:08:48.893 "assigned_rate_limits": { 00:08:48.893 "rw_ios_per_sec": 0, 00:08:48.893 "rw_mbytes_per_sec": 0, 00:08:48.893 "r_mbytes_per_sec": 0, 00:08:48.893 "w_mbytes_per_sec": 0 00:08:48.893 }, 00:08:48.893 "claimed": true, 00:08:48.893 "claim_type": "exclusive_write", 00:08:48.893 "zoned": false, 00:08:48.893 "supported_io_types": { 00:08:48.893 "read": true, 00:08:48.893 "write": true, 00:08:48.893 "unmap": true, 00:08:48.893 "flush": true, 00:08:48.893 "reset": true, 00:08:48.893 "nvme_admin": false, 00:08:48.893 "nvme_io": false, 00:08:48.893 "nvme_io_md": false, 00:08:48.893 "write_zeroes": true, 00:08:48.893 "zcopy": true, 00:08:48.893 "get_zone_info": false, 00:08:48.893 "zone_management": false, 00:08:48.893 "zone_append": false, 00:08:48.893 "compare": false, 00:08:48.893 "compare_and_write": false, 00:08:48.893 "abort": true, 00:08:48.893 "seek_hole": false, 00:08:48.893 "seek_data": false, 00:08:48.893 "copy": true, 00:08:48.893 "nvme_iov_md": false 00:08:48.893 }, 00:08:48.893 "memory_domains": [ 00:08:48.893 { 00:08:48.893 "dma_device_id": "system", 00:08:48.893 "dma_device_type": 1 00:08:48.893 }, 00:08:48.893 { 00:08:48.893 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:48.893 "dma_device_type": 2 00:08:48.893 } 00:08:48.893 ], 00:08:48.893 "driver_specific": {} 00:08:48.893 }, 00:08:48.893 { 00:08:48.893 "name": "Passthru0", 00:08:48.893 "aliases": [ 00:08:48.893 "7fdd2c99-b59f-59c9-8111-e8c5cf64ee60" 00:08:48.893 ], 00:08:48.893 "product_name": "passthru", 00:08:48.893 "block_size": 512, 00:08:48.893 "num_blocks": 16384, 00:08:48.893 "uuid": "7fdd2c99-b59f-59c9-8111-e8c5cf64ee60", 00:08:48.893 "assigned_rate_limits": { 00:08:48.893 "rw_ios_per_sec": 0, 00:08:48.893 "rw_mbytes_per_sec": 0, 00:08:48.893 "r_mbytes_per_sec": 0, 00:08:48.893 "w_mbytes_per_sec": 0 00:08:48.893 }, 00:08:48.893 "claimed": false, 00:08:48.893 "zoned": false, 00:08:48.893 "supported_io_types": { 00:08:48.893 "read": true, 00:08:48.893 "write": true, 00:08:48.894 "unmap": true, 00:08:48.894 "flush": true, 00:08:48.894 "reset": true, 00:08:48.894 "nvme_admin": false, 00:08:48.894 "nvme_io": false, 00:08:48.894 "nvme_io_md": false, 00:08:48.894 "write_zeroes": true, 00:08:48.894 "zcopy": true, 00:08:48.894 "get_zone_info": false, 00:08:48.894 "zone_management": false, 00:08:48.894 "zone_append": false, 00:08:48.894 "compare": false, 00:08:48.894 "compare_and_write": false, 00:08:48.894 "abort": true, 00:08:48.894 "seek_hole": false, 00:08:48.894 "seek_data": false, 00:08:48.894 "copy": true, 00:08:48.894 "nvme_iov_md": false 00:08:48.894 }, 00:08:48.894 "memory_domains": [ 00:08:48.894 { 00:08:48.894 "dma_device_id": "system", 00:08:48.894 "dma_device_type": 1 00:08:48.894 }, 00:08:48.894 { 00:08:48.894 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:48.894 "dma_device_type": 2 00:08:48.894 } 00:08:48.894 ], 00:08:48.894 "driver_specific": { 00:08:48.894 "passthru": { 00:08:48.894 "name": "Passthru0", 00:08:48.894 "base_bdev_name": "Malloc0" 00:08:48.894 } 00:08:48.894 } 00:08:48.894 } 00:08:48.894 ]' 00:08:48.894 17:24:30 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # jq length 00:08:48.894 17:24:30 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:08:48.894 17:24:30 rpc.rpc_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:08:48.894 17:24:30 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.894 17:24:30 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:08:48.894 17:24:30 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.894 17:24:30 rpc.rpc_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc0 00:08:48.894 17:24:30 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.894 17:24:30 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:08:48.894 17:24:30 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.894 17:24:30 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:08:48.894 17:24:30 rpc.rpc_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:48.894 17:24:30 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:08:48.894 17:24:30 rpc.rpc_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:48.894 17:24:30 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:08:49.152 17:24:30 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # jq length 00:08:49.152 17:24:30 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:08:49.152 00:08:49.152 real 0m0.228s 00:08:49.152 user 0m0.147s 00:08:49.152 sys 0m0.022s 00:08:49.152 17:24:30 rpc.rpc_integrity -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:49.152 17:24:30 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:08:49.152 ************************************ 00:08:49.152 END TEST rpc_integrity 00:08:49.152 ************************************ 00:08:49.152 17:24:30 rpc -- rpc/rpc.sh@74 -- # run_test rpc_plugins rpc_plugins 00:08:49.152 17:24:30 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:08:49.152 17:24:30 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:49.152 17:24:30 rpc -- common/autotest_common.sh@10 -- # set +x 00:08:49.152 ************************************ 00:08:49.152 START TEST rpc_plugins 00:08:49.152 ************************************ 00:08:49.152 17:24:30 rpc.rpc_plugins -- common/autotest_common.sh@1129 -- # rpc_plugins 00:08:49.152 17:24:30 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # rpc_cmd --plugin rpc_plugin create_malloc 00:08:49.152 17:24:30 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.152 17:24:30 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:08:49.152 17:24:30 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.152 17:24:30 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # malloc=Malloc1 00:08:49.152 17:24:30 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # rpc_cmd bdev_get_bdevs 00:08:49.152 17:24:30 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.152 17:24:30 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:08:49.152 17:24:30 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.152 17:24:30 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # bdevs='[ 00:08:49.152 { 00:08:49.152 "name": "Malloc1", 00:08:49.152 "aliases": [ 00:08:49.152 "f3c71b98-0039-4fa0-9cb5-93efd8668914" 00:08:49.152 ], 00:08:49.152 "product_name": "Malloc disk", 00:08:49.152 "block_size": 4096, 00:08:49.152 "num_blocks": 256, 00:08:49.152 "uuid": "f3c71b98-0039-4fa0-9cb5-93efd8668914", 00:08:49.152 "assigned_rate_limits": { 00:08:49.152 "rw_ios_per_sec": 0, 00:08:49.152 "rw_mbytes_per_sec": 0, 00:08:49.152 "r_mbytes_per_sec": 0, 00:08:49.152 "w_mbytes_per_sec": 0 00:08:49.152 }, 00:08:49.152 "claimed": false, 00:08:49.152 "zoned": false, 00:08:49.152 "supported_io_types": { 00:08:49.152 "read": true, 00:08:49.152 "write": true, 00:08:49.152 "unmap": true, 00:08:49.152 "flush": true, 00:08:49.152 "reset": true, 00:08:49.152 "nvme_admin": false, 00:08:49.152 "nvme_io": false, 00:08:49.152 "nvme_io_md": false, 00:08:49.152 "write_zeroes": true, 00:08:49.152 "zcopy": true, 00:08:49.152 "get_zone_info": false, 00:08:49.152 "zone_management": false, 00:08:49.152 "zone_append": false, 00:08:49.152 "compare": false, 00:08:49.152 "compare_and_write": false, 00:08:49.152 "abort": true, 00:08:49.152 "seek_hole": false, 00:08:49.152 "seek_data": false, 00:08:49.152 "copy": true, 00:08:49.152 "nvme_iov_md": false 00:08:49.152 }, 00:08:49.152 "memory_domains": [ 00:08:49.152 { 00:08:49.152 "dma_device_id": "system", 00:08:49.152 "dma_device_type": 1 00:08:49.152 }, 00:08:49.152 { 00:08:49.152 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:49.152 "dma_device_type": 2 00:08:49.152 } 00:08:49.152 ], 00:08:49.152 "driver_specific": {} 00:08:49.152 } 00:08:49.152 ]' 00:08:49.152 17:24:30 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # jq length 00:08:49.152 17:24:30 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # '[' 1 == 1 ']' 00:08:49.152 17:24:30 rpc.rpc_plugins -- rpc/rpc.sh@34 -- # rpc_cmd --plugin rpc_plugin delete_malloc Malloc1 00:08:49.152 17:24:30 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.152 17:24:30 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:08:49.152 17:24:30 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.152 17:24:30 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # rpc_cmd bdev_get_bdevs 00:08:49.152 17:24:30 rpc.rpc_plugins -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.152 17:24:30 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:08:49.152 17:24:30 rpc.rpc_plugins -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.152 17:24:30 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # bdevs='[]' 00:08:49.152 17:24:30 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # jq length 00:08:49.152 17:24:30 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # '[' 0 == 0 ']' 00:08:49.152 00:08:49.152 real 0m0.108s 00:08:49.152 user 0m0.071s 00:08:49.152 sys 0m0.007s 00:08:49.153 17:24:30 rpc.rpc_plugins -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:49.153 17:24:30 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:08:49.153 ************************************ 00:08:49.153 END TEST rpc_plugins 00:08:49.153 ************************************ 00:08:49.153 17:24:30 rpc -- rpc/rpc.sh@75 -- # run_test rpc_trace_cmd_test rpc_trace_cmd_test 00:08:49.153 17:24:30 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:08:49.153 17:24:30 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:49.153 17:24:30 rpc -- common/autotest_common.sh@10 -- # set +x 00:08:49.153 ************************************ 00:08:49.153 START TEST rpc_trace_cmd_test 00:08:49.153 ************************************ 00:08:49.153 17:24:30 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1129 -- # rpc_trace_cmd_test 00:08:49.153 17:24:30 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@40 -- # local info 00:08:49.153 17:24:30 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # rpc_cmd trace_get_info 00:08:49.153 17:24:30 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.153 17:24:30 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.153 17:24:30 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.153 17:24:30 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # info='{ 00:08:49.153 "tpoint_shm_path": "/dev/shm/spdk_tgt_trace.pid110959", 00:08:49.153 "tpoint_group_mask": "0x8", 00:08:49.153 "iscsi_conn": { 00:08:49.153 "mask": "0x2", 00:08:49.153 "tpoint_mask": "0x0" 00:08:49.153 }, 00:08:49.153 "scsi": { 00:08:49.153 "mask": "0x4", 00:08:49.153 "tpoint_mask": "0x0" 00:08:49.153 }, 00:08:49.153 "bdev": { 00:08:49.153 "mask": "0x8", 00:08:49.153 "tpoint_mask": "0xffffffffffffffff" 00:08:49.153 }, 00:08:49.153 "nvmf_rdma": { 00:08:49.153 "mask": "0x10", 00:08:49.153 "tpoint_mask": "0x0" 00:08:49.153 }, 00:08:49.153 "nvmf_tcp": { 00:08:49.153 "mask": "0x20", 00:08:49.153 "tpoint_mask": "0x0" 00:08:49.153 }, 00:08:49.153 "ftl": { 00:08:49.153 "mask": "0x40", 00:08:49.153 "tpoint_mask": "0x0" 00:08:49.153 }, 00:08:49.153 "blobfs": { 00:08:49.153 "mask": "0x80", 00:08:49.153 "tpoint_mask": "0x0" 00:08:49.153 }, 00:08:49.153 "dsa": { 00:08:49.153 "mask": "0x200", 00:08:49.153 "tpoint_mask": "0x0" 00:08:49.153 }, 00:08:49.153 "thread": { 00:08:49.153 "mask": "0x400", 00:08:49.153 "tpoint_mask": "0x0" 00:08:49.153 }, 00:08:49.153 "nvme_pcie": { 00:08:49.153 "mask": "0x800", 00:08:49.153 "tpoint_mask": "0x0" 00:08:49.153 }, 00:08:49.153 "iaa": { 00:08:49.153 "mask": "0x1000", 00:08:49.153 "tpoint_mask": "0x0" 00:08:49.153 }, 00:08:49.153 "nvme_tcp": { 00:08:49.153 "mask": "0x2000", 00:08:49.153 "tpoint_mask": "0x0" 00:08:49.153 }, 00:08:49.153 "bdev_nvme": { 00:08:49.153 "mask": "0x4000", 00:08:49.153 "tpoint_mask": "0x0" 00:08:49.153 }, 00:08:49.153 "sock": { 00:08:49.153 "mask": "0x8000", 00:08:49.153 "tpoint_mask": "0x0" 00:08:49.153 }, 00:08:49.153 "blob": { 00:08:49.153 "mask": "0x10000", 00:08:49.153 "tpoint_mask": "0x0" 00:08:49.153 }, 00:08:49.153 "bdev_raid": { 00:08:49.153 "mask": "0x20000", 00:08:49.153 "tpoint_mask": "0x0" 00:08:49.153 }, 00:08:49.153 "scheduler": { 00:08:49.153 "mask": "0x40000", 00:08:49.153 "tpoint_mask": "0x0" 00:08:49.153 } 00:08:49.153 }' 00:08:49.153 17:24:30 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # jq length 00:08:49.411 17:24:31 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # '[' 19 -gt 2 ']' 00:08:49.411 17:24:31 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # jq 'has("tpoint_group_mask")' 00:08:49.411 17:24:31 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # '[' true = true ']' 00:08:49.411 17:24:31 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # jq 'has("tpoint_shm_path")' 00:08:49.411 17:24:31 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # '[' true = true ']' 00:08:49.411 17:24:31 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # jq 'has("bdev")' 00:08:49.411 17:24:31 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # '[' true = true ']' 00:08:49.411 17:24:31 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # jq -r .bdev.tpoint_mask 00:08:49.411 17:24:31 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # '[' 0xffffffffffffffff '!=' 0x0 ']' 00:08:49.411 00:08:49.411 real 0m0.185s 00:08:49.411 user 0m0.155s 00:08:49.411 sys 0m0.021s 00:08:49.411 17:24:31 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:49.411 17:24:31 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:08:49.411 ************************************ 00:08:49.411 END TEST rpc_trace_cmd_test 00:08:49.411 ************************************ 00:08:49.411 17:24:31 rpc -- rpc/rpc.sh@76 -- # [[ 0 -eq 1 ]] 00:08:49.411 17:24:31 rpc -- rpc/rpc.sh@80 -- # rpc=rpc_cmd 00:08:49.411 17:24:31 rpc -- rpc/rpc.sh@81 -- # run_test rpc_daemon_integrity rpc_integrity 00:08:49.411 17:24:31 rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:08:49.411 17:24:31 rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:49.411 17:24:31 rpc -- common/autotest_common.sh@10 -- # set +x 00:08:49.411 ************************************ 00:08:49.411 START TEST rpc_daemon_integrity 00:08:49.411 ************************************ 00:08:49.411 17:24:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1129 -- # rpc_integrity 00:08:49.411 17:24:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:08:49.411 17:24:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.411 17:24:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:08:49.411 17:24:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.411 17:24:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:08:49.411 17:24:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # jq length 00:08:49.411 17:24:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:08:49.411 17:24:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:08:49.411 17:24:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.411 17:24:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:08:49.411 17:24:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.411 17:24:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc2 00:08:49.411 17:24:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:08:49.669 17:24:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.669 17:24:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:08:49.669 17:24:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.669 17:24:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:08:49.669 { 00:08:49.669 "name": "Malloc2", 00:08:49.669 "aliases": [ 00:08:49.669 "30268aec-fbd7-4bb2-a1cf-0b03c4a9813d" 00:08:49.669 ], 00:08:49.669 "product_name": "Malloc disk", 00:08:49.669 "block_size": 512, 00:08:49.669 "num_blocks": 16384, 00:08:49.669 "uuid": "30268aec-fbd7-4bb2-a1cf-0b03c4a9813d", 00:08:49.669 "assigned_rate_limits": { 00:08:49.669 "rw_ios_per_sec": 0, 00:08:49.669 "rw_mbytes_per_sec": 0, 00:08:49.669 "r_mbytes_per_sec": 0, 00:08:49.669 "w_mbytes_per_sec": 0 00:08:49.669 }, 00:08:49.669 "claimed": false, 00:08:49.669 "zoned": false, 00:08:49.669 "supported_io_types": { 00:08:49.669 "read": true, 00:08:49.669 "write": true, 00:08:49.669 "unmap": true, 00:08:49.669 "flush": true, 00:08:49.669 "reset": true, 00:08:49.670 "nvme_admin": false, 00:08:49.670 "nvme_io": false, 00:08:49.670 "nvme_io_md": false, 00:08:49.670 "write_zeroes": true, 00:08:49.670 "zcopy": true, 00:08:49.670 "get_zone_info": false, 00:08:49.670 "zone_management": false, 00:08:49.670 "zone_append": false, 00:08:49.670 "compare": false, 00:08:49.670 "compare_and_write": false, 00:08:49.670 "abort": true, 00:08:49.670 "seek_hole": false, 00:08:49.670 "seek_data": false, 00:08:49.670 "copy": true, 00:08:49.670 "nvme_iov_md": false 00:08:49.670 }, 00:08:49.670 "memory_domains": [ 00:08:49.670 { 00:08:49.670 "dma_device_id": "system", 00:08:49.670 "dma_device_type": 1 00:08:49.670 }, 00:08:49.670 { 00:08:49.670 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:49.670 "dma_device_type": 2 00:08:49.670 } 00:08:49.670 ], 00:08:49.670 "driver_specific": {} 00:08:49.670 } 00:08:49.670 ]' 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # jq length 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc2 -p Passthru0 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:08:49.670 [2024-12-06 17:24:31.299232] vbdev_passthru.c: 608:vbdev_passthru_register: *NOTICE*: Match on Malloc2 00:08:49.670 [2024-12-06 17:24:31.299287] vbdev_passthru.c: 636:vbdev_passthru_register: *NOTICE*: base bdev opened 00:08:49.670 [2024-12-06 17:24:31.299315] vbdev_passthru.c: 682:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x2261a80 00:08:49.670 [2024-12-06 17:24:31.299328] vbdev_passthru.c: 697:vbdev_passthru_register: *NOTICE*: bdev claimed 00:08:49.670 [2024-12-06 17:24:31.300504] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:08:49.670 [2024-12-06 17:24:31.300526] vbdev_passthru.c: 711:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:08:49.670 Passthru0 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:08:49.670 { 00:08:49.670 "name": "Malloc2", 00:08:49.670 "aliases": [ 00:08:49.670 "30268aec-fbd7-4bb2-a1cf-0b03c4a9813d" 00:08:49.670 ], 00:08:49.670 "product_name": "Malloc disk", 00:08:49.670 "block_size": 512, 00:08:49.670 "num_blocks": 16384, 00:08:49.670 "uuid": "30268aec-fbd7-4bb2-a1cf-0b03c4a9813d", 00:08:49.670 "assigned_rate_limits": { 00:08:49.670 "rw_ios_per_sec": 0, 00:08:49.670 "rw_mbytes_per_sec": 0, 00:08:49.670 "r_mbytes_per_sec": 0, 00:08:49.670 "w_mbytes_per_sec": 0 00:08:49.670 }, 00:08:49.670 "claimed": true, 00:08:49.670 "claim_type": "exclusive_write", 00:08:49.670 "zoned": false, 00:08:49.670 "supported_io_types": { 00:08:49.670 "read": true, 00:08:49.670 "write": true, 00:08:49.670 "unmap": true, 00:08:49.670 "flush": true, 00:08:49.670 "reset": true, 00:08:49.670 "nvme_admin": false, 00:08:49.670 "nvme_io": false, 00:08:49.670 "nvme_io_md": false, 00:08:49.670 "write_zeroes": true, 00:08:49.670 "zcopy": true, 00:08:49.670 "get_zone_info": false, 00:08:49.670 "zone_management": false, 00:08:49.670 "zone_append": false, 00:08:49.670 "compare": false, 00:08:49.670 "compare_and_write": false, 00:08:49.670 "abort": true, 00:08:49.670 "seek_hole": false, 00:08:49.670 "seek_data": false, 00:08:49.670 "copy": true, 00:08:49.670 "nvme_iov_md": false 00:08:49.670 }, 00:08:49.670 "memory_domains": [ 00:08:49.670 { 00:08:49.670 "dma_device_id": "system", 00:08:49.670 "dma_device_type": 1 00:08:49.670 }, 00:08:49.670 { 00:08:49.670 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:49.670 "dma_device_type": 2 00:08:49.670 } 00:08:49.670 ], 00:08:49.670 "driver_specific": {} 00:08:49.670 }, 00:08:49.670 { 00:08:49.670 "name": "Passthru0", 00:08:49.670 "aliases": [ 00:08:49.670 "5f0868cf-d7f8-5893-b24b-5ae8b5cd9ebb" 00:08:49.670 ], 00:08:49.670 "product_name": "passthru", 00:08:49.670 "block_size": 512, 00:08:49.670 "num_blocks": 16384, 00:08:49.670 "uuid": "5f0868cf-d7f8-5893-b24b-5ae8b5cd9ebb", 00:08:49.670 "assigned_rate_limits": { 00:08:49.670 "rw_ios_per_sec": 0, 00:08:49.670 "rw_mbytes_per_sec": 0, 00:08:49.670 "r_mbytes_per_sec": 0, 00:08:49.670 "w_mbytes_per_sec": 0 00:08:49.670 }, 00:08:49.670 "claimed": false, 00:08:49.670 "zoned": false, 00:08:49.670 "supported_io_types": { 00:08:49.670 "read": true, 00:08:49.670 "write": true, 00:08:49.670 "unmap": true, 00:08:49.670 "flush": true, 00:08:49.670 "reset": true, 00:08:49.670 "nvme_admin": false, 00:08:49.670 "nvme_io": false, 00:08:49.670 "nvme_io_md": false, 00:08:49.670 "write_zeroes": true, 00:08:49.670 "zcopy": true, 00:08:49.670 "get_zone_info": false, 00:08:49.670 "zone_management": false, 00:08:49.670 "zone_append": false, 00:08:49.670 "compare": false, 00:08:49.670 "compare_and_write": false, 00:08:49.670 "abort": true, 00:08:49.670 "seek_hole": false, 00:08:49.670 "seek_data": false, 00:08:49.670 "copy": true, 00:08:49.670 "nvme_iov_md": false 00:08:49.670 }, 00:08:49.670 "memory_domains": [ 00:08:49.670 { 00:08:49.670 "dma_device_id": "system", 00:08:49.670 "dma_device_type": 1 00:08:49.670 }, 00:08:49.670 { 00:08:49.670 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:08:49.670 "dma_device_type": 2 00:08:49.670 } 00:08:49.670 ], 00:08:49.670 "driver_specific": { 00:08:49.670 "passthru": { 00:08:49.670 "name": "Passthru0", 00:08:49.670 "base_bdev_name": "Malloc2" 00:08:49.670 } 00:08:49.670 } 00:08:49.670 } 00:08:49.670 ]' 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # jq length 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc2 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # jq length 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:08:49.670 00:08:49.670 real 0m0.214s 00:08:49.670 user 0m0.137s 00:08:49.670 sys 0m0.020s 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:49.670 17:24:31 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:08:49.670 ************************************ 00:08:49.670 END TEST rpc_daemon_integrity 00:08:49.670 ************************************ 00:08:49.670 17:24:31 rpc -- rpc/rpc.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:08:49.670 17:24:31 rpc -- rpc/rpc.sh@84 -- # killprocess 110959 00:08:49.670 17:24:31 rpc -- common/autotest_common.sh@954 -- # '[' -z 110959 ']' 00:08:49.670 17:24:31 rpc -- common/autotest_common.sh@958 -- # kill -0 110959 00:08:49.670 17:24:31 rpc -- common/autotest_common.sh@959 -- # uname 00:08:49.670 17:24:31 rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:49.670 17:24:31 rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 110959 00:08:49.670 17:24:31 rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:49.670 17:24:31 rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:49.670 17:24:31 rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 110959' 00:08:49.670 killing process with pid 110959 00:08:49.670 17:24:31 rpc -- common/autotest_common.sh@973 -- # kill 110959 00:08:49.670 17:24:31 rpc -- common/autotest_common.sh@978 -- # wait 110959 00:08:50.237 00:08:50.237 real 0m1.942s 00:08:50.237 user 0m2.423s 00:08:50.237 sys 0m0.600s 00:08:50.237 17:24:31 rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:50.237 17:24:31 rpc -- common/autotest_common.sh@10 -- # set +x 00:08:50.237 ************************************ 00:08:50.237 END TEST rpc 00:08:50.237 ************************************ 00:08:50.237 17:24:31 -- spdk/autotest.sh@157 -- # run_test skip_rpc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/skip_rpc.sh 00:08:50.237 17:24:31 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:08:50.237 17:24:31 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:50.237 17:24:31 -- common/autotest_common.sh@10 -- # set +x 00:08:50.237 ************************************ 00:08:50.237 START TEST skip_rpc 00:08:50.237 ************************************ 00:08:50.237 17:24:31 skip_rpc -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/skip_rpc.sh 00:08:50.237 * Looking for test storage... 00:08:50.237 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc 00:08:50.237 17:24:31 skip_rpc -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:08:50.237 17:24:31 skip_rpc -- common/autotest_common.sh@1711 -- # lcov --version 00:08:50.237 17:24:31 skip_rpc -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:08:50.237 17:24:32 skip_rpc -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:08:50.237 17:24:32 skip_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:50.237 17:24:32 skip_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:50.237 17:24:32 skip_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:50.237 17:24:32 skip_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:08:50.237 17:24:32 skip_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:08:50.237 17:24:32 skip_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:08:50.237 17:24:32 skip_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:08:50.237 17:24:32 skip_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:08:50.237 17:24:32 skip_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:08:50.237 17:24:32 skip_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:08:50.237 17:24:32 skip_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:50.237 17:24:32 skip_rpc -- scripts/common.sh@344 -- # case "$op" in 00:08:50.237 17:24:32 skip_rpc -- scripts/common.sh@345 -- # : 1 00:08:50.237 17:24:32 skip_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:50.237 17:24:32 skip_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:50.237 17:24:32 skip_rpc -- scripts/common.sh@365 -- # decimal 1 00:08:50.237 17:24:32 skip_rpc -- scripts/common.sh@353 -- # local d=1 00:08:50.237 17:24:32 skip_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:50.237 17:24:32 skip_rpc -- scripts/common.sh@355 -- # echo 1 00:08:50.237 17:24:32 skip_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:08:50.237 17:24:32 skip_rpc -- scripts/common.sh@366 -- # decimal 2 00:08:50.237 17:24:32 skip_rpc -- scripts/common.sh@353 -- # local d=2 00:08:50.238 17:24:32 skip_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:50.238 17:24:32 skip_rpc -- scripts/common.sh@355 -- # echo 2 00:08:50.238 17:24:32 skip_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:08:50.238 17:24:32 skip_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:50.238 17:24:32 skip_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:50.238 17:24:32 skip_rpc -- scripts/common.sh@368 -- # return 0 00:08:50.238 17:24:32 skip_rpc -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:50.238 17:24:32 skip_rpc -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:08:50.238 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:50.238 --rc genhtml_branch_coverage=1 00:08:50.238 --rc genhtml_function_coverage=1 00:08:50.238 --rc genhtml_legend=1 00:08:50.238 --rc geninfo_all_blocks=1 00:08:50.238 --rc geninfo_unexecuted_blocks=1 00:08:50.238 00:08:50.238 ' 00:08:50.238 17:24:32 skip_rpc -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:08:50.238 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:50.238 --rc genhtml_branch_coverage=1 00:08:50.238 --rc genhtml_function_coverage=1 00:08:50.238 --rc genhtml_legend=1 00:08:50.238 --rc geninfo_all_blocks=1 00:08:50.238 --rc geninfo_unexecuted_blocks=1 00:08:50.238 00:08:50.238 ' 00:08:50.238 17:24:32 skip_rpc -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:08:50.238 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:50.238 --rc genhtml_branch_coverage=1 00:08:50.238 --rc genhtml_function_coverage=1 00:08:50.238 --rc genhtml_legend=1 00:08:50.238 --rc geninfo_all_blocks=1 00:08:50.238 --rc geninfo_unexecuted_blocks=1 00:08:50.238 00:08:50.238 ' 00:08:50.238 17:24:32 skip_rpc -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:08:50.238 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:50.238 --rc genhtml_branch_coverage=1 00:08:50.238 --rc genhtml_function_coverage=1 00:08:50.238 --rc genhtml_legend=1 00:08:50.238 --rc geninfo_all_blocks=1 00:08:50.238 --rc geninfo_unexecuted_blocks=1 00:08:50.238 00:08:50.238 ' 00:08:50.238 17:24:32 skip_rpc -- rpc/skip_rpc.sh@11 -- # CONFIG_PATH=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/config.json 00:08:50.238 17:24:32 skip_rpc -- rpc/skip_rpc.sh@12 -- # LOG_PATH=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/log.txt 00:08:50.238 17:24:32 skip_rpc -- rpc/skip_rpc.sh@73 -- # run_test skip_rpc test_skip_rpc 00:08:50.238 17:24:32 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:08:50.238 17:24:32 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:50.238 17:24:32 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:50.496 ************************************ 00:08:50.496 START TEST skip_rpc 00:08:50.496 ************************************ 00:08:50.496 17:24:32 skip_rpc.skip_rpc -- common/autotest_common.sh@1129 -- # test_skip_rpc 00:08:50.496 17:24:32 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@16 -- # local spdk_pid=111280 00:08:50.496 17:24:32 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 00:08:50.496 17:24:32 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@18 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:08:50.496 17:24:32 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@19 -- # sleep 5 00:08:50.496 [2024-12-06 17:24:32.139267] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:08:50.496 [2024-12-06 17:24:32.139351] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid111280 ] 00:08:50.496 [2024-12-06 17:24:32.205733] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:50.496 [2024-12-06 17:24:32.255147] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:55.753 17:24:37 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@21 -- # NOT rpc_cmd spdk_get_version 00:08:55.753 17:24:37 skip_rpc.skip_rpc -- common/autotest_common.sh@652 -- # local es=0 00:08:55.753 17:24:37 skip_rpc.skip_rpc -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd spdk_get_version 00:08:55.753 17:24:37 skip_rpc.skip_rpc -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:08:55.753 17:24:37 skip_rpc.skip_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:55.753 17:24:37 skip_rpc.skip_rpc -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:08:55.753 17:24:37 skip_rpc.skip_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:08:55.753 17:24:37 skip_rpc.skip_rpc -- common/autotest_common.sh@655 -- # rpc_cmd spdk_get_version 00:08:55.753 17:24:37 skip_rpc.skip_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:55.753 17:24:37 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:55.753 17:24:37 skip_rpc.skip_rpc -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:08:55.753 17:24:37 skip_rpc.skip_rpc -- common/autotest_common.sh@655 -- # es=1 00:08:55.753 17:24:37 skip_rpc.skip_rpc -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:08:55.754 17:24:37 skip_rpc.skip_rpc -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:08:55.754 17:24:37 skip_rpc.skip_rpc -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:08:55.754 17:24:37 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@22 -- # trap - SIGINT SIGTERM EXIT 00:08:55.754 17:24:37 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@23 -- # killprocess 111280 00:08:55.754 17:24:37 skip_rpc.skip_rpc -- common/autotest_common.sh@954 -- # '[' -z 111280 ']' 00:08:55.754 17:24:37 skip_rpc.skip_rpc -- common/autotest_common.sh@958 -- # kill -0 111280 00:08:55.754 17:24:37 skip_rpc.skip_rpc -- common/autotest_common.sh@959 -- # uname 00:08:55.754 17:24:37 skip_rpc.skip_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:55.754 17:24:37 skip_rpc.skip_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 111280 00:08:55.754 17:24:37 skip_rpc.skip_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:55.754 17:24:37 skip_rpc.skip_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:55.754 17:24:37 skip_rpc.skip_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 111280' 00:08:55.754 killing process with pid 111280 00:08:55.754 17:24:37 skip_rpc.skip_rpc -- common/autotest_common.sh@973 -- # kill 111280 00:08:55.754 17:24:37 skip_rpc.skip_rpc -- common/autotest_common.sh@978 -- # wait 111280 00:08:55.754 00:08:55.754 real 0m5.427s 00:08:55.754 user 0m5.141s 00:08:55.754 sys 0m0.305s 00:08:55.754 17:24:37 skip_rpc.skip_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:08:55.754 17:24:37 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:55.754 ************************************ 00:08:55.754 END TEST skip_rpc 00:08:55.754 ************************************ 00:08:55.754 17:24:37 skip_rpc -- rpc/skip_rpc.sh@74 -- # run_test skip_rpc_with_json test_skip_rpc_with_json 00:08:55.754 17:24:37 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:08:55.754 17:24:37 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:08:55.754 17:24:37 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:55.754 ************************************ 00:08:55.754 START TEST skip_rpc_with_json 00:08:55.754 ************************************ 00:08:55.754 17:24:37 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1129 -- # test_skip_rpc_with_json 00:08:55.754 17:24:37 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@44 -- # gen_json_config 00:08:55.754 17:24:37 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@28 -- # local spdk_pid=111967 00:08:55.754 17:24:37 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:08:55.754 17:24:37 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@30 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:08:55.754 17:24:37 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@31 -- # waitforlisten 111967 00:08:55.754 17:24:37 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@835 -- # '[' -z 111967 ']' 00:08:55.754 17:24:37 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:55.754 17:24:37 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@840 -- # local max_retries=100 00:08:55.754 17:24:37 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:55.754 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:55.754 17:24:37 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@844 -- # xtrace_disable 00:08:55.754 17:24:37 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:08:56.012 [2024-12-06 17:24:37.624849] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:08:56.012 [2024-12-06 17:24:37.624954] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid111967 ] 00:08:56.012 [2024-12-06 17:24:37.698458] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:56.012 [2024-12-06 17:24:37.746145] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:56.271 17:24:37 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:08:56.271 17:24:37 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@868 -- # return 0 00:08:56.271 17:24:37 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_get_transports --trtype tcp 00:08:56.271 17:24:37 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.271 17:24:37 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:08:56.271 [2024-12-06 17:24:38.004038] nvmf_rpc.c:2707:rpc_nvmf_get_transports: *ERROR*: transport 'tcp' does not exist 00:08:56.271 request: 00:08:56.271 { 00:08:56.271 "trtype": "tcp", 00:08:56.271 "method": "nvmf_get_transports", 00:08:56.271 "req_id": 1 00:08:56.271 } 00:08:56.271 Got JSON-RPC error response 00:08:56.271 response: 00:08:56.271 { 00:08:56.271 "code": -19, 00:08:56.271 "message": "No such device" 00:08:56.271 } 00:08:56.271 17:24:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:08:56.271 17:24:38 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_create_transport -t tcp 00:08:56.271 17:24:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.271 17:24:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:08:56.271 [2024-12-06 17:24:38.012147] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:08:56.271 17:24:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:56.271 17:24:38 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@36 -- # rpc_cmd save_config 00:08:56.271 17:24:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@563 -- # xtrace_disable 00:08:56.271 17:24:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:08:56.530 17:24:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:08:56.530 17:24:38 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@37 -- # cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/config.json 00:08:56.530 { 00:08:56.530 "subsystems": [ 00:08:56.530 { 00:08:56.530 "subsystem": "fsdev", 00:08:56.530 "config": [ 00:08:56.530 { 00:08:56.530 "method": "fsdev_set_opts", 00:08:56.530 "params": { 00:08:56.530 "fsdev_io_pool_size": 65535, 00:08:56.530 "fsdev_io_cache_size": 256 00:08:56.530 } 00:08:56.530 } 00:08:56.530 ] 00:08:56.530 }, 00:08:56.530 { 00:08:56.530 "subsystem": "vfio_user_target", 00:08:56.530 "config": null 00:08:56.530 }, 00:08:56.530 { 00:08:56.530 "subsystem": "keyring", 00:08:56.530 "config": [] 00:08:56.530 }, 00:08:56.530 { 00:08:56.530 "subsystem": "iobuf", 00:08:56.530 "config": [ 00:08:56.530 { 00:08:56.530 "method": "iobuf_set_options", 00:08:56.530 "params": { 00:08:56.530 "small_pool_count": 8192, 00:08:56.530 "large_pool_count": 1024, 00:08:56.530 "small_bufsize": 8192, 00:08:56.530 "large_bufsize": 135168, 00:08:56.530 "enable_numa": false 00:08:56.530 } 00:08:56.530 } 00:08:56.530 ] 00:08:56.530 }, 00:08:56.530 { 00:08:56.531 "subsystem": "sock", 00:08:56.531 "config": [ 00:08:56.531 { 00:08:56.531 "method": "sock_set_default_impl", 00:08:56.531 "params": { 00:08:56.531 "impl_name": "posix" 00:08:56.531 } 00:08:56.531 }, 00:08:56.531 { 00:08:56.531 "method": "sock_impl_set_options", 00:08:56.531 "params": { 00:08:56.531 "impl_name": "ssl", 00:08:56.531 "recv_buf_size": 4096, 00:08:56.531 "send_buf_size": 4096, 00:08:56.531 "enable_recv_pipe": true, 00:08:56.531 "enable_quickack": false, 00:08:56.531 "enable_placement_id": 0, 00:08:56.531 "enable_zerocopy_send_server": true, 00:08:56.531 "enable_zerocopy_send_client": false, 00:08:56.531 "zerocopy_threshold": 0, 00:08:56.531 "tls_version": 0, 00:08:56.531 "enable_ktls": false 00:08:56.531 } 00:08:56.531 }, 00:08:56.531 { 00:08:56.531 "method": "sock_impl_set_options", 00:08:56.531 "params": { 00:08:56.531 "impl_name": "posix", 00:08:56.531 "recv_buf_size": 2097152, 00:08:56.531 "send_buf_size": 2097152, 00:08:56.531 "enable_recv_pipe": true, 00:08:56.531 "enable_quickack": false, 00:08:56.531 "enable_placement_id": 0, 00:08:56.531 "enable_zerocopy_send_server": true, 00:08:56.531 "enable_zerocopy_send_client": false, 00:08:56.531 "zerocopy_threshold": 0, 00:08:56.531 "tls_version": 0, 00:08:56.531 "enable_ktls": false 00:08:56.531 } 00:08:56.531 } 00:08:56.531 ] 00:08:56.531 }, 00:08:56.531 { 00:08:56.531 "subsystem": "vmd", 00:08:56.531 "config": [] 00:08:56.531 }, 00:08:56.531 { 00:08:56.531 "subsystem": "accel", 00:08:56.531 "config": [ 00:08:56.531 { 00:08:56.531 "method": "accel_set_options", 00:08:56.531 "params": { 00:08:56.531 "small_cache_size": 128, 00:08:56.531 "large_cache_size": 16, 00:08:56.531 "task_count": 2048, 00:08:56.531 "sequence_count": 2048, 00:08:56.531 "buf_count": 2048 00:08:56.531 } 00:08:56.531 } 00:08:56.531 ] 00:08:56.531 }, 00:08:56.531 { 00:08:56.531 "subsystem": "bdev", 00:08:56.531 "config": [ 00:08:56.531 { 00:08:56.531 "method": "bdev_set_options", 00:08:56.531 "params": { 00:08:56.531 "bdev_io_pool_size": 65535, 00:08:56.531 "bdev_io_cache_size": 256, 00:08:56.531 "bdev_auto_examine": true, 00:08:56.531 "iobuf_small_cache_size": 128, 00:08:56.531 "iobuf_large_cache_size": 16 00:08:56.531 } 00:08:56.531 }, 00:08:56.531 { 00:08:56.531 "method": "bdev_raid_set_options", 00:08:56.531 "params": { 00:08:56.531 "process_window_size_kb": 1024, 00:08:56.531 "process_max_bandwidth_mb_sec": 0 00:08:56.531 } 00:08:56.531 }, 00:08:56.531 { 00:08:56.531 "method": "bdev_iscsi_set_options", 00:08:56.531 "params": { 00:08:56.531 "timeout_sec": 30 00:08:56.531 } 00:08:56.531 }, 00:08:56.531 { 00:08:56.531 "method": "bdev_nvme_set_options", 00:08:56.531 "params": { 00:08:56.531 "action_on_timeout": "none", 00:08:56.531 "timeout_us": 0, 00:08:56.531 "timeout_admin_us": 0, 00:08:56.531 "keep_alive_timeout_ms": 10000, 00:08:56.531 "arbitration_burst": 0, 00:08:56.531 "low_priority_weight": 0, 00:08:56.531 "medium_priority_weight": 0, 00:08:56.531 "high_priority_weight": 0, 00:08:56.531 "nvme_adminq_poll_period_us": 10000, 00:08:56.531 "nvme_ioq_poll_period_us": 0, 00:08:56.531 "io_queue_requests": 0, 00:08:56.531 "delay_cmd_submit": true, 00:08:56.531 "transport_retry_count": 4, 00:08:56.531 "bdev_retry_count": 3, 00:08:56.531 "transport_ack_timeout": 0, 00:08:56.531 "ctrlr_loss_timeout_sec": 0, 00:08:56.531 "reconnect_delay_sec": 0, 00:08:56.531 "fast_io_fail_timeout_sec": 0, 00:08:56.531 "disable_auto_failback": false, 00:08:56.531 "generate_uuids": false, 00:08:56.531 "transport_tos": 0, 00:08:56.531 "nvme_error_stat": false, 00:08:56.531 "rdma_srq_size": 0, 00:08:56.531 "io_path_stat": false, 00:08:56.531 "allow_accel_sequence": false, 00:08:56.531 "rdma_max_cq_size": 0, 00:08:56.531 "rdma_cm_event_timeout_ms": 0, 00:08:56.531 "dhchap_digests": [ 00:08:56.531 "sha256", 00:08:56.531 "sha384", 00:08:56.531 "sha512" 00:08:56.531 ], 00:08:56.531 "dhchap_dhgroups": [ 00:08:56.531 "null", 00:08:56.531 "ffdhe2048", 00:08:56.531 "ffdhe3072", 00:08:56.531 "ffdhe4096", 00:08:56.531 "ffdhe6144", 00:08:56.531 "ffdhe8192" 00:08:56.531 ] 00:08:56.531 } 00:08:56.531 }, 00:08:56.531 { 00:08:56.531 "method": "bdev_nvme_set_hotplug", 00:08:56.531 "params": { 00:08:56.531 "period_us": 100000, 00:08:56.531 "enable": false 00:08:56.531 } 00:08:56.531 }, 00:08:56.531 { 00:08:56.531 "method": "bdev_wait_for_examine" 00:08:56.531 } 00:08:56.531 ] 00:08:56.531 }, 00:08:56.531 { 00:08:56.531 "subsystem": "scsi", 00:08:56.531 "config": null 00:08:56.531 }, 00:08:56.531 { 00:08:56.531 "subsystem": "scheduler", 00:08:56.531 "config": [ 00:08:56.531 { 00:08:56.531 "method": "framework_set_scheduler", 00:08:56.531 "params": { 00:08:56.531 "name": "static" 00:08:56.531 } 00:08:56.531 } 00:08:56.531 ] 00:08:56.531 }, 00:08:56.531 { 00:08:56.531 "subsystem": "vhost_scsi", 00:08:56.531 "config": [] 00:08:56.531 }, 00:08:56.531 { 00:08:56.531 "subsystem": "vhost_blk", 00:08:56.531 "config": [] 00:08:56.531 }, 00:08:56.531 { 00:08:56.531 "subsystem": "ublk", 00:08:56.531 "config": [] 00:08:56.531 }, 00:08:56.531 { 00:08:56.531 "subsystem": "nbd", 00:08:56.531 "config": [] 00:08:56.531 }, 00:08:56.531 { 00:08:56.531 "subsystem": "nvmf", 00:08:56.531 "config": [ 00:08:56.531 { 00:08:56.531 "method": "nvmf_set_config", 00:08:56.531 "params": { 00:08:56.531 "discovery_filter": "match_any", 00:08:56.531 "admin_cmd_passthru": { 00:08:56.531 "identify_ctrlr": false 00:08:56.531 }, 00:08:56.531 "dhchap_digests": [ 00:08:56.531 "sha256", 00:08:56.531 "sha384", 00:08:56.531 "sha512" 00:08:56.531 ], 00:08:56.531 "dhchap_dhgroups": [ 00:08:56.531 "null", 00:08:56.531 "ffdhe2048", 00:08:56.531 "ffdhe3072", 00:08:56.531 "ffdhe4096", 00:08:56.531 "ffdhe6144", 00:08:56.531 "ffdhe8192" 00:08:56.531 ] 00:08:56.531 } 00:08:56.531 }, 00:08:56.531 { 00:08:56.531 "method": "nvmf_set_max_subsystems", 00:08:56.531 "params": { 00:08:56.531 "max_subsystems": 1024 00:08:56.531 } 00:08:56.531 }, 00:08:56.531 { 00:08:56.531 "method": "nvmf_set_crdt", 00:08:56.531 "params": { 00:08:56.531 "crdt1": 0, 00:08:56.531 "crdt2": 0, 00:08:56.531 "crdt3": 0 00:08:56.531 } 00:08:56.531 }, 00:08:56.531 { 00:08:56.531 "method": "nvmf_create_transport", 00:08:56.531 "params": { 00:08:56.531 "trtype": "TCP", 00:08:56.531 "max_queue_depth": 128, 00:08:56.531 "max_io_qpairs_per_ctrlr": 127, 00:08:56.531 "in_capsule_data_size": 4096, 00:08:56.531 "max_io_size": 131072, 00:08:56.531 "io_unit_size": 131072, 00:08:56.531 "max_aq_depth": 128, 00:08:56.531 "num_shared_buffers": 511, 00:08:56.531 "buf_cache_size": 4294967295, 00:08:56.531 "dif_insert_or_strip": false, 00:08:56.531 "zcopy": false, 00:08:56.531 "c2h_success": true, 00:08:56.531 "sock_priority": 0, 00:08:56.531 "abort_timeout_sec": 1, 00:08:56.531 "ack_timeout": 0, 00:08:56.531 "data_wr_pool_size": 0 00:08:56.531 } 00:08:56.531 } 00:08:56.531 ] 00:08:56.531 }, 00:08:56.531 { 00:08:56.531 "subsystem": "iscsi", 00:08:56.531 "config": [ 00:08:56.531 { 00:08:56.531 "method": "iscsi_set_options", 00:08:56.531 "params": { 00:08:56.531 "node_base": "iqn.2016-06.io.spdk", 00:08:56.531 "max_sessions": 128, 00:08:56.531 "max_connections_per_session": 2, 00:08:56.531 "max_queue_depth": 64, 00:08:56.531 "default_time2wait": 2, 00:08:56.531 "default_time2retain": 20, 00:08:56.531 "first_burst_length": 8192, 00:08:56.531 "immediate_data": true, 00:08:56.531 "allow_duplicated_isid": false, 00:08:56.531 "error_recovery_level": 0, 00:08:56.531 "nop_timeout": 60, 00:08:56.531 "nop_in_interval": 30, 00:08:56.531 "disable_chap": false, 00:08:56.531 "require_chap": false, 00:08:56.531 "mutual_chap": false, 00:08:56.531 "chap_group": 0, 00:08:56.531 "max_large_datain_per_connection": 64, 00:08:56.531 "max_r2t_per_connection": 4, 00:08:56.531 "pdu_pool_size": 36864, 00:08:56.531 "immediate_data_pool_size": 16384, 00:08:56.531 "data_out_pool_size": 2048 00:08:56.531 } 00:08:56.531 } 00:08:56.531 ] 00:08:56.531 } 00:08:56.531 ] 00:08:56.531 } 00:08:56.531 17:24:38 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@39 -- # trap - SIGINT SIGTERM EXIT 00:08:56.531 17:24:38 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@40 -- # killprocess 111967 00:08:56.531 17:24:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # '[' -z 111967 ']' 00:08:56.531 17:24:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@958 -- # kill -0 111967 00:08:56.531 17:24:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # uname 00:08:56.531 17:24:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:08:56.531 17:24:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 111967 00:08:56.531 17:24:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:08:56.532 17:24:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:08:56.532 17:24:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@972 -- # echo 'killing process with pid 111967' 00:08:56.532 killing process with pid 111967 00:08:56.532 17:24:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@973 -- # kill 111967 00:08:56.532 17:24:38 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@978 -- # wait 111967 00:08:56.792 17:24:38 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@47 -- # local spdk_pid=112109 00:08:56.793 17:24:38 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/config.json 00:08:56.793 17:24:38 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@48 -- # sleep 5 00:09:02.061 17:24:43 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@50 -- # killprocess 112109 00:09:02.061 17:24:43 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # '[' -z 112109 ']' 00:09:02.061 17:24:43 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@958 -- # kill -0 112109 00:09:02.061 17:24:43 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # uname 00:09:02.061 17:24:43 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:02.061 17:24:43 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 112109 00:09:02.061 17:24:43 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:02.061 17:24:43 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:02.061 17:24:43 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@972 -- # echo 'killing process with pid 112109' 00:09:02.061 killing process with pid 112109 00:09:02.061 17:24:43 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@973 -- # kill 112109 00:09:02.061 17:24:43 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@978 -- # wait 112109 00:09:02.324 17:24:44 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@51 -- # grep -q 'TCP Transport Init' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/log.txt 00:09:02.324 17:24:44 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@52 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/log.txt 00:09:02.324 00:09:02.324 real 0m6.462s 00:09:02.324 user 0m6.113s 00:09:02.324 sys 0m0.670s 00:09:02.324 17:24:44 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:02.324 17:24:44 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:09:02.324 ************************************ 00:09:02.324 END TEST skip_rpc_with_json 00:09:02.324 ************************************ 00:09:02.324 17:24:44 skip_rpc -- rpc/skip_rpc.sh@75 -- # run_test skip_rpc_with_delay test_skip_rpc_with_delay 00:09:02.324 17:24:44 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:09:02.324 17:24:44 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:02.324 17:24:44 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:09:02.324 ************************************ 00:09:02.324 START TEST skip_rpc_with_delay 00:09:02.324 ************************************ 00:09:02.324 17:24:44 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1129 -- # test_skip_rpc_with_delay 00:09:02.324 17:24:44 skip_rpc.skip_rpc_with_delay -- rpc/skip_rpc.sh@57 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:09:02.324 17:24:44 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@652 -- # local es=0 00:09:02.324 17:24:44 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@654 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:09:02.324 17:24:44 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@640 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:09:02.324 17:24:44 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:02.324 17:24:44 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:09:02.324 17:24:44 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:02.324 17:24:44 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@646 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:09:02.324 17:24:44 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:02.324 17:24:44 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@646 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:09:02.324 17:24:44 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@646 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt ]] 00:09:02.324 17:24:44 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@655 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:09:02.324 [2024-12-06 17:24:44.142720] app.c: 842:spdk_app_start: *ERROR*: Cannot use '--wait-for-rpc' if no RPC server is going to be started. 00:09:02.324 17:24:44 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@655 -- # es=1 00:09:02.324 17:24:44 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:09:02.324 17:24:44 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:09:02.324 17:24:44 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:09:02.324 00:09:02.324 real 0m0.076s 00:09:02.324 user 0m0.047s 00:09:02.324 sys 0m0.029s 00:09:02.324 17:24:44 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:02.324 17:24:44 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@10 -- # set +x 00:09:02.324 ************************************ 00:09:02.324 END TEST skip_rpc_with_delay 00:09:02.324 ************************************ 00:09:02.584 17:24:44 skip_rpc -- rpc/skip_rpc.sh@77 -- # uname 00:09:02.584 17:24:44 skip_rpc -- rpc/skip_rpc.sh@77 -- # '[' Linux '!=' FreeBSD ']' 00:09:02.584 17:24:44 skip_rpc -- rpc/skip_rpc.sh@78 -- # run_test exit_on_failed_rpc_init test_exit_on_failed_rpc_init 00:09:02.584 17:24:44 skip_rpc -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:09:02.584 17:24:44 skip_rpc -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:02.584 17:24:44 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:09:02.584 ************************************ 00:09:02.584 START TEST exit_on_failed_rpc_init 00:09:02.584 ************************************ 00:09:02.584 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1129 -- # test_exit_on_failed_rpc_init 00:09:02.584 17:24:44 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@62 -- # local spdk_pid=112829 00:09:02.584 17:24:44 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:09:02.584 17:24:44 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@63 -- # waitforlisten 112829 00:09:02.584 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@835 -- # '[' -z 112829 ']' 00:09:02.584 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:02.584 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:02.584 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:02.584 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:02.584 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:02.584 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:09:02.584 [2024-12-06 17:24:44.264860] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:09:02.584 [2024-12-06 17:24:44.264965] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid112829 ] 00:09:02.584 [2024-12-06 17:24:44.329863] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:02.584 [2024-12-06 17:24:44.373187] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:02.843 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:02.843 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@868 -- # return 0 00:09:02.843 17:24:44 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@65 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:09:02.843 17:24:44 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@67 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x2 00:09:02.843 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@652 -- # local es=0 00:09:02.843 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@654 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x2 00:09:02.843 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@640 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:09:02.843 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:02.843 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:09:02.843 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:02.843 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@646 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:09:02.843 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:02.843 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@646 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:09:02.843 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@646 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt ]] 00:09:02.843 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@655 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x2 00:09:03.102 [2024-12-06 17:24:44.682715] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:09:03.102 [2024-12-06 17:24:44.682811] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid112914 ] 00:09:03.102 [2024-12-06 17:24:44.748307] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:03.102 [2024-12-06 17:24:44.793486] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:09:03.102 [2024-12-06 17:24:44.793615] rpc.c: 180:_spdk_rpc_listen: *ERROR*: RPC Unix domain socket path /var/tmp/spdk.sock in use. Specify another. 00:09:03.102 [2024-12-06 17:24:44.793636] rpc.c: 166:spdk_rpc_initialize: *ERROR*: Unable to start RPC service at /var/tmp/spdk.sock 00:09:03.102 [2024-12-06 17:24:44.793648] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:09:03.102 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@655 -- # es=234 00:09:03.102 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:09:03.102 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@664 -- # es=106 00:09:03.102 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@665 -- # case "$es" in 00:09:03.102 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@672 -- # es=1 00:09:03.102 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:09:03.102 17:24:44 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:09:03.102 17:24:44 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@70 -- # killprocess 112829 00:09:03.102 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@954 -- # '[' -z 112829 ']' 00:09:03.102 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@958 -- # kill -0 112829 00:09:03.102 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@959 -- # uname 00:09:03.102 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:03.102 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 112829 00:09:03.102 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:03.102 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:03.102 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@972 -- # echo 'killing process with pid 112829' 00:09:03.102 killing process with pid 112829 00:09:03.102 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@973 -- # kill 112829 00:09:03.102 17:24:44 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@978 -- # wait 112829 00:09:03.669 00:09:03.669 real 0m1.065s 00:09:03.669 user 0m1.157s 00:09:03.669 sys 0m0.417s 00:09:03.669 17:24:45 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:03.669 17:24:45 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:09:03.669 ************************************ 00:09:03.669 END TEST exit_on_failed_rpc_init 00:09:03.669 ************************************ 00:09:03.669 17:24:45 skip_rpc -- rpc/skip_rpc.sh@81 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/config.json 00:09:03.669 00:09:03.669 real 0m13.386s 00:09:03.669 user 0m12.635s 00:09:03.669 sys 0m1.620s 00:09:03.669 17:24:45 skip_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:03.669 17:24:45 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:09:03.669 ************************************ 00:09:03.669 END TEST skip_rpc 00:09:03.669 ************************************ 00:09:03.669 17:24:45 -- spdk/autotest.sh@158 -- # run_test rpc_client /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_client/rpc_client.sh 00:09:03.669 17:24:45 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:09:03.669 17:24:45 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:03.670 17:24:45 -- common/autotest_common.sh@10 -- # set +x 00:09:03.670 ************************************ 00:09:03.670 START TEST rpc_client 00:09:03.670 ************************************ 00:09:03.670 17:24:45 rpc_client -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_client/rpc_client.sh 00:09:03.670 * Looking for test storage... 00:09:03.670 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_client 00:09:03.670 17:24:45 rpc_client -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:09:03.670 17:24:45 rpc_client -- common/autotest_common.sh@1711 -- # lcov --version 00:09:03.670 17:24:45 rpc_client -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:09:03.670 17:24:45 rpc_client -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:09:03.670 17:24:45 rpc_client -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:09:03.670 17:24:45 rpc_client -- scripts/common.sh@333 -- # local ver1 ver1_l 00:09:03.670 17:24:45 rpc_client -- scripts/common.sh@334 -- # local ver2 ver2_l 00:09:03.670 17:24:45 rpc_client -- scripts/common.sh@336 -- # IFS=.-: 00:09:03.670 17:24:45 rpc_client -- scripts/common.sh@336 -- # read -ra ver1 00:09:03.670 17:24:45 rpc_client -- scripts/common.sh@337 -- # IFS=.-: 00:09:03.670 17:24:45 rpc_client -- scripts/common.sh@337 -- # read -ra ver2 00:09:03.670 17:24:45 rpc_client -- scripts/common.sh@338 -- # local 'op=<' 00:09:03.670 17:24:45 rpc_client -- scripts/common.sh@340 -- # ver1_l=2 00:09:03.670 17:24:45 rpc_client -- scripts/common.sh@341 -- # ver2_l=1 00:09:03.670 17:24:45 rpc_client -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:09:03.670 17:24:45 rpc_client -- scripts/common.sh@344 -- # case "$op" in 00:09:03.670 17:24:45 rpc_client -- scripts/common.sh@345 -- # : 1 00:09:03.670 17:24:45 rpc_client -- scripts/common.sh@364 -- # (( v = 0 )) 00:09:03.670 17:24:45 rpc_client -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:09:03.670 17:24:45 rpc_client -- scripts/common.sh@365 -- # decimal 1 00:09:03.670 17:24:45 rpc_client -- scripts/common.sh@353 -- # local d=1 00:09:03.670 17:24:45 rpc_client -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:09:03.670 17:24:45 rpc_client -- scripts/common.sh@355 -- # echo 1 00:09:03.670 17:24:45 rpc_client -- scripts/common.sh@365 -- # ver1[v]=1 00:09:03.670 17:24:45 rpc_client -- scripts/common.sh@366 -- # decimal 2 00:09:03.670 17:24:45 rpc_client -- scripts/common.sh@353 -- # local d=2 00:09:03.670 17:24:45 rpc_client -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:09:03.670 17:24:45 rpc_client -- scripts/common.sh@355 -- # echo 2 00:09:03.670 17:24:45 rpc_client -- scripts/common.sh@366 -- # ver2[v]=2 00:09:03.670 17:24:45 rpc_client -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:09:03.670 17:24:45 rpc_client -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:09:03.670 17:24:45 rpc_client -- scripts/common.sh@368 -- # return 0 00:09:03.670 17:24:45 rpc_client -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:09:03.670 17:24:45 rpc_client -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:09:03.670 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:03.670 --rc genhtml_branch_coverage=1 00:09:03.670 --rc genhtml_function_coverage=1 00:09:03.670 --rc genhtml_legend=1 00:09:03.670 --rc geninfo_all_blocks=1 00:09:03.670 --rc geninfo_unexecuted_blocks=1 00:09:03.670 00:09:03.670 ' 00:09:03.670 17:24:45 rpc_client -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:09:03.670 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:03.670 --rc genhtml_branch_coverage=1 00:09:03.670 --rc genhtml_function_coverage=1 00:09:03.670 --rc genhtml_legend=1 00:09:03.670 --rc geninfo_all_blocks=1 00:09:03.670 --rc geninfo_unexecuted_blocks=1 00:09:03.670 00:09:03.670 ' 00:09:03.670 17:24:45 rpc_client -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:09:03.670 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:03.670 --rc genhtml_branch_coverage=1 00:09:03.670 --rc genhtml_function_coverage=1 00:09:03.670 --rc genhtml_legend=1 00:09:03.670 --rc geninfo_all_blocks=1 00:09:03.670 --rc geninfo_unexecuted_blocks=1 00:09:03.670 00:09:03.670 ' 00:09:03.670 17:24:45 rpc_client -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:09:03.670 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:03.670 --rc genhtml_branch_coverage=1 00:09:03.670 --rc genhtml_function_coverage=1 00:09:03.670 --rc genhtml_legend=1 00:09:03.670 --rc geninfo_all_blocks=1 00:09:03.670 --rc geninfo_unexecuted_blocks=1 00:09:03.670 00:09:03.670 ' 00:09:03.670 17:24:45 rpc_client -- rpc_client/rpc_client.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_client/rpc_client_test 00:09:03.670 OK 00:09:03.670 17:24:45 rpc_client -- rpc_client/rpc_client.sh@12 -- # trap - SIGINT SIGTERM EXIT 00:09:03.929 00:09:03.929 real 0m0.161s 00:09:03.929 user 0m0.113s 00:09:03.929 sys 0m0.057s 00:09:03.929 17:24:45 rpc_client -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:03.929 17:24:45 rpc_client -- common/autotest_common.sh@10 -- # set +x 00:09:03.929 ************************************ 00:09:03.929 END TEST rpc_client 00:09:03.929 ************************************ 00:09:03.929 17:24:45 -- spdk/autotest.sh@159 -- # run_test json_config /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_config.sh 00:09:03.929 17:24:45 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:09:03.929 17:24:45 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:03.929 17:24:45 -- common/autotest_common.sh@10 -- # set +x 00:09:03.929 ************************************ 00:09:03.929 START TEST json_config 00:09:03.929 ************************************ 00:09:03.929 17:24:45 json_config -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_config.sh 00:09:03.929 17:24:45 json_config -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:09:03.929 17:24:45 json_config -- common/autotest_common.sh@1711 -- # lcov --version 00:09:03.929 17:24:45 json_config -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:09:03.929 17:24:45 json_config -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:09:03.929 17:24:45 json_config -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:09:03.929 17:24:45 json_config -- scripts/common.sh@333 -- # local ver1 ver1_l 00:09:03.929 17:24:45 json_config -- scripts/common.sh@334 -- # local ver2 ver2_l 00:09:03.929 17:24:45 json_config -- scripts/common.sh@336 -- # IFS=.-: 00:09:03.929 17:24:45 json_config -- scripts/common.sh@336 -- # read -ra ver1 00:09:03.929 17:24:45 json_config -- scripts/common.sh@337 -- # IFS=.-: 00:09:03.929 17:24:45 json_config -- scripts/common.sh@337 -- # read -ra ver2 00:09:03.929 17:24:45 json_config -- scripts/common.sh@338 -- # local 'op=<' 00:09:03.929 17:24:45 json_config -- scripts/common.sh@340 -- # ver1_l=2 00:09:03.929 17:24:45 json_config -- scripts/common.sh@341 -- # ver2_l=1 00:09:03.929 17:24:45 json_config -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:09:03.929 17:24:45 json_config -- scripts/common.sh@344 -- # case "$op" in 00:09:03.929 17:24:45 json_config -- scripts/common.sh@345 -- # : 1 00:09:03.929 17:24:45 json_config -- scripts/common.sh@364 -- # (( v = 0 )) 00:09:03.929 17:24:45 json_config -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:09:03.929 17:24:45 json_config -- scripts/common.sh@365 -- # decimal 1 00:09:03.929 17:24:45 json_config -- scripts/common.sh@353 -- # local d=1 00:09:03.929 17:24:45 json_config -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:09:03.929 17:24:45 json_config -- scripts/common.sh@355 -- # echo 1 00:09:03.929 17:24:45 json_config -- scripts/common.sh@365 -- # ver1[v]=1 00:09:03.929 17:24:45 json_config -- scripts/common.sh@366 -- # decimal 2 00:09:03.929 17:24:45 json_config -- scripts/common.sh@353 -- # local d=2 00:09:03.929 17:24:45 json_config -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:09:03.929 17:24:45 json_config -- scripts/common.sh@355 -- # echo 2 00:09:03.929 17:24:45 json_config -- scripts/common.sh@366 -- # ver2[v]=2 00:09:03.929 17:24:45 json_config -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:09:03.929 17:24:45 json_config -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:09:03.929 17:24:45 json_config -- scripts/common.sh@368 -- # return 0 00:09:03.929 17:24:45 json_config -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:09:03.929 17:24:45 json_config -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:09:03.929 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:03.929 --rc genhtml_branch_coverage=1 00:09:03.929 --rc genhtml_function_coverage=1 00:09:03.930 --rc genhtml_legend=1 00:09:03.930 --rc geninfo_all_blocks=1 00:09:03.930 --rc geninfo_unexecuted_blocks=1 00:09:03.930 00:09:03.930 ' 00:09:03.930 17:24:45 json_config -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:09:03.930 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:03.930 --rc genhtml_branch_coverage=1 00:09:03.930 --rc genhtml_function_coverage=1 00:09:03.930 --rc genhtml_legend=1 00:09:03.930 --rc geninfo_all_blocks=1 00:09:03.930 --rc geninfo_unexecuted_blocks=1 00:09:03.930 00:09:03.930 ' 00:09:03.930 17:24:45 json_config -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:09:03.930 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:03.930 --rc genhtml_branch_coverage=1 00:09:03.930 --rc genhtml_function_coverage=1 00:09:03.930 --rc genhtml_legend=1 00:09:03.930 --rc geninfo_all_blocks=1 00:09:03.930 --rc geninfo_unexecuted_blocks=1 00:09:03.930 00:09:03.930 ' 00:09:03.930 17:24:45 json_config -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:09:03.930 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:03.930 --rc genhtml_branch_coverage=1 00:09:03.930 --rc genhtml_function_coverage=1 00:09:03.930 --rc genhtml_legend=1 00:09:03.930 --rc geninfo_all_blocks=1 00:09:03.930 --rc geninfo_unexecuted_blocks=1 00:09:03.930 00:09:03.930 ' 00:09:03.930 17:24:45 json_config -- json_config/json_config.sh@8 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:09:03.930 17:24:45 json_config -- nvmf/common.sh@7 -- # uname -s 00:09:03.930 17:24:45 json_config -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:09:03.930 17:24:45 json_config -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:09:03.930 17:24:45 json_config -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:09:03.930 17:24:45 json_config -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:09:03.930 17:24:45 json_config -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:09:03.930 17:24:45 json_config -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:09:03.930 17:24:45 json_config -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:09:03.930 17:24:45 json_config -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:09:03.930 17:24:45 json_config -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:09:03.930 17:24:45 json_config -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:09:03.930 17:24:45 json_config -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:09:03.930 17:24:45 json_config -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:09:03.930 17:24:45 json_config -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:09:03.930 17:24:45 json_config -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:09:03.930 17:24:45 json_config -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:09:03.930 17:24:45 json_config -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:09:03.930 17:24:45 json_config -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:09:03.930 17:24:45 json_config -- scripts/common.sh@15 -- # shopt -s extglob 00:09:03.930 17:24:45 json_config -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:09:03.930 17:24:45 json_config -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:09:03.930 17:24:45 json_config -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:09:03.930 17:24:45 json_config -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:03.930 17:24:45 json_config -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:03.930 17:24:45 json_config -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:03.930 17:24:45 json_config -- paths/export.sh@5 -- # export PATH 00:09:03.930 17:24:45 json_config -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:03.930 17:24:45 json_config -- nvmf/common.sh@51 -- # : 0 00:09:03.930 17:24:45 json_config -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:09:03.930 17:24:45 json_config -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:09:03.930 17:24:45 json_config -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:09:03.930 17:24:45 json_config -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:09:03.930 17:24:45 json_config -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:09:03.930 17:24:45 json_config -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:09:03.930 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:09:03.930 17:24:45 json_config -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:09:03.930 17:24:45 json_config -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:09:03.930 17:24:45 json_config -- nvmf/common.sh@55 -- # have_pci_nics=0 00:09:03.930 17:24:45 json_config -- json_config/json_config.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/common.sh 00:09:03.930 17:24:45 json_config -- json_config/json_config.sh@11 -- # [[ 0 -eq 1 ]] 00:09:03.930 17:24:45 json_config -- json_config/json_config.sh@15 -- # [[ 0 -ne 1 ]] 00:09:03.930 17:24:45 json_config -- json_config/json_config.sh@15 -- # [[ 0 -eq 1 ]] 00:09:03.930 17:24:45 json_config -- json_config/json_config.sh@26 -- # (( SPDK_TEST_BLOCKDEV + SPDK_TEST_ISCSI + SPDK_TEST_NVMF + SPDK_TEST_VHOST + SPDK_TEST_VHOST_INIT + SPDK_TEST_RBD == 0 )) 00:09:03.930 17:24:45 json_config -- json_config/json_config.sh@31 -- # app_pid=(['target']='' ['initiator']='') 00:09:03.930 17:24:45 json_config -- json_config/json_config.sh@31 -- # declare -A app_pid 00:09:03.930 17:24:45 json_config -- json_config/json_config.sh@32 -- # app_socket=(['target']='/var/tmp/spdk_tgt.sock' ['initiator']='/var/tmp/spdk_initiator.sock') 00:09:03.930 17:24:45 json_config -- json_config/json_config.sh@32 -- # declare -A app_socket 00:09:03.930 17:24:45 json_config -- json_config/json_config.sh@33 -- # app_params=(['target']='-m 0x1 -s 1024' ['initiator']='-m 0x2 -g -u -s 1024') 00:09:03.930 17:24:45 json_config -- json_config/json_config.sh@33 -- # declare -A app_params 00:09:03.930 17:24:45 json_config -- json_config/json_config.sh@34 -- # configs_path=(['target']='/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json' ['initiator']='/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_initiator_config.json') 00:09:03.930 17:24:45 json_config -- json_config/json_config.sh@34 -- # declare -A configs_path 00:09:03.930 17:24:45 json_config -- json_config/json_config.sh@40 -- # last_event_id=0 00:09:03.930 17:24:45 json_config -- json_config/json_config.sh@362 -- # trap 'on_error_exit "${FUNCNAME}" "${LINENO}"' ERR 00:09:03.930 17:24:45 json_config -- json_config/json_config.sh@363 -- # echo 'INFO: JSON configuration test init' 00:09:03.930 INFO: JSON configuration test init 00:09:03.930 17:24:45 json_config -- json_config/json_config.sh@364 -- # json_config_test_init 00:09:03.930 17:24:45 json_config -- json_config/json_config.sh@269 -- # timing_enter json_config_test_init 00:09:03.930 17:24:45 json_config -- common/autotest_common.sh@726 -- # xtrace_disable 00:09:03.930 17:24:45 json_config -- common/autotest_common.sh@10 -- # set +x 00:09:03.930 17:24:45 json_config -- json_config/json_config.sh@270 -- # timing_enter json_config_setup_target 00:09:03.930 17:24:45 json_config -- common/autotest_common.sh@726 -- # xtrace_disable 00:09:03.930 17:24:45 json_config -- common/autotest_common.sh@10 -- # set +x 00:09:03.930 17:24:45 json_config -- json_config/json_config.sh@272 -- # json_config_test_start_app target --wait-for-rpc 00:09:03.930 17:24:45 json_config -- json_config/common.sh@9 -- # local app=target 00:09:03.930 17:24:45 json_config -- json_config/common.sh@10 -- # shift 00:09:03.931 17:24:45 json_config -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:09:03.931 17:24:45 json_config -- json_config/common.sh@13 -- # [[ -z '' ]] 00:09:03.931 17:24:45 json_config -- json_config/common.sh@15 -- # local app_extra_params= 00:09:03.931 17:24:45 json_config -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:09:03.931 17:24:45 json_config -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:09:03.931 17:24:45 json_config -- json_config/common.sh@22 -- # app_pid["$app"]=113121 00:09:03.931 17:24:45 json_config -- json_config/common.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --wait-for-rpc 00:09:03.931 17:24:45 json_config -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:09:03.931 Waiting for target to run... 00:09:03.931 17:24:45 json_config -- json_config/common.sh@25 -- # waitforlisten 113121 /var/tmp/spdk_tgt.sock 00:09:03.931 17:24:45 json_config -- common/autotest_common.sh@835 -- # '[' -z 113121 ']' 00:09:03.931 17:24:45 json_config -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:09:03.931 17:24:45 json_config -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:03.931 17:24:45 json_config -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:09:03.931 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:09:03.931 17:24:45 json_config -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:03.931 17:24:45 json_config -- common/autotest_common.sh@10 -- # set +x 00:09:03.931 [2024-12-06 17:24:45.757234] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:09:03.931 [2024-12-06 17:24:45.757315] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid113121 ] 00:09:04.499 [2024-12-06 17:24:46.265183] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:04.499 [2024-12-06 17:24:46.306562] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:05.065 17:24:46 json_config -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:05.065 17:24:46 json_config -- common/autotest_common.sh@868 -- # return 0 00:09:05.065 17:24:46 json_config -- json_config/common.sh@26 -- # echo '' 00:09:05.065 00:09:05.065 17:24:46 json_config -- json_config/json_config.sh@276 -- # create_accel_config 00:09:05.065 17:24:46 json_config -- json_config/json_config.sh@100 -- # timing_enter create_accel_config 00:09:05.065 17:24:46 json_config -- common/autotest_common.sh@726 -- # xtrace_disable 00:09:05.065 17:24:46 json_config -- common/autotest_common.sh@10 -- # set +x 00:09:05.065 17:24:46 json_config -- json_config/json_config.sh@102 -- # [[ 0 -eq 1 ]] 00:09:05.065 17:24:46 json_config -- json_config/json_config.sh@108 -- # timing_exit create_accel_config 00:09:05.065 17:24:46 json_config -- common/autotest_common.sh@732 -- # xtrace_disable 00:09:05.065 17:24:46 json_config -- common/autotest_common.sh@10 -- # set +x 00:09:05.065 17:24:46 json_config -- json_config/json_config.sh@280 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh --json-with-subsystems 00:09:05.065 17:24:46 json_config -- json_config/json_config.sh@281 -- # tgt_rpc load_config 00:09:05.065 17:24:46 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock load_config 00:09:08.357 17:24:49 json_config -- json_config/json_config.sh@283 -- # tgt_check_notification_types 00:09:08.357 17:24:49 json_config -- json_config/json_config.sh@43 -- # timing_enter tgt_check_notification_types 00:09:08.357 17:24:49 json_config -- common/autotest_common.sh@726 -- # xtrace_disable 00:09:08.357 17:24:49 json_config -- common/autotest_common.sh@10 -- # set +x 00:09:08.357 17:24:49 json_config -- json_config/json_config.sh@45 -- # local ret=0 00:09:08.357 17:24:49 json_config -- json_config/json_config.sh@46 -- # enabled_types=('bdev_register' 'bdev_unregister') 00:09:08.357 17:24:49 json_config -- json_config/json_config.sh@46 -- # local enabled_types 00:09:08.357 17:24:49 json_config -- json_config/json_config.sh@47 -- # [[ y == y ]] 00:09:08.357 17:24:49 json_config -- json_config/json_config.sh@48 -- # enabled_types+=("fsdev_register" "fsdev_unregister") 00:09:08.357 17:24:49 json_config -- json_config/json_config.sh@51 -- # tgt_rpc notify_get_types 00:09:08.357 17:24:49 json_config -- json_config/json_config.sh@51 -- # jq -r '.[]' 00:09:08.357 17:24:49 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock notify_get_types 00:09:08.614 17:24:50 json_config -- json_config/json_config.sh@51 -- # get_types=('fsdev_register' 'fsdev_unregister' 'bdev_register' 'bdev_unregister') 00:09:08.614 17:24:50 json_config -- json_config/json_config.sh@51 -- # local get_types 00:09:08.614 17:24:50 json_config -- json_config/json_config.sh@53 -- # local type_diff 00:09:08.614 17:24:50 json_config -- json_config/json_config.sh@54 -- # echo bdev_register bdev_unregister fsdev_register fsdev_unregister fsdev_register fsdev_unregister bdev_register bdev_unregister 00:09:08.614 17:24:50 json_config -- json_config/json_config.sh@54 -- # tr ' ' '\n' 00:09:08.614 17:24:50 json_config -- json_config/json_config.sh@54 -- # sort 00:09:08.614 17:24:50 json_config -- json_config/json_config.sh@54 -- # uniq -u 00:09:08.614 17:24:50 json_config -- json_config/json_config.sh@54 -- # type_diff= 00:09:08.614 17:24:50 json_config -- json_config/json_config.sh@56 -- # [[ -n '' ]] 00:09:08.614 17:24:50 json_config -- json_config/json_config.sh@61 -- # timing_exit tgt_check_notification_types 00:09:08.614 17:24:50 json_config -- common/autotest_common.sh@732 -- # xtrace_disable 00:09:08.614 17:24:50 json_config -- common/autotest_common.sh@10 -- # set +x 00:09:08.614 17:24:50 json_config -- json_config/json_config.sh@62 -- # return 0 00:09:08.614 17:24:50 json_config -- json_config/json_config.sh@285 -- # [[ 0 -eq 1 ]] 00:09:08.614 17:24:50 json_config -- json_config/json_config.sh@289 -- # [[ 0 -eq 1 ]] 00:09:08.614 17:24:50 json_config -- json_config/json_config.sh@293 -- # [[ 0 -eq 1 ]] 00:09:08.614 17:24:50 json_config -- json_config/json_config.sh@297 -- # [[ 1 -eq 1 ]] 00:09:08.614 17:24:50 json_config -- json_config/json_config.sh@298 -- # create_nvmf_subsystem_config 00:09:08.614 17:24:50 json_config -- json_config/json_config.sh@237 -- # timing_enter create_nvmf_subsystem_config 00:09:08.614 17:24:50 json_config -- common/autotest_common.sh@726 -- # xtrace_disable 00:09:08.614 17:24:50 json_config -- common/autotest_common.sh@10 -- # set +x 00:09:08.614 17:24:50 json_config -- json_config/json_config.sh@239 -- # NVMF_FIRST_TARGET_IP=127.0.0.1 00:09:08.614 17:24:50 json_config -- json_config/json_config.sh@240 -- # [[ tcp == \r\d\m\a ]] 00:09:08.614 17:24:50 json_config -- json_config/json_config.sh@244 -- # [[ -z 127.0.0.1 ]] 00:09:08.614 17:24:50 json_config -- json_config/json_config.sh@249 -- # tgt_rpc bdev_malloc_create 8 512 --name MallocForNvmf0 00:09:08.614 17:24:50 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_create 8 512 --name MallocForNvmf0 00:09:08.871 MallocForNvmf0 00:09:08.871 17:24:50 json_config -- json_config/json_config.sh@250 -- # tgt_rpc bdev_malloc_create 4 1024 --name MallocForNvmf1 00:09:08.871 17:24:50 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_create 4 1024 --name MallocForNvmf1 00:09:09.128 MallocForNvmf1 00:09:09.128 17:24:50 json_config -- json_config/json_config.sh@252 -- # tgt_rpc nvmf_create_transport -t tcp -u 8192 -c 0 00:09:09.128 17:24:50 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_create_transport -t tcp -u 8192 -c 0 00:09:09.386 [2024-12-06 17:24:51.038482] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:09:09.386 17:24:51 json_config -- json_config/json_config.sh@253 -- # tgt_rpc nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:09:09.386 17:24:51 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:09:09.643 17:24:51 json_config -- json_config/json_config.sh@254 -- # tgt_rpc nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 MallocForNvmf0 00:09:09.643 17:24:51 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 MallocForNvmf0 00:09:09.915 17:24:51 json_config -- json_config/json_config.sh@255 -- # tgt_rpc nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 MallocForNvmf1 00:09:09.915 17:24:51 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 MallocForNvmf1 00:09:10.172 17:24:51 json_config -- json_config/json_config.sh@256 -- # tgt_rpc nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 127.0.0.1 -s 4420 00:09:10.172 17:24:51 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 127.0.0.1 -s 4420 00:09:10.429 [2024-12-06 17:24:52.089883] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4420 *** 00:09:10.429 17:24:52 json_config -- json_config/json_config.sh@258 -- # timing_exit create_nvmf_subsystem_config 00:09:10.429 17:24:52 json_config -- common/autotest_common.sh@732 -- # xtrace_disable 00:09:10.429 17:24:52 json_config -- common/autotest_common.sh@10 -- # set +x 00:09:10.429 17:24:52 json_config -- json_config/json_config.sh@300 -- # timing_exit json_config_setup_target 00:09:10.429 17:24:52 json_config -- common/autotest_common.sh@732 -- # xtrace_disable 00:09:10.429 17:24:52 json_config -- common/autotest_common.sh@10 -- # set +x 00:09:10.429 17:24:52 json_config -- json_config/json_config.sh@302 -- # [[ 0 -eq 1 ]] 00:09:10.429 17:24:52 json_config -- json_config/json_config.sh@307 -- # tgt_rpc bdev_malloc_create 8 512 --name MallocBdevForConfigChangeCheck 00:09:10.429 17:24:52 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_create 8 512 --name MallocBdevForConfigChangeCheck 00:09:10.687 MallocBdevForConfigChangeCheck 00:09:10.687 17:24:52 json_config -- json_config/json_config.sh@309 -- # timing_exit json_config_test_init 00:09:10.687 17:24:52 json_config -- common/autotest_common.sh@732 -- # xtrace_disable 00:09:10.687 17:24:52 json_config -- common/autotest_common.sh@10 -- # set +x 00:09:10.687 17:24:52 json_config -- json_config/json_config.sh@366 -- # tgt_rpc save_config 00:09:10.687 17:24:52 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock save_config 00:09:11.251 17:24:52 json_config -- json_config/json_config.sh@368 -- # echo 'INFO: shutting down applications...' 00:09:11.251 INFO: shutting down applications... 00:09:11.251 17:24:52 json_config -- json_config/json_config.sh@369 -- # [[ 0 -eq 1 ]] 00:09:11.251 17:24:52 json_config -- json_config/json_config.sh@375 -- # json_config_clear target 00:09:11.251 17:24:52 json_config -- json_config/json_config.sh@339 -- # [[ -n 22 ]] 00:09:11.251 17:24:52 json_config -- json_config/json_config.sh@340 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/clear_config.py -s /var/tmp/spdk_tgt.sock clear_config 00:09:13.150 Calling clear_iscsi_subsystem 00:09:13.150 Calling clear_nvmf_subsystem 00:09:13.150 Calling clear_nbd_subsystem 00:09:13.150 Calling clear_ublk_subsystem 00:09:13.150 Calling clear_vhost_blk_subsystem 00:09:13.150 Calling clear_vhost_scsi_subsystem 00:09:13.150 Calling clear_bdev_subsystem 00:09:13.150 17:24:54 json_config -- json_config/json_config.sh@344 -- # local config_filter=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py 00:09:13.150 17:24:54 json_config -- json_config/json_config.sh@350 -- # count=100 00:09:13.150 17:24:54 json_config -- json_config/json_config.sh@351 -- # '[' 100 -gt 0 ']' 00:09:13.150 17:24:54 json_config -- json_config/json_config.sh@352 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock save_config 00:09:13.150 17:24:54 json_config -- json_config/json_config.sh@352 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method delete_global_parameters 00:09:13.150 17:24:54 json_config -- json_config/json_config.sh@352 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method check_empty 00:09:13.150 17:24:54 json_config -- json_config/json_config.sh@352 -- # break 00:09:13.150 17:24:54 json_config -- json_config/json_config.sh@357 -- # '[' 100 -eq 0 ']' 00:09:13.150 17:24:54 json_config -- json_config/json_config.sh@376 -- # json_config_test_shutdown_app target 00:09:13.150 17:24:54 json_config -- json_config/common.sh@31 -- # local app=target 00:09:13.150 17:24:54 json_config -- json_config/common.sh@34 -- # [[ -n 22 ]] 00:09:13.150 17:24:54 json_config -- json_config/common.sh@35 -- # [[ -n 113121 ]] 00:09:13.150 17:24:54 json_config -- json_config/common.sh@38 -- # kill -SIGINT 113121 00:09:13.150 17:24:54 json_config -- json_config/common.sh@40 -- # (( i = 0 )) 00:09:13.150 17:24:54 json_config -- json_config/common.sh@40 -- # (( i < 30 )) 00:09:13.150 17:24:54 json_config -- json_config/common.sh@41 -- # kill -0 113121 00:09:13.150 17:24:54 json_config -- json_config/common.sh@45 -- # sleep 0.5 00:09:13.722 17:24:55 json_config -- json_config/common.sh@40 -- # (( i++ )) 00:09:13.722 17:24:55 json_config -- json_config/common.sh@40 -- # (( i < 30 )) 00:09:13.722 17:24:55 json_config -- json_config/common.sh@41 -- # kill -0 113121 00:09:13.722 17:24:55 json_config -- json_config/common.sh@42 -- # app_pid["$app"]= 00:09:13.722 17:24:55 json_config -- json_config/common.sh@43 -- # break 00:09:13.722 17:24:55 json_config -- json_config/common.sh@48 -- # [[ -n '' ]] 00:09:13.722 17:24:55 json_config -- json_config/common.sh@53 -- # echo 'SPDK target shutdown done' 00:09:13.722 SPDK target shutdown done 00:09:13.722 17:24:55 json_config -- json_config/json_config.sh@378 -- # echo 'INFO: relaunching applications...' 00:09:13.722 INFO: relaunching applications... 00:09:13.722 17:24:55 json_config -- json_config/json_config.sh@379 -- # json_config_test_start_app target --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:09:13.722 17:24:55 json_config -- json_config/common.sh@9 -- # local app=target 00:09:13.722 17:24:55 json_config -- json_config/common.sh@10 -- # shift 00:09:13.722 17:24:55 json_config -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:09:13.722 17:24:55 json_config -- json_config/common.sh@13 -- # [[ -z '' ]] 00:09:13.722 17:24:55 json_config -- json_config/common.sh@15 -- # local app_extra_params= 00:09:13.722 17:24:55 json_config -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:09:13.722 17:24:55 json_config -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:09:13.722 17:24:55 json_config -- json_config/common.sh@22 -- # app_pid["$app"]=114419 00:09:13.722 17:24:55 json_config -- json_config/common.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:09:13.722 17:24:55 json_config -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:09:13.722 Waiting for target to run... 00:09:13.722 17:24:55 json_config -- json_config/common.sh@25 -- # waitforlisten 114419 /var/tmp/spdk_tgt.sock 00:09:13.722 17:24:55 json_config -- common/autotest_common.sh@835 -- # '[' -z 114419 ']' 00:09:13.722 17:24:55 json_config -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:09:13.722 17:24:55 json_config -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:13.722 17:24:55 json_config -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:09:13.722 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:09:13.722 17:24:55 json_config -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:13.722 17:24:55 json_config -- common/autotest_common.sh@10 -- # set +x 00:09:13.722 [2024-12-06 17:24:55.466341] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:09:13.722 [2024-12-06 17:24:55.466436] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid114419 ] 00:09:14.290 [2024-12-06 17:24:56.012294] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:14.290 [2024-12-06 17:24:56.053468] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:17.582 [2024-12-06 17:24:59.103239] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:09:17.582 [2024-12-06 17:24:59.135643] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4420 *** 00:09:17.582 17:24:59 json_config -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:17.582 17:24:59 json_config -- common/autotest_common.sh@868 -- # return 0 00:09:17.582 17:24:59 json_config -- json_config/common.sh@26 -- # echo '' 00:09:17.582 00:09:17.582 17:24:59 json_config -- json_config/json_config.sh@380 -- # [[ 0 -eq 1 ]] 00:09:17.582 17:24:59 json_config -- json_config/json_config.sh@384 -- # echo 'INFO: Checking if target configuration is the same...' 00:09:17.582 INFO: Checking if target configuration is the same... 00:09:17.582 17:24:59 json_config -- json_config/json_config.sh@385 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_diff.sh /dev/fd/62 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:09:17.582 17:24:59 json_config -- json_config/json_config.sh@385 -- # tgt_rpc save_config 00:09:17.582 17:24:59 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock save_config 00:09:17.582 + '[' 2 -ne 2 ']' 00:09:17.582 +++ dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_diff.sh 00:09:17.582 ++ readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/../.. 00:09:17.582 + rootdir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:09:17.582 +++ basename /dev/fd/62 00:09:17.582 ++ mktemp /tmp/62.XXX 00:09:17.582 + tmp_file_1=/tmp/62.jB4 00:09:17.582 +++ basename /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:09:17.582 ++ mktemp /tmp/spdk_tgt_config.json.XXX 00:09:17.582 + tmp_file_2=/tmp/spdk_tgt_config.json.rqD 00:09:17.582 + ret=0 00:09:17.582 + /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method sort 00:09:17.841 + /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method sort 00:09:17.841 + diff -u /tmp/62.jB4 /tmp/spdk_tgt_config.json.rqD 00:09:17.841 + echo 'INFO: JSON config files are the same' 00:09:17.841 INFO: JSON config files are the same 00:09:17.841 + rm /tmp/62.jB4 /tmp/spdk_tgt_config.json.rqD 00:09:17.841 + exit 0 00:09:17.841 17:24:59 json_config -- json_config/json_config.sh@386 -- # [[ 0 -eq 1 ]] 00:09:17.841 17:24:59 json_config -- json_config/json_config.sh@391 -- # echo 'INFO: changing configuration and checking if this can be detected...' 00:09:17.841 INFO: changing configuration and checking if this can be detected... 00:09:17.841 17:24:59 json_config -- json_config/json_config.sh@393 -- # tgt_rpc bdev_malloc_delete MallocBdevForConfigChangeCheck 00:09:17.841 17:24:59 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_delete MallocBdevForConfigChangeCheck 00:09:18.100 17:24:59 json_config -- json_config/json_config.sh@394 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_diff.sh /dev/fd/62 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:09:18.100 17:24:59 json_config -- json_config/json_config.sh@394 -- # tgt_rpc save_config 00:09:18.100 17:24:59 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock save_config 00:09:18.100 + '[' 2 -ne 2 ']' 00:09:18.100 +++ dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_diff.sh 00:09:18.100 ++ readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/../.. 00:09:18.100 + rootdir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:09:18.100 +++ basename /dev/fd/62 00:09:18.100 ++ mktemp /tmp/62.XXX 00:09:18.100 + tmp_file_1=/tmp/62.iK5 00:09:18.100 +++ basename /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:09:18.100 ++ mktemp /tmp/spdk_tgt_config.json.XXX 00:09:18.100 + tmp_file_2=/tmp/spdk_tgt_config.json.upr 00:09:18.100 + ret=0 00:09:18.100 + /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method sort 00:09:18.668 + /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method sort 00:09:18.668 + diff -u /tmp/62.iK5 /tmp/spdk_tgt_config.json.upr 00:09:18.668 + ret=1 00:09:18.668 + echo '=== Start of file: /tmp/62.iK5 ===' 00:09:18.668 + cat /tmp/62.iK5 00:09:18.668 + echo '=== End of file: /tmp/62.iK5 ===' 00:09:18.668 + echo '' 00:09:18.668 + echo '=== Start of file: /tmp/spdk_tgt_config.json.upr ===' 00:09:18.668 + cat /tmp/spdk_tgt_config.json.upr 00:09:18.668 + echo '=== End of file: /tmp/spdk_tgt_config.json.upr ===' 00:09:18.668 + echo '' 00:09:18.668 + rm /tmp/62.iK5 /tmp/spdk_tgt_config.json.upr 00:09:18.668 + exit 1 00:09:18.668 17:25:00 json_config -- json_config/json_config.sh@398 -- # echo 'INFO: configuration change detected.' 00:09:18.668 INFO: configuration change detected. 00:09:18.668 17:25:00 json_config -- json_config/json_config.sh@401 -- # json_config_test_fini 00:09:18.668 17:25:00 json_config -- json_config/json_config.sh@313 -- # timing_enter json_config_test_fini 00:09:18.668 17:25:00 json_config -- common/autotest_common.sh@726 -- # xtrace_disable 00:09:18.668 17:25:00 json_config -- common/autotest_common.sh@10 -- # set +x 00:09:18.668 17:25:00 json_config -- json_config/json_config.sh@314 -- # local ret=0 00:09:18.668 17:25:00 json_config -- json_config/json_config.sh@316 -- # [[ -n '' ]] 00:09:18.668 17:25:00 json_config -- json_config/json_config.sh@324 -- # [[ -n 114419 ]] 00:09:18.668 17:25:00 json_config -- json_config/json_config.sh@327 -- # cleanup_bdev_subsystem_config 00:09:18.668 17:25:00 json_config -- json_config/json_config.sh@191 -- # timing_enter cleanup_bdev_subsystem_config 00:09:18.668 17:25:00 json_config -- common/autotest_common.sh@726 -- # xtrace_disable 00:09:18.668 17:25:00 json_config -- common/autotest_common.sh@10 -- # set +x 00:09:18.668 17:25:00 json_config -- json_config/json_config.sh@193 -- # [[ 0 -eq 1 ]] 00:09:18.668 17:25:00 json_config -- json_config/json_config.sh@200 -- # uname -s 00:09:18.668 17:25:00 json_config -- json_config/json_config.sh@200 -- # [[ Linux = Linux ]] 00:09:18.668 17:25:00 json_config -- json_config/json_config.sh@201 -- # rm -f /sample_aio 00:09:18.668 17:25:00 json_config -- json_config/json_config.sh@204 -- # [[ 0 -eq 1 ]] 00:09:18.668 17:25:00 json_config -- json_config/json_config.sh@208 -- # timing_exit cleanup_bdev_subsystem_config 00:09:18.668 17:25:00 json_config -- common/autotest_common.sh@732 -- # xtrace_disable 00:09:18.668 17:25:00 json_config -- common/autotest_common.sh@10 -- # set +x 00:09:18.668 17:25:00 json_config -- json_config/json_config.sh@330 -- # killprocess 114419 00:09:18.668 17:25:00 json_config -- common/autotest_common.sh@954 -- # '[' -z 114419 ']' 00:09:18.668 17:25:00 json_config -- common/autotest_common.sh@958 -- # kill -0 114419 00:09:18.668 17:25:00 json_config -- common/autotest_common.sh@959 -- # uname 00:09:18.668 17:25:00 json_config -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:18.668 17:25:00 json_config -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 114419 00:09:18.668 17:25:00 json_config -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:18.668 17:25:00 json_config -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:18.668 17:25:00 json_config -- common/autotest_common.sh@972 -- # echo 'killing process with pid 114419' 00:09:18.668 killing process with pid 114419 00:09:18.668 17:25:00 json_config -- common/autotest_common.sh@973 -- # kill 114419 00:09:18.668 17:25:00 json_config -- common/autotest_common.sh@978 -- # wait 114419 00:09:20.571 17:25:02 json_config -- json_config/json_config.sh@333 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_initiator_config.json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:09:20.571 17:25:02 json_config -- json_config/json_config.sh@334 -- # timing_exit json_config_test_fini 00:09:20.571 17:25:02 json_config -- common/autotest_common.sh@732 -- # xtrace_disable 00:09:20.571 17:25:02 json_config -- common/autotest_common.sh@10 -- # set +x 00:09:20.571 17:25:02 json_config -- json_config/json_config.sh@335 -- # return 0 00:09:20.571 17:25:02 json_config -- json_config/json_config.sh@403 -- # echo 'INFO: Success' 00:09:20.571 INFO: Success 00:09:20.571 00:09:20.571 real 0m16.489s 00:09:20.571 user 0m18.434s 00:09:20.571 sys 0m2.256s 00:09:20.571 17:25:02 json_config -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:20.571 17:25:02 json_config -- common/autotest_common.sh@10 -- # set +x 00:09:20.571 ************************************ 00:09:20.571 END TEST json_config 00:09:20.571 ************************************ 00:09:20.571 17:25:02 -- spdk/autotest.sh@160 -- # run_test json_config_extra_key /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_config_extra_key.sh 00:09:20.571 17:25:02 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:09:20.571 17:25:02 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:20.571 17:25:02 -- common/autotest_common.sh@10 -- # set +x 00:09:20.571 ************************************ 00:09:20.571 START TEST json_config_extra_key 00:09:20.571 ************************************ 00:09:20.571 17:25:02 json_config_extra_key -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_config_extra_key.sh 00:09:20.571 17:25:02 json_config_extra_key -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:09:20.571 17:25:02 json_config_extra_key -- common/autotest_common.sh@1711 -- # lcov --version 00:09:20.571 17:25:02 json_config_extra_key -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:09:20.571 17:25:02 json_config_extra_key -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:09:20.571 17:25:02 json_config_extra_key -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:09:20.571 17:25:02 json_config_extra_key -- scripts/common.sh@333 -- # local ver1 ver1_l 00:09:20.571 17:25:02 json_config_extra_key -- scripts/common.sh@334 -- # local ver2 ver2_l 00:09:20.571 17:25:02 json_config_extra_key -- scripts/common.sh@336 -- # IFS=.-: 00:09:20.571 17:25:02 json_config_extra_key -- scripts/common.sh@336 -- # read -ra ver1 00:09:20.571 17:25:02 json_config_extra_key -- scripts/common.sh@337 -- # IFS=.-: 00:09:20.571 17:25:02 json_config_extra_key -- scripts/common.sh@337 -- # read -ra ver2 00:09:20.571 17:25:02 json_config_extra_key -- scripts/common.sh@338 -- # local 'op=<' 00:09:20.571 17:25:02 json_config_extra_key -- scripts/common.sh@340 -- # ver1_l=2 00:09:20.571 17:25:02 json_config_extra_key -- scripts/common.sh@341 -- # ver2_l=1 00:09:20.571 17:25:02 json_config_extra_key -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:09:20.571 17:25:02 json_config_extra_key -- scripts/common.sh@344 -- # case "$op" in 00:09:20.571 17:25:02 json_config_extra_key -- scripts/common.sh@345 -- # : 1 00:09:20.571 17:25:02 json_config_extra_key -- scripts/common.sh@364 -- # (( v = 0 )) 00:09:20.571 17:25:02 json_config_extra_key -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:09:20.571 17:25:02 json_config_extra_key -- scripts/common.sh@365 -- # decimal 1 00:09:20.571 17:25:02 json_config_extra_key -- scripts/common.sh@353 -- # local d=1 00:09:20.571 17:25:02 json_config_extra_key -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:09:20.571 17:25:02 json_config_extra_key -- scripts/common.sh@355 -- # echo 1 00:09:20.571 17:25:02 json_config_extra_key -- scripts/common.sh@365 -- # ver1[v]=1 00:09:20.571 17:25:02 json_config_extra_key -- scripts/common.sh@366 -- # decimal 2 00:09:20.571 17:25:02 json_config_extra_key -- scripts/common.sh@353 -- # local d=2 00:09:20.571 17:25:02 json_config_extra_key -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:09:20.571 17:25:02 json_config_extra_key -- scripts/common.sh@355 -- # echo 2 00:09:20.571 17:25:02 json_config_extra_key -- scripts/common.sh@366 -- # ver2[v]=2 00:09:20.571 17:25:02 json_config_extra_key -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:09:20.571 17:25:02 json_config_extra_key -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:09:20.571 17:25:02 json_config_extra_key -- scripts/common.sh@368 -- # return 0 00:09:20.571 17:25:02 json_config_extra_key -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:09:20.571 17:25:02 json_config_extra_key -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:09:20.571 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:20.571 --rc genhtml_branch_coverage=1 00:09:20.571 --rc genhtml_function_coverage=1 00:09:20.571 --rc genhtml_legend=1 00:09:20.571 --rc geninfo_all_blocks=1 00:09:20.571 --rc geninfo_unexecuted_blocks=1 00:09:20.571 00:09:20.571 ' 00:09:20.571 17:25:02 json_config_extra_key -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:09:20.571 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:20.571 --rc genhtml_branch_coverage=1 00:09:20.571 --rc genhtml_function_coverage=1 00:09:20.571 --rc genhtml_legend=1 00:09:20.571 --rc geninfo_all_blocks=1 00:09:20.571 --rc geninfo_unexecuted_blocks=1 00:09:20.571 00:09:20.571 ' 00:09:20.571 17:25:02 json_config_extra_key -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:09:20.571 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:20.571 --rc genhtml_branch_coverage=1 00:09:20.571 --rc genhtml_function_coverage=1 00:09:20.571 --rc genhtml_legend=1 00:09:20.571 --rc geninfo_all_blocks=1 00:09:20.571 --rc geninfo_unexecuted_blocks=1 00:09:20.571 00:09:20.571 ' 00:09:20.571 17:25:02 json_config_extra_key -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:09:20.571 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:20.571 --rc genhtml_branch_coverage=1 00:09:20.571 --rc genhtml_function_coverage=1 00:09:20.571 --rc genhtml_legend=1 00:09:20.571 --rc geninfo_all_blocks=1 00:09:20.571 --rc geninfo_unexecuted_blocks=1 00:09:20.571 00:09:20.571 ' 00:09:20.572 17:25:02 json_config_extra_key -- json_config/json_config_extra_key.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:09:20.572 17:25:02 json_config_extra_key -- nvmf/common.sh@7 -- # uname -s 00:09:20.572 17:25:02 json_config_extra_key -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:09:20.572 17:25:02 json_config_extra_key -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:09:20.572 17:25:02 json_config_extra_key -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:09:20.572 17:25:02 json_config_extra_key -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:09:20.572 17:25:02 json_config_extra_key -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:09:20.572 17:25:02 json_config_extra_key -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:09:20.572 17:25:02 json_config_extra_key -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:09:20.572 17:25:02 json_config_extra_key -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:09:20.572 17:25:02 json_config_extra_key -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:09:20.572 17:25:02 json_config_extra_key -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:09:20.572 17:25:02 json_config_extra_key -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:09:20.572 17:25:02 json_config_extra_key -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:09:20.572 17:25:02 json_config_extra_key -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:09:20.572 17:25:02 json_config_extra_key -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:09:20.572 17:25:02 json_config_extra_key -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:09:20.572 17:25:02 json_config_extra_key -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:09:20.572 17:25:02 json_config_extra_key -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:09:20.572 17:25:02 json_config_extra_key -- scripts/common.sh@15 -- # shopt -s extglob 00:09:20.572 17:25:02 json_config_extra_key -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:09:20.572 17:25:02 json_config_extra_key -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:09:20.572 17:25:02 json_config_extra_key -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:09:20.572 17:25:02 json_config_extra_key -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:20.572 17:25:02 json_config_extra_key -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:20.572 17:25:02 json_config_extra_key -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:20.572 17:25:02 json_config_extra_key -- paths/export.sh@5 -- # export PATH 00:09:20.572 17:25:02 json_config_extra_key -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:20.572 17:25:02 json_config_extra_key -- nvmf/common.sh@51 -- # : 0 00:09:20.572 17:25:02 json_config_extra_key -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:09:20.572 17:25:02 json_config_extra_key -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:09:20.572 17:25:02 json_config_extra_key -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:09:20.572 17:25:02 json_config_extra_key -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:09:20.572 17:25:02 json_config_extra_key -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:09:20.572 17:25:02 json_config_extra_key -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:09:20.572 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:09:20.572 17:25:02 json_config_extra_key -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:09:20.572 17:25:02 json_config_extra_key -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:09:20.572 17:25:02 json_config_extra_key -- nvmf/common.sh@55 -- # have_pci_nics=0 00:09:20.572 17:25:02 json_config_extra_key -- json_config/json_config_extra_key.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/common.sh 00:09:20.572 17:25:02 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # app_pid=(['target']='') 00:09:20.572 17:25:02 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # declare -A app_pid 00:09:20.572 17:25:02 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # app_socket=(['target']='/var/tmp/spdk_tgt.sock') 00:09:20.572 17:25:02 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # declare -A app_socket 00:09:20.572 17:25:02 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # app_params=(['target']='-m 0x1 -s 1024') 00:09:20.572 17:25:02 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # declare -A app_params 00:09:20.572 17:25:02 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # configs_path=(['target']='/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/extra_key.json') 00:09:20.572 17:25:02 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # declare -A configs_path 00:09:20.572 17:25:02 json_config_extra_key -- json_config/json_config_extra_key.sh@22 -- # trap 'on_error_exit "${FUNCNAME}" "${LINENO}"' ERR 00:09:20.572 17:25:02 json_config_extra_key -- json_config/json_config_extra_key.sh@24 -- # echo 'INFO: launching applications...' 00:09:20.572 INFO: launching applications... 00:09:20.572 17:25:02 json_config_extra_key -- json_config/json_config_extra_key.sh@25 -- # json_config_test_start_app target --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/extra_key.json 00:09:20.572 17:25:02 json_config_extra_key -- json_config/common.sh@9 -- # local app=target 00:09:20.572 17:25:02 json_config_extra_key -- json_config/common.sh@10 -- # shift 00:09:20.572 17:25:02 json_config_extra_key -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:09:20.572 17:25:02 json_config_extra_key -- json_config/common.sh@13 -- # [[ -z '' ]] 00:09:20.572 17:25:02 json_config_extra_key -- json_config/common.sh@15 -- # local app_extra_params= 00:09:20.572 17:25:02 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:09:20.572 17:25:02 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:09:20.572 17:25:02 json_config_extra_key -- json_config/common.sh@22 -- # app_pid["$app"]=115341 00:09:20.572 17:25:02 json_config_extra_key -- json_config/common.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/extra_key.json 00:09:20.572 17:25:02 json_config_extra_key -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:09:20.572 Waiting for target to run... 00:09:20.572 17:25:02 json_config_extra_key -- json_config/common.sh@25 -- # waitforlisten 115341 /var/tmp/spdk_tgt.sock 00:09:20.572 17:25:02 json_config_extra_key -- common/autotest_common.sh@835 -- # '[' -z 115341 ']' 00:09:20.572 17:25:02 json_config_extra_key -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:09:20.572 17:25:02 json_config_extra_key -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:20.572 17:25:02 json_config_extra_key -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:09:20.572 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:09:20.572 17:25:02 json_config_extra_key -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:20.572 17:25:02 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:09:20.572 [2024-12-06 17:25:02.293987] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:09:20.572 [2024-12-06 17:25:02.294081] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid115341 ] 00:09:20.832 [2024-12-06 17:25:02.630788] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:20.832 [2024-12-06 17:25:02.662087] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:21.767 17:25:03 json_config_extra_key -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:21.767 17:25:03 json_config_extra_key -- common/autotest_common.sh@868 -- # return 0 00:09:21.768 17:25:03 json_config_extra_key -- json_config/common.sh@26 -- # echo '' 00:09:21.768 00:09:21.768 17:25:03 json_config_extra_key -- json_config/json_config_extra_key.sh@27 -- # echo 'INFO: shutting down applications...' 00:09:21.768 INFO: shutting down applications... 00:09:21.768 17:25:03 json_config_extra_key -- json_config/json_config_extra_key.sh@28 -- # json_config_test_shutdown_app target 00:09:21.768 17:25:03 json_config_extra_key -- json_config/common.sh@31 -- # local app=target 00:09:21.768 17:25:03 json_config_extra_key -- json_config/common.sh@34 -- # [[ -n 22 ]] 00:09:21.768 17:25:03 json_config_extra_key -- json_config/common.sh@35 -- # [[ -n 115341 ]] 00:09:21.768 17:25:03 json_config_extra_key -- json_config/common.sh@38 -- # kill -SIGINT 115341 00:09:21.768 17:25:03 json_config_extra_key -- json_config/common.sh@40 -- # (( i = 0 )) 00:09:21.768 17:25:03 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:09:21.768 17:25:03 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 115341 00:09:21.768 17:25:03 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:09:22.027 17:25:03 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:09:22.027 17:25:03 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:09:22.027 17:25:03 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 115341 00:09:22.027 17:25:03 json_config_extra_key -- json_config/common.sh@42 -- # app_pid["$app"]= 00:09:22.027 17:25:03 json_config_extra_key -- json_config/common.sh@43 -- # break 00:09:22.027 17:25:03 json_config_extra_key -- json_config/common.sh@48 -- # [[ -n '' ]] 00:09:22.027 17:25:03 json_config_extra_key -- json_config/common.sh@53 -- # echo 'SPDK target shutdown done' 00:09:22.027 SPDK target shutdown done 00:09:22.027 17:25:03 json_config_extra_key -- json_config/json_config_extra_key.sh@30 -- # echo Success 00:09:22.027 Success 00:09:22.027 00:09:22.027 real 0m1.684s 00:09:22.027 user 0m1.656s 00:09:22.027 sys 0m0.449s 00:09:22.027 17:25:03 json_config_extra_key -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:22.027 17:25:03 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:09:22.027 ************************************ 00:09:22.027 END TEST json_config_extra_key 00:09:22.027 ************************************ 00:09:22.027 17:25:03 -- spdk/autotest.sh@161 -- # run_test alias_rpc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:09:22.027 17:25:03 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:09:22.027 17:25:03 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:22.027 17:25:03 -- common/autotest_common.sh@10 -- # set +x 00:09:22.027 ************************************ 00:09:22.027 START TEST alias_rpc 00:09:22.027 ************************************ 00:09:22.027 17:25:03 alias_rpc -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:09:22.287 * Looking for test storage... 00:09:22.287 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/alias_rpc 00:09:22.287 17:25:03 alias_rpc -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:09:22.287 17:25:03 alias_rpc -- common/autotest_common.sh@1711 -- # lcov --version 00:09:22.287 17:25:03 alias_rpc -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:09:22.287 17:25:03 alias_rpc -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:09:22.287 17:25:03 alias_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:09:22.287 17:25:03 alias_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:09:22.287 17:25:03 alias_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:09:22.287 17:25:03 alias_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:09:22.287 17:25:03 alias_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:09:22.287 17:25:03 alias_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:09:22.287 17:25:03 alias_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:09:22.287 17:25:03 alias_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:09:22.287 17:25:03 alias_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:09:22.287 17:25:03 alias_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:09:22.287 17:25:03 alias_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:09:22.287 17:25:03 alias_rpc -- scripts/common.sh@344 -- # case "$op" in 00:09:22.287 17:25:03 alias_rpc -- scripts/common.sh@345 -- # : 1 00:09:22.287 17:25:03 alias_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:09:22.287 17:25:03 alias_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:09:22.287 17:25:03 alias_rpc -- scripts/common.sh@365 -- # decimal 1 00:09:22.287 17:25:03 alias_rpc -- scripts/common.sh@353 -- # local d=1 00:09:22.287 17:25:03 alias_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:09:22.287 17:25:03 alias_rpc -- scripts/common.sh@355 -- # echo 1 00:09:22.287 17:25:03 alias_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:09:22.287 17:25:03 alias_rpc -- scripts/common.sh@366 -- # decimal 2 00:09:22.287 17:25:03 alias_rpc -- scripts/common.sh@353 -- # local d=2 00:09:22.287 17:25:03 alias_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:09:22.287 17:25:03 alias_rpc -- scripts/common.sh@355 -- # echo 2 00:09:22.287 17:25:03 alias_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:09:22.287 17:25:03 alias_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:09:22.287 17:25:03 alias_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:09:22.287 17:25:03 alias_rpc -- scripts/common.sh@368 -- # return 0 00:09:22.287 17:25:03 alias_rpc -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:09:22.287 17:25:03 alias_rpc -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:09:22.287 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:22.287 --rc genhtml_branch_coverage=1 00:09:22.287 --rc genhtml_function_coverage=1 00:09:22.287 --rc genhtml_legend=1 00:09:22.287 --rc geninfo_all_blocks=1 00:09:22.287 --rc geninfo_unexecuted_blocks=1 00:09:22.287 00:09:22.287 ' 00:09:22.287 17:25:03 alias_rpc -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:09:22.287 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:22.287 --rc genhtml_branch_coverage=1 00:09:22.287 --rc genhtml_function_coverage=1 00:09:22.287 --rc genhtml_legend=1 00:09:22.287 --rc geninfo_all_blocks=1 00:09:22.287 --rc geninfo_unexecuted_blocks=1 00:09:22.287 00:09:22.287 ' 00:09:22.287 17:25:03 alias_rpc -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:09:22.287 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:22.287 --rc genhtml_branch_coverage=1 00:09:22.287 --rc genhtml_function_coverage=1 00:09:22.287 --rc genhtml_legend=1 00:09:22.287 --rc geninfo_all_blocks=1 00:09:22.287 --rc geninfo_unexecuted_blocks=1 00:09:22.287 00:09:22.287 ' 00:09:22.287 17:25:03 alias_rpc -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:09:22.287 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:22.287 --rc genhtml_branch_coverage=1 00:09:22.287 --rc genhtml_function_coverage=1 00:09:22.287 --rc genhtml_legend=1 00:09:22.287 --rc geninfo_all_blocks=1 00:09:22.287 --rc geninfo_unexecuted_blocks=1 00:09:22.287 00:09:22.287 ' 00:09:22.287 17:25:03 alias_rpc -- alias_rpc/alias_rpc.sh@10 -- # trap 'killprocess $spdk_tgt_pid; exit 1' ERR 00:09:22.287 17:25:03 alias_rpc -- alias_rpc/alias_rpc.sh@13 -- # spdk_tgt_pid=115652 00:09:22.287 17:25:03 alias_rpc -- alias_rpc/alias_rpc.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:09:22.287 17:25:03 alias_rpc -- alias_rpc/alias_rpc.sh@14 -- # waitforlisten 115652 00:09:22.287 17:25:03 alias_rpc -- common/autotest_common.sh@835 -- # '[' -z 115652 ']' 00:09:22.287 17:25:03 alias_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:22.287 17:25:03 alias_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:22.287 17:25:03 alias_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:22.287 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:22.287 17:25:03 alias_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:22.287 17:25:03 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:09:22.287 [2024-12-06 17:25:04.023357] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:09:22.287 [2024-12-06 17:25:04.023464] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid115652 ] 00:09:22.287 [2024-12-06 17:25:04.098714] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:22.547 [2024-12-06 17:25:04.148185] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:22.806 17:25:04 alias_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:22.806 17:25:04 alias_rpc -- common/autotest_common.sh@868 -- # return 0 00:09:22.806 17:25:04 alias_rpc -- alias_rpc/alias_rpc.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py load_config -i 00:09:23.065 17:25:04 alias_rpc -- alias_rpc/alias_rpc.sh@19 -- # killprocess 115652 00:09:23.065 17:25:04 alias_rpc -- common/autotest_common.sh@954 -- # '[' -z 115652 ']' 00:09:23.065 17:25:04 alias_rpc -- common/autotest_common.sh@958 -- # kill -0 115652 00:09:23.065 17:25:04 alias_rpc -- common/autotest_common.sh@959 -- # uname 00:09:23.065 17:25:04 alias_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:23.065 17:25:04 alias_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 115652 00:09:23.065 17:25:04 alias_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:23.065 17:25:04 alias_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:23.065 17:25:04 alias_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 115652' 00:09:23.065 killing process with pid 115652 00:09:23.065 17:25:04 alias_rpc -- common/autotest_common.sh@973 -- # kill 115652 00:09:23.065 17:25:04 alias_rpc -- common/autotest_common.sh@978 -- # wait 115652 00:09:23.326 00:09:23.326 real 0m1.283s 00:09:23.326 user 0m1.403s 00:09:23.326 sys 0m0.425s 00:09:23.326 17:25:05 alias_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:23.326 17:25:05 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:09:23.326 ************************************ 00:09:23.326 END TEST alias_rpc 00:09:23.326 ************************************ 00:09:23.326 17:25:05 -- spdk/autotest.sh@163 -- # [[ 0 -eq 0 ]] 00:09:23.326 17:25:05 -- spdk/autotest.sh@164 -- # run_test spdkcli_tcp /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/tcp.sh 00:09:23.326 17:25:05 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:09:23.326 17:25:05 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:23.326 17:25:05 -- common/autotest_common.sh@10 -- # set +x 00:09:23.326 ************************************ 00:09:23.326 START TEST spdkcli_tcp 00:09:23.326 ************************************ 00:09:23.326 17:25:05 spdkcli_tcp -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/tcp.sh 00:09:23.585 * Looking for test storage... 00:09:23.585 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli 00:09:23.585 17:25:05 spdkcli_tcp -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:09:23.585 17:25:05 spdkcli_tcp -- common/autotest_common.sh@1711 -- # lcov --version 00:09:23.585 17:25:05 spdkcli_tcp -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:09:23.585 17:25:05 spdkcli_tcp -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:09:23.585 17:25:05 spdkcli_tcp -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:09:23.585 17:25:05 spdkcli_tcp -- scripts/common.sh@333 -- # local ver1 ver1_l 00:09:23.585 17:25:05 spdkcli_tcp -- scripts/common.sh@334 -- # local ver2 ver2_l 00:09:23.585 17:25:05 spdkcli_tcp -- scripts/common.sh@336 -- # IFS=.-: 00:09:23.585 17:25:05 spdkcli_tcp -- scripts/common.sh@336 -- # read -ra ver1 00:09:23.585 17:25:05 spdkcli_tcp -- scripts/common.sh@337 -- # IFS=.-: 00:09:23.585 17:25:05 spdkcli_tcp -- scripts/common.sh@337 -- # read -ra ver2 00:09:23.585 17:25:05 spdkcli_tcp -- scripts/common.sh@338 -- # local 'op=<' 00:09:23.585 17:25:05 spdkcli_tcp -- scripts/common.sh@340 -- # ver1_l=2 00:09:23.585 17:25:05 spdkcli_tcp -- scripts/common.sh@341 -- # ver2_l=1 00:09:23.585 17:25:05 spdkcli_tcp -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:09:23.585 17:25:05 spdkcli_tcp -- scripts/common.sh@344 -- # case "$op" in 00:09:23.585 17:25:05 spdkcli_tcp -- scripts/common.sh@345 -- # : 1 00:09:23.585 17:25:05 spdkcli_tcp -- scripts/common.sh@364 -- # (( v = 0 )) 00:09:23.585 17:25:05 spdkcli_tcp -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:09:23.585 17:25:05 spdkcli_tcp -- scripts/common.sh@365 -- # decimal 1 00:09:23.585 17:25:05 spdkcli_tcp -- scripts/common.sh@353 -- # local d=1 00:09:23.585 17:25:05 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:09:23.585 17:25:05 spdkcli_tcp -- scripts/common.sh@355 -- # echo 1 00:09:23.585 17:25:05 spdkcli_tcp -- scripts/common.sh@365 -- # ver1[v]=1 00:09:23.585 17:25:05 spdkcli_tcp -- scripts/common.sh@366 -- # decimal 2 00:09:23.585 17:25:05 spdkcli_tcp -- scripts/common.sh@353 -- # local d=2 00:09:23.585 17:25:05 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:09:23.585 17:25:05 spdkcli_tcp -- scripts/common.sh@355 -- # echo 2 00:09:23.585 17:25:05 spdkcli_tcp -- scripts/common.sh@366 -- # ver2[v]=2 00:09:23.585 17:25:05 spdkcli_tcp -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:09:23.585 17:25:05 spdkcli_tcp -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:09:23.585 17:25:05 spdkcli_tcp -- scripts/common.sh@368 -- # return 0 00:09:23.585 17:25:05 spdkcli_tcp -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:09:23.585 17:25:05 spdkcli_tcp -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:09:23.585 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:23.585 --rc genhtml_branch_coverage=1 00:09:23.585 --rc genhtml_function_coverage=1 00:09:23.585 --rc genhtml_legend=1 00:09:23.585 --rc geninfo_all_blocks=1 00:09:23.585 --rc geninfo_unexecuted_blocks=1 00:09:23.585 00:09:23.585 ' 00:09:23.585 17:25:05 spdkcli_tcp -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:09:23.585 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:23.585 --rc genhtml_branch_coverage=1 00:09:23.585 --rc genhtml_function_coverage=1 00:09:23.585 --rc genhtml_legend=1 00:09:23.585 --rc geninfo_all_blocks=1 00:09:23.585 --rc geninfo_unexecuted_blocks=1 00:09:23.585 00:09:23.585 ' 00:09:23.585 17:25:05 spdkcli_tcp -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:09:23.585 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:23.585 --rc genhtml_branch_coverage=1 00:09:23.585 --rc genhtml_function_coverage=1 00:09:23.585 --rc genhtml_legend=1 00:09:23.585 --rc geninfo_all_blocks=1 00:09:23.585 --rc geninfo_unexecuted_blocks=1 00:09:23.585 00:09:23.585 ' 00:09:23.585 17:25:05 spdkcli_tcp -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:09:23.585 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:23.585 --rc genhtml_branch_coverage=1 00:09:23.585 --rc genhtml_function_coverage=1 00:09:23.585 --rc genhtml_legend=1 00:09:23.585 --rc geninfo_all_blocks=1 00:09:23.585 --rc geninfo_unexecuted_blocks=1 00:09:23.585 00:09:23.585 ' 00:09:23.585 17:25:05 spdkcli_tcp -- spdkcli/tcp.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/common.sh 00:09:23.585 17:25:05 spdkcli_tcp -- spdkcli/common.sh@6 -- # spdkcli_job=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_job.py 00:09:23.585 17:25:05 spdkcli_tcp -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/clear_config.py 00:09:23.585 17:25:05 spdkcli_tcp -- spdkcli/tcp.sh@18 -- # IP_ADDRESS=127.0.0.1 00:09:23.585 17:25:05 spdkcli_tcp -- spdkcli/tcp.sh@19 -- # PORT=9998 00:09:23.585 17:25:05 spdkcli_tcp -- spdkcli/tcp.sh@21 -- # trap 'err_cleanup; exit 1' SIGINT SIGTERM EXIT 00:09:23.585 17:25:05 spdkcli_tcp -- spdkcli/tcp.sh@23 -- # timing_enter run_spdk_tgt_tcp 00:09:23.585 17:25:05 spdkcli_tcp -- common/autotest_common.sh@726 -- # xtrace_disable 00:09:23.585 17:25:05 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:09:23.585 17:25:05 spdkcli_tcp -- spdkcli/tcp.sh@25 -- # spdk_tgt_pid=115853 00:09:23.585 17:25:05 spdkcli_tcp -- spdkcli/tcp.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:09:23.585 17:25:05 spdkcli_tcp -- spdkcli/tcp.sh@27 -- # waitforlisten 115853 00:09:23.585 17:25:05 spdkcli_tcp -- common/autotest_common.sh@835 -- # '[' -z 115853 ']' 00:09:23.585 17:25:05 spdkcli_tcp -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:23.585 17:25:05 spdkcli_tcp -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:23.585 17:25:05 spdkcli_tcp -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:23.585 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:23.585 17:25:05 spdkcli_tcp -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:23.585 17:25:05 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:09:23.585 [2024-12-06 17:25:05.359862] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:09:23.585 [2024-12-06 17:25:05.359979] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid115853 ] 00:09:23.844 [2024-12-06 17:25:05.429543] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:09:23.844 [2024-12-06 17:25:05.477213] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:09:23.844 [2024-12-06 17:25:05.477217] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:24.101 17:25:05 spdkcli_tcp -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:24.101 17:25:05 spdkcli_tcp -- common/autotest_common.sh@868 -- # return 0 00:09:24.101 17:25:05 spdkcli_tcp -- spdkcli/tcp.sh@31 -- # socat_pid=115862 00:09:24.101 17:25:05 spdkcli_tcp -- spdkcli/tcp.sh@30 -- # socat TCP-LISTEN:9998 UNIX-CONNECT:/var/tmp/spdk.sock 00:09:24.101 17:25:05 spdkcli_tcp -- spdkcli/tcp.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -r 100 -t 2 -s 127.0.0.1 -p 9998 rpc_get_methods 00:09:24.359 [ 00:09:24.359 "bdev_malloc_delete", 00:09:24.359 "bdev_malloc_create", 00:09:24.359 "bdev_null_resize", 00:09:24.359 "bdev_null_delete", 00:09:24.359 "bdev_null_create", 00:09:24.359 "bdev_nvme_cuse_unregister", 00:09:24.359 "bdev_nvme_cuse_register", 00:09:24.359 "bdev_opal_new_user", 00:09:24.359 "bdev_opal_set_lock_state", 00:09:24.359 "bdev_opal_delete", 00:09:24.359 "bdev_opal_get_info", 00:09:24.359 "bdev_opal_create", 00:09:24.359 "bdev_nvme_opal_revert", 00:09:24.359 "bdev_nvme_opal_init", 00:09:24.359 "bdev_nvme_send_cmd", 00:09:24.359 "bdev_nvme_set_keys", 00:09:24.359 "bdev_nvme_get_path_iostat", 00:09:24.359 "bdev_nvme_get_mdns_discovery_info", 00:09:24.359 "bdev_nvme_stop_mdns_discovery", 00:09:24.359 "bdev_nvme_start_mdns_discovery", 00:09:24.359 "bdev_nvme_set_multipath_policy", 00:09:24.359 "bdev_nvme_set_preferred_path", 00:09:24.359 "bdev_nvme_get_io_paths", 00:09:24.359 "bdev_nvme_remove_error_injection", 00:09:24.359 "bdev_nvme_add_error_injection", 00:09:24.360 "bdev_nvme_get_discovery_info", 00:09:24.360 "bdev_nvme_stop_discovery", 00:09:24.360 "bdev_nvme_start_discovery", 00:09:24.360 "bdev_nvme_get_controller_health_info", 00:09:24.360 "bdev_nvme_disable_controller", 00:09:24.360 "bdev_nvme_enable_controller", 00:09:24.360 "bdev_nvme_reset_controller", 00:09:24.360 "bdev_nvme_get_transport_statistics", 00:09:24.360 "bdev_nvme_apply_firmware", 00:09:24.360 "bdev_nvme_detach_controller", 00:09:24.360 "bdev_nvme_get_controllers", 00:09:24.360 "bdev_nvme_attach_controller", 00:09:24.360 "bdev_nvme_set_hotplug", 00:09:24.360 "bdev_nvme_set_options", 00:09:24.360 "bdev_passthru_delete", 00:09:24.360 "bdev_passthru_create", 00:09:24.360 "bdev_lvol_set_parent_bdev", 00:09:24.360 "bdev_lvol_set_parent", 00:09:24.360 "bdev_lvol_check_shallow_copy", 00:09:24.360 "bdev_lvol_start_shallow_copy", 00:09:24.360 "bdev_lvol_grow_lvstore", 00:09:24.360 "bdev_lvol_get_lvols", 00:09:24.360 "bdev_lvol_get_lvstores", 00:09:24.360 "bdev_lvol_delete", 00:09:24.360 "bdev_lvol_set_read_only", 00:09:24.360 "bdev_lvol_resize", 00:09:24.360 "bdev_lvol_decouple_parent", 00:09:24.360 "bdev_lvol_inflate", 00:09:24.360 "bdev_lvol_rename", 00:09:24.360 "bdev_lvol_clone_bdev", 00:09:24.360 "bdev_lvol_clone", 00:09:24.360 "bdev_lvol_snapshot", 00:09:24.360 "bdev_lvol_create", 00:09:24.360 "bdev_lvol_delete_lvstore", 00:09:24.360 "bdev_lvol_rename_lvstore", 00:09:24.360 "bdev_lvol_create_lvstore", 00:09:24.360 "bdev_raid_set_options", 00:09:24.360 "bdev_raid_remove_base_bdev", 00:09:24.360 "bdev_raid_add_base_bdev", 00:09:24.360 "bdev_raid_delete", 00:09:24.360 "bdev_raid_create", 00:09:24.360 "bdev_raid_get_bdevs", 00:09:24.360 "bdev_error_inject_error", 00:09:24.360 "bdev_error_delete", 00:09:24.360 "bdev_error_create", 00:09:24.360 "bdev_split_delete", 00:09:24.360 "bdev_split_create", 00:09:24.360 "bdev_delay_delete", 00:09:24.360 "bdev_delay_create", 00:09:24.360 "bdev_delay_update_latency", 00:09:24.360 "bdev_zone_block_delete", 00:09:24.360 "bdev_zone_block_create", 00:09:24.360 "blobfs_create", 00:09:24.360 "blobfs_detect", 00:09:24.360 "blobfs_set_cache_size", 00:09:24.360 "bdev_aio_delete", 00:09:24.360 "bdev_aio_rescan", 00:09:24.360 "bdev_aio_create", 00:09:24.360 "bdev_ftl_set_property", 00:09:24.360 "bdev_ftl_get_properties", 00:09:24.360 "bdev_ftl_get_stats", 00:09:24.360 "bdev_ftl_unmap", 00:09:24.360 "bdev_ftl_unload", 00:09:24.360 "bdev_ftl_delete", 00:09:24.360 "bdev_ftl_load", 00:09:24.360 "bdev_ftl_create", 00:09:24.360 "bdev_virtio_attach_controller", 00:09:24.360 "bdev_virtio_scsi_get_devices", 00:09:24.360 "bdev_virtio_detach_controller", 00:09:24.360 "bdev_virtio_blk_set_hotplug", 00:09:24.360 "bdev_iscsi_delete", 00:09:24.360 "bdev_iscsi_create", 00:09:24.360 "bdev_iscsi_set_options", 00:09:24.360 "accel_error_inject_error", 00:09:24.360 "ioat_scan_accel_module", 00:09:24.360 "dsa_scan_accel_module", 00:09:24.360 "iaa_scan_accel_module", 00:09:24.360 "vfu_virtio_create_fs_endpoint", 00:09:24.360 "vfu_virtio_create_scsi_endpoint", 00:09:24.360 "vfu_virtio_scsi_remove_target", 00:09:24.360 "vfu_virtio_scsi_add_target", 00:09:24.360 "vfu_virtio_create_blk_endpoint", 00:09:24.360 "vfu_virtio_delete_endpoint", 00:09:24.360 "keyring_file_remove_key", 00:09:24.360 "keyring_file_add_key", 00:09:24.360 "keyring_linux_set_options", 00:09:24.360 "fsdev_aio_delete", 00:09:24.360 "fsdev_aio_create", 00:09:24.360 "iscsi_get_histogram", 00:09:24.360 "iscsi_enable_histogram", 00:09:24.360 "iscsi_set_options", 00:09:24.360 "iscsi_get_auth_groups", 00:09:24.360 "iscsi_auth_group_remove_secret", 00:09:24.360 "iscsi_auth_group_add_secret", 00:09:24.360 "iscsi_delete_auth_group", 00:09:24.360 "iscsi_create_auth_group", 00:09:24.360 "iscsi_set_discovery_auth", 00:09:24.360 "iscsi_get_options", 00:09:24.360 "iscsi_target_node_request_logout", 00:09:24.360 "iscsi_target_node_set_redirect", 00:09:24.360 "iscsi_target_node_set_auth", 00:09:24.360 "iscsi_target_node_add_lun", 00:09:24.360 "iscsi_get_stats", 00:09:24.360 "iscsi_get_connections", 00:09:24.360 "iscsi_portal_group_set_auth", 00:09:24.360 "iscsi_start_portal_group", 00:09:24.360 "iscsi_delete_portal_group", 00:09:24.360 "iscsi_create_portal_group", 00:09:24.360 "iscsi_get_portal_groups", 00:09:24.360 "iscsi_delete_target_node", 00:09:24.360 "iscsi_target_node_remove_pg_ig_maps", 00:09:24.360 "iscsi_target_node_add_pg_ig_maps", 00:09:24.360 "iscsi_create_target_node", 00:09:24.360 "iscsi_get_target_nodes", 00:09:24.360 "iscsi_delete_initiator_group", 00:09:24.360 "iscsi_initiator_group_remove_initiators", 00:09:24.360 "iscsi_initiator_group_add_initiators", 00:09:24.360 "iscsi_create_initiator_group", 00:09:24.360 "iscsi_get_initiator_groups", 00:09:24.360 "nvmf_set_crdt", 00:09:24.360 "nvmf_set_config", 00:09:24.360 "nvmf_set_max_subsystems", 00:09:24.360 "nvmf_stop_mdns_prr", 00:09:24.360 "nvmf_publish_mdns_prr", 00:09:24.360 "nvmf_subsystem_get_listeners", 00:09:24.360 "nvmf_subsystem_get_qpairs", 00:09:24.360 "nvmf_subsystem_get_controllers", 00:09:24.360 "nvmf_get_stats", 00:09:24.360 "nvmf_get_transports", 00:09:24.360 "nvmf_create_transport", 00:09:24.360 "nvmf_get_targets", 00:09:24.360 "nvmf_delete_target", 00:09:24.360 "nvmf_create_target", 00:09:24.360 "nvmf_subsystem_allow_any_host", 00:09:24.360 "nvmf_subsystem_set_keys", 00:09:24.360 "nvmf_subsystem_remove_host", 00:09:24.360 "nvmf_subsystem_add_host", 00:09:24.360 "nvmf_ns_remove_host", 00:09:24.360 "nvmf_ns_add_host", 00:09:24.360 "nvmf_subsystem_remove_ns", 00:09:24.360 "nvmf_subsystem_set_ns_ana_group", 00:09:24.360 "nvmf_subsystem_add_ns", 00:09:24.360 "nvmf_subsystem_listener_set_ana_state", 00:09:24.360 "nvmf_discovery_get_referrals", 00:09:24.360 "nvmf_discovery_remove_referral", 00:09:24.360 "nvmf_discovery_add_referral", 00:09:24.360 "nvmf_subsystem_remove_listener", 00:09:24.360 "nvmf_subsystem_add_listener", 00:09:24.360 "nvmf_delete_subsystem", 00:09:24.360 "nvmf_create_subsystem", 00:09:24.360 "nvmf_get_subsystems", 00:09:24.360 "env_dpdk_get_mem_stats", 00:09:24.360 "nbd_get_disks", 00:09:24.360 "nbd_stop_disk", 00:09:24.360 "nbd_start_disk", 00:09:24.360 "ublk_recover_disk", 00:09:24.360 "ublk_get_disks", 00:09:24.360 "ublk_stop_disk", 00:09:24.360 "ublk_start_disk", 00:09:24.360 "ublk_destroy_target", 00:09:24.360 "ublk_create_target", 00:09:24.360 "virtio_blk_create_transport", 00:09:24.360 "virtio_blk_get_transports", 00:09:24.360 "vhost_controller_set_coalescing", 00:09:24.360 "vhost_get_controllers", 00:09:24.360 "vhost_delete_controller", 00:09:24.360 "vhost_create_blk_controller", 00:09:24.360 "vhost_scsi_controller_remove_target", 00:09:24.360 "vhost_scsi_controller_add_target", 00:09:24.360 "vhost_start_scsi_controller", 00:09:24.360 "vhost_create_scsi_controller", 00:09:24.360 "thread_set_cpumask", 00:09:24.360 "scheduler_set_options", 00:09:24.360 "framework_get_governor", 00:09:24.360 "framework_get_scheduler", 00:09:24.360 "framework_set_scheduler", 00:09:24.360 "framework_get_reactors", 00:09:24.360 "thread_get_io_channels", 00:09:24.360 "thread_get_pollers", 00:09:24.360 "thread_get_stats", 00:09:24.360 "framework_monitor_context_switch", 00:09:24.360 "spdk_kill_instance", 00:09:24.360 "log_enable_timestamps", 00:09:24.360 "log_get_flags", 00:09:24.360 "log_clear_flag", 00:09:24.360 "log_set_flag", 00:09:24.360 "log_get_level", 00:09:24.360 "log_set_level", 00:09:24.360 "log_get_print_level", 00:09:24.360 "log_set_print_level", 00:09:24.360 "framework_enable_cpumask_locks", 00:09:24.360 "framework_disable_cpumask_locks", 00:09:24.360 "framework_wait_init", 00:09:24.360 "framework_start_init", 00:09:24.360 "scsi_get_devices", 00:09:24.360 "bdev_get_histogram", 00:09:24.360 "bdev_enable_histogram", 00:09:24.360 "bdev_set_qos_limit", 00:09:24.360 "bdev_set_qd_sampling_period", 00:09:24.360 "bdev_get_bdevs", 00:09:24.360 "bdev_reset_iostat", 00:09:24.360 "bdev_get_iostat", 00:09:24.360 "bdev_examine", 00:09:24.360 "bdev_wait_for_examine", 00:09:24.360 "bdev_set_options", 00:09:24.360 "accel_get_stats", 00:09:24.360 "accel_set_options", 00:09:24.360 "accel_set_driver", 00:09:24.360 "accel_crypto_key_destroy", 00:09:24.360 "accel_crypto_keys_get", 00:09:24.360 "accel_crypto_key_create", 00:09:24.360 "accel_assign_opc", 00:09:24.360 "accel_get_module_info", 00:09:24.360 "accel_get_opc_assignments", 00:09:24.360 "vmd_rescan", 00:09:24.360 "vmd_remove_device", 00:09:24.360 "vmd_enable", 00:09:24.360 "sock_get_default_impl", 00:09:24.360 "sock_set_default_impl", 00:09:24.360 "sock_impl_set_options", 00:09:24.360 "sock_impl_get_options", 00:09:24.360 "iobuf_get_stats", 00:09:24.360 "iobuf_set_options", 00:09:24.360 "keyring_get_keys", 00:09:24.360 "vfu_tgt_set_base_path", 00:09:24.360 "framework_get_pci_devices", 00:09:24.360 "framework_get_config", 00:09:24.360 "framework_get_subsystems", 00:09:24.360 "fsdev_set_opts", 00:09:24.360 "fsdev_get_opts", 00:09:24.360 "trace_get_info", 00:09:24.360 "trace_get_tpoint_group_mask", 00:09:24.360 "trace_disable_tpoint_group", 00:09:24.360 "trace_enable_tpoint_group", 00:09:24.360 "trace_clear_tpoint_mask", 00:09:24.360 "trace_set_tpoint_mask", 00:09:24.360 "notify_get_notifications", 00:09:24.360 "notify_get_types", 00:09:24.360 "spdk_get_version", 00:09:24.360 "rpc_get_methods" 00:09:24.360 ] 00:09:24.360 17:25:06 spdkcli_tcp -- spdkcli/tcp.sh@35 -- # timing_exit run_spdk_tgt_tcp 00:09:24.360 17:25:06 spdkcli_tcp -- common/autotest_common.sh@732 -- # xtrace_disable 00:09:24.360 17:25:06 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:09:24.360 17:25:06 spdkcli_tcp -- spdkcli/tcp.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:09:24.360 17:25:06 spdkcli_tcp -- spdkcli/tcp.sh@38 -- # killprocess 115853 00:09:24.360 17:25:06 spdkcli_tcp -- common/autotest_common.sh@954 -- # '[' -z 115853 ']' 00:09:24.360 17:25:06 spdkcli_tcp -- common/autotest_common.sh@958 -- # kill -0 115853 00:09:24.360 17:25:06 spdkcli_tcp -- common/autotest_common.sh@959 -- # uname 00:09:24.360 17:25:06 spdkcli_tcp -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:24.360 17:25:06 spdkcli_tcp -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 115853 00:09:24.360 17:25:06 spdkcli_tcp -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:24.361 17:25:06 spdkcli_tcp -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:24.361 17:25:06 spdkcli_tcp -- common/autotest_common.sh@972 -- # echo 'killing process with pid 115853' 00:09:24.361 killing process with pid 115853 00:09:24.361 17:25:06 spdkcli_tcp -- common/autotest_common.sh@973 -- # kill 115853 00:09:24.361 17:25:06 spdkcli_tcp -- common/autotest_common.sh@978 -- # wait 115853 00:09:24.927 00:09:24.927 real 0m1.321s 00:09:24.927 user 0m2.389s 00:09:24.927 sys 0m0.479s 00:09:24.927 17:25:06 spdkcli_tcp -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:24.927 17:25:06 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:09:24.927 ************************************ 00:09:24.927 END TEST spdkcli_tcp 00:09:24.927 ************************************ 00:09:24.927 17:25:06 -- spdk/autotest.sh@167 -- # run_test dpdk_mem_utility /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:09:24.927 17:25:06 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:09:24.927 17:25:06 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:24.927 17:25:06 -- common/autotest_common.sh@10 -- # set +x 00:09:24.927 ************************************ 00:09:24.927 START TEST dpdk_mem_utility 00:09:24.927 ************************************ 00:09:24.927 17:25:06 dpdk_mem_utility -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:09:24.927 * Looking for test storage... 00:09:24.927 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/dpdk_memory_utility 00:09:24.927 17:25:06 dpdk_mem_utility -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:09:24.927 17:25:06 dpdk_mem_utility -- common/autotest_common.sh@1711 -- # lcov --version 00:09:24.927 17:25:06 dpdk_mem_utility -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:09:24.927 17:25:06 dpdk_mem_utility -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:09:24.927 17:25:06 dpdk_mem_utility -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:09:24.927 17:25:06 dpdk_mem_utility -- scripts/common.sh@333 -- # local ver1 ver1_l 00:09:24.927 17:25:06 dpdk_mem_utility -- scripts/common.sh@334 -- # local ver2 ver2_l 00:09:24.927 17:25:06 dpdk_mem_utility -- scripts/common.sh@336 -- # IFS=.-: 00:09:24.927 17:25:06 dpdk_mem_utility -- scripts/common.sh@336 -- # read -ra ver1 00:09:24.927 17:25:06 dpdk_mem_utility -- scripts/common.sh@337 -- # IFS=.-: 00:09:24.927 17:25:06 dpdk_mem_utility -- scripts/common.sh@337 -- # read -ra ver2 00:09:24.927 17:25:06 dpdk_mem_utility -- scripts/common.sh@338 -- # local 'op=<' 00:09:24.927 17:25:06 dpdk_mem_utility -- scripts/common.sh@340 -- # ver1_l=2 00:09:24.927 17:25:06 dpdk_mem_utility -- scripts/common.sh@341 -- # ver2_l=1 00:09:24.927 17:25:06 dpdk_mem_utility -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:09:24.927 17:25:06 dpdk_mem_utility -- scripts/common.sh@344 -- # case "$op" in 00:09:24.927 17:25:06 dpdk_mem_utility -- scripts/common.sh@345 -- # : 1 00:09:24.927 17:25:06 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v = 0 )) 00:09:24.927 17:25:06 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:09:24.927 17:25:06 dpdk_mem_utility -- scripts/common.sh@365 -- # decimal 1 00:09:24.927 17:25:06 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=1 00:09:24.927 17:25:06 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:09:24.927 17:25:06 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 1 00:09:24.927 17:25:06 dpdk_mem_utility -- scripts/common.sh@365 -- # ver1[v]=1 00:09:24.927 17:25:06 dpdk_mem_utility -- scripts/common.sh@366 -- # decimal 2 00:09:24.927 17:25:06 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=2 00:09:24.927 17:25:06 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:09:24.927 17:25:06 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 2 00:09:24.927 17:25:06 dpdk_mem_utility -- scripts/common.sh@366 -- # ver2[v]=2 00:09:24.927 17:25:06 dpdk_mem_utility -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:09:24.927 17:25:06 dpdk_mem_utility -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:09:24.927 17:25:06 dpdk_mem_utility -- scripts/common.sh@368 -- # return 0 00:09:24.927 17:25:06 dpdk_mem_utility -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:09:24.927 17:25:06 dpdk_mem_utility -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:09:24.927 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:24.927 --rc genhtml_branch_coverage=1 00:09:24.927 --rc genhtml_function_coverage=1 00:09:24.927 --rc genhtml_legend=1 00:09:24.927 --rc geninfo_all_blocks=1 00:09:24.927 --rc geninfo_unexecuted_blocks=1 00:09:24.927 00:09:24.927 ' 00:09:24.927 17:25:06 dpdk_mem_utility -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:09:24.927 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:24.927 --rc genhtml_branch_coverage=1 00:09:24.927 --rc genhtml_function_coverage=1 00:09:24.927 --rc genhtml_legend=1 00:09:24.927 --rc geninfo_all_blocks=1 00:09:24.927 --rc geninfo_unexecuted_blocks=1 00:09:24.927 00:09:24.927 ' 00:09:24.927 17:25:06 dpdk_mem_utility -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:09:24.927 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:24.927 --rc genhtml_branch_coverage=1 00:09:24.927 --rc genhtml_function_coverage=1 00:09:24.927 --rc genhtml_legend=1 00:09:24.927 --rc geninfo_all_blocks=1 00:09:24.927 --rc geninfo_unexecuted_blocks=1 00:09:24.927 00:09:24.927 ' 00:09:24.927 17:25:06 dpdk_mem_utility -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:09:24.927 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:24.927 --rc genhtml_branch_coverage=1 00:09:24.927 --rc genhtml_function_coverage=1 00:09:24.927 --rc genhtml_legend=1 00:09:24.927 --rc geninfo_all_blocks=1 00:09:24.927 --rc geninfo_unexecuted_blocks=1 00:09:24.927 00:09:24.927 ' 00:09:24.927 17:25:06 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@10 -- # MEM_SCRIPT=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/dpdk_mem_info.py 00:09:24.927 17:25:06 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@13 -- # spdkpid=116066 00:09:24.927 17:25:06 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:09:24.927 17:25:06 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@15 -- # waitforlisten 116066 00:09:24.927 17:25:06 dpdk_mem_utility -- common/autotest_common.sh@835 -- # '[' -z 116066 ']' 00:09:24.927 17:25:06 dpdk_mem_utility -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:24.927 17:25:06 dpdk_mem_utility -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:24.927 17:25:06 dpdk_mem_utility -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:24.927 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:24.927 17:25:06 dpdk_mem_utility -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:24.927 17:25:06 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:09:24.927 [2024-12-06 17:25:06.727398] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:09:24.927 [2024-12-06 17:25:06.727499] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid116066 ] 00:09:25.186 [2024-12-06 17:25:06.794411] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:25.186 [2024-12-06 17:25:06.841874] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:25.445 17:25:07 dpdk_mem_utility -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:25.445 17:25:07 dpdk_mem_utility -- common/autotest_common.sh@868 -- # return 0 00:09:25.445 17:25:07 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@17 -- # trap 'killprocess $spdkpid' SIGINT SIGTERM EXIT 00:09:25.445 17:25:07 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@19 -- # rpc_cmd env_dpdk_get_mem_stats 00:09:25.445 17:25:07 dpdk_mem_utility -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:25.445 17:25:07 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:09:25.445 { 00:09:25.445 "filename": "/tmp/spdk_mem_dump.txt" 00:09:25.445 } 00:09:25.445 17:25:07 dpdk_mem_utility -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:25.445 17:25:07 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/dpdk_mem_info.py 00:09:25.445 DPDK memory size 818.000000 MiB in 1 heap(s) 00:09:25.445 1 heaps totaling size 818.000000 MiB 00:09:25.445 size: 818.000000 MiB heap id: 0 00:09:25.445 end heaps---------- 00:09:25.445 9 mempools totaling size 603.782043 MiB 00:09:25.445 size: 212.674988 MiB name: PDU_immediate_data_Pool 00:09:25.445 size: 158.602051 MiB name: PDU_data_out_Pool 00:09:25.445 size: 100.555481 MiB name: bdev_io_116066 00:09:25.445 size: 50.003479 MiB name: msgpool_116066 00:09:25.445 size: 36.509338 MiB name: fsdev_io_116066 00:09:25.445 size: 21.763794 MiB name: PDU_Pool 00:09:25.445 size: 19.513306 MiB name: SCSI_TASK_Pool 00:09:25.445 size: 4.133484 MiB name: evtpool_116066 00:09:25.445 size: 0.026123 MiB name: Session_Pool 00:09:25.445 end mempools------- 00:09:25.445 6 memzones totaling size 4.142822 MiB 00:09:25.445 size: 1.000366 MiB name: RG_ring_0_116066 00:09:25.445 size: 1.000366 MiB name: RG_ring_1_116066 00:09:25.445 size: 1.000366 MiB name: RG_ring_4_116066 00:09:25.445 size: 1.000366 MiB name: RG_ring_5_116066 00:09:25.445 size: 0.125366 MiB name: RG_ring_2_116066 00:09:25.445 size: 0.015991 MiB name: RG_ring_3_116066 00:09:25.445 end memzones------- 00:09:25.445 17:25:07 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@23 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/dpdk_mem_info.py -m 0 00:09:25.445 heap id: 0 total size: 818.000000 MiB number of busy elements: 44 number of free elements: 15 00:09:25.445 list of free elements. size: 10.852478 MiB 00:09:25.445 element at address: 0x200019200000 with size: 0.999878 MiB 00:09:25.445 element at address: 0x200019400000 with size: 0.999878 MiB 00:09:25.445 element at address: 0x200000400000 with size: 0.998535 MiB 00:09:25.445 element at address: 0x200032000000 with size: 0.994446 MiB 00:09:25.445 element at address: 0x200006400000 with size: 0.959839 MiB 00:09:25.445 element at address: 0x200012c00000 with size: 0.944275 MiB 00:09:25.445 element at address: 0x200019600000 with size: 0.936584 MiB 00:09:25.445 element at address: 0x200000200000 with size: 0.717346 MiB 00:09:25.445 element at address: 0x20001ae00000 with size: 0.582886 MiB 00:09:25.445 element at address: 0x200000c00000 with size: 0.495422 MiB 00:09:25.445 element at address: 0x20000a600000 with size: 0.490723 MiB 00:09:25.445 element at address: 0x200019800000 with size: 0.485657 MiB 00:09:25.445 element at address: 0x200003e00000 with size: 0.481934 MiB 00:09:25.445 element at address: 0x200028200000 with size: 0.410034 MiB 00:09:25.445 element at address: 0x200000800000 with size: 0.355042 MiB 00:09:25.445 list of standard malloc elements. size: 199.218628 MiB 00:09:25.445 element at address: 0x20000a7fff80 with size: 132.000122 MiB 00:09:25.445 element at address: 0x2000065fff80 with size: 64.000122 MiB 00:09:25.445 element at address: 0x2000192fff80 with size: 1.000122 MiB 00:09:25.445 element at address: 0x2000194fff80 with size: 1.000122 MiB 00:09:25.445 element at address: 0x2000196fff80 with size: 1.000122 MiB 00:09:25.445 element at address: 0x2000003d9f00 with size: 0.140747 MiB 00:09:25.445 element at address: 0x2000196eff00 with size: 0.062622 MiB 00:09:25.445 element at address: 0x2000003fdf80 with size: 0.007935 MiB 00:09:25.445 element at address: 0x2000196efdc0 with size: 0.000305 MiB 00:09:25.445 element at address: 0x2000002d7c40 with size: 0.000183 MiB 00:09:25.445 element at address: 0x2000003d9e40 with size: 0.000183 MiB 00:09:25.445 element at address: 0x2000004ffa00 with size: 0.000183 MiB 00:09:25.445 element at address: 0x2000004ffac0 with size: 0.000183 MiB 00:09:25.445 element at address: 0x2000004ffb80 with size: 0.000183 MiB 00:09:25.445 element at address: 0x2000004ffd80 with size: 0.000183 MiB 00:09:25.445 element at address: 0x2000004ffe40 with size: 0.000183 MiB 00:09:25.445 element at address: 0x20000085ae40 with size: 0.000183 MiB 00:09:25.445 element at address: 0x20000085b040 with size: 0.000183 MiB 00:09:25.445 element at address: 0x20000085f300 with size: 0.000183 MiB 00:09:25.445 element at address: 0x20000087f5c0 with size: 0.000183 MiB 00:09:25.445 element at address: 0x20000087f680 with size: 0.000183 MiB 00:09:25.445 element at address: 0x2000008ff940 with size: 0.000183 MiB 00:09:25.445 element at address: 0x2000008ffb40 with size: 0.000183 MiB 00:09:25.445 element at address: 0x200000c7ed40 with size: 0.000183 MiB 00:09:25.445 element at address: 0x200000cff000 with size: 0.000183 MiB 00:09:25.445 element at address: 0x200000cff0c0 with size: 0.000183 MiB 00:09:25.445 element at address: 0x200003e7b600 with size: 0.000183 MiB 00:09:25.446 element at address: 0x200003e7b6c0 with size: 0.000183 MiB 00:09:25.446 element at address: 0x200003efb980 with size: 0.000183 MiB 00:09:25.446 element at address: 0x2000064fdd80 with size: 0.000183 MiB 00:09:25.446 element at address: 0x20000a67da00 with size: 0.000183 MiB 00:09:25.446 element at address: 0x20000a67dac0 with size: 0.000183 MiB 00:09:25.446 element at address: 0x20000a6fdd80 with size: 0.000183 MiB 00:09:25.446 element at address: 0x200012cf1bc0 with size: 0.000183 MiB 00:09:25.446 element at address: 0x2000196efc40 with size: 0.000183 MiB 00:09:25.446 element at address: 0x2000196efd00 with size: 0.000183 MiB 00:09:25.446 element at address: 0x2000198bc740 with size: 0.000183 MiB 00:09:25.446 element at address: 0x20001ae95380 with size: 0.000183 MiB 00:09:25.446 element at address: 0x20001ae95440 with size: 0.000183 MiB 00:09:25.446 element at address: 0x200028268f80 with size: 0.000183 MiB 00:09:25.446 element at address: 0x200028269040 with size: 0.000183 MiB 00:09:25.446 element at address: 0x20002826fc40 with size: 0.000183 MiB 00:09:25.446 element at address: 0x20002826fe40 with size: 0.000183 MiB 00:09:25.446 element at address: 0x20002826ff00 with size: 0.000183 MiB 00:09:25.446 list of memzone associated elements. size: 607.928894 MiB 00:09:25.446 element at address: 0x20001ae95500 with size: 211.416748 MiB 00:09:25.446 associated memzone info: size: 211.416626 MiB name: MP_PDU_immediate_data_Pool_0 00:09:25.446 element at address: 0x20002826ffc0 with size: 157.562561 MiB 00:09:25.446 associated memzone info: size: 157.562439 MiB name: MP_PDU_data_out_Pool_0 00:09:25.446 element at address: 0x200012df1e80 with size: 100.055054 MiB 00:09:25.446 associated memzone info: size: 100.054932 MiB name: MP_bdev_io_116066_0 00:09:25.446 element at address: 0x200000dff380 with size: 48.003052 MiB 00:09:25.446 associated memzone info: size: 48.002930 MiB name: MP_msgpool_116066_0 00:09:25.446 element at address: 0x200003ffdb80 with size: 36.008911 MiB 00:09:25.446 associated memzone info: size: 36.008789 MiB name: MP_fsdev_io_116066_0 00:09:25.446 element at address: 0x2000199be940 with size: 20.255554 MiB 00:09:25.446 associated memzone info: size: 20.255432 MiB name: MP_PDU_Pool_0 00:09:25.446 element at address: 0x2000321feb40 with size: 18.005066 MiB 00:09:25.446 associated memzone info: size: 18.004944 MiB name: MP_SCSI_TASK_Pool_0 00:09:25.446 element at address: 0x2000004fff00 with size: 3.000244 MiB 00:09:25.446 associated memzone info: size: 3.000122 MiB name: MP_evtpool_116066_0 00:09:25.446 element at address: 0x2000009ffe00 with size: 2.000488 MiB 00:09:25.446 associated memzone info: size: 2.000366 MiB name: RG_MP_msgpool_116066 00:09:25.446 element at address: 0x2000002d7d00 with size: 1.008118 MiB 00:09:25.446 associated memzone info: size: 1.007996 MiB name: MP_evtpool_116066 00:09:25.446 element at address: 0x20000a6fde40 with size: 1.008118 MiB 00:09:25.446 associated memzone info: size: 1.007996 MiB name: MP_PDU_Pool 00:09:25.446 element at address: 0x2000198bc800 with size: 1.008118 MiB 00:09:25.446 associated memzone info: size: 1.007996 MiB name: MP_PDU_immediate_data_Pool 00:09:25.446 element at address: 0x2000064fde40 with size: 1.008118 MiB 00:09:25.446 associated memzone info: size: 1.007996 MiB name: MP_PDU_data_out_Pool 00:09:25.446 element at address: 0x200003efba40 with size: 1.008118 MiB 00:09:25.446 associated memzone info: size: 1.007996 MiB name: MP_SCSI_TASK_Pool 00:09:25.446 element at address: 0x200000cff180 with size: 1.000488 MiB 00:09:25.446 associated memzone info: size: 1.000366 MiB name: RG_ring_0_116066 00:09:25.446 element at address: 0x2000008ffc00 with size: 1.000488 MiB 00:09:25.446 associated memzone info: size: 1.000366 MiB name: RG_ring_1_116066 00:09:25.446 element at address: 0x200012cf1c80 with size: 1.000488 MiB 00:09:25.446 associated memzone info: size: 1.000366 MiB name: RG_ring_4_116066 00:09:25.446 element at address: 0x2000320fe940 with size: 1.000488 MiB 00:09:25.446 associated memzone info: size: 1.000366 MiB name: RG_ring_5_116066 00:09:25.446 element at address: 0x20000087f740 with size: 0.500488 MiB 00:09:25.446 associated memzone info: size: 0.500366 MiB name: RG_MP_fsdev_io_116066 00:09:25.446 element at address: 0x200000c7ee00 with size: 0.500488 MiB 00:09:25.446 associated memzone info: size: 0.500366 MiB name: RG_MP_bdev_io_116066 00:09:25.446 element at address: 0x20000a67db80 with size: 0.500488 MiB 00:09:25.446 associated memzone info: size: 0.500366 MiB name: RG_MP_PDU_Pool 00:09:25.446 element at address: 0x200003e7b780 with size: 0.500488 MiB 00:09:25.446 associated memzone info: size: 0.500366 MiB name: RG_MP_SCSI_TASK_Pool 00:09:25.446 element at address: 0x20001987c540 with size: 0.250488 MiB 00:09:25.446 associated memzone info: size: 0.250366 MiB name: RG_MP_PDU_immediate_data_Pool 00:09:25.446 element at address: 0x2000002b7a40 with size: 0.125488 MiB 00:09:25.446 associated memzone info: size: 0.125366 MiB name: RG_MP_evtpool_116066 00:09:25.446 element at address: 0x20000085f3c0 with size: 0.125488 MiB 00:09:25.446 associated memzone info: size: 0.125366 MiB name: RG_ring_2_116066 00:09:25.446 element at address: 0x2000064f5b80 with size: 0.031738 MiB 00:09:25.446 associated memzone info: size: 0.031616 MiB name: RG_MP_PDU_data_out_Pool 00:09:25.446 element at address: 0x200028269100 with size: 0.023743 MiB 00:09:25.446 associated memzone info: size: 0.023621 MiB name: MP_Session_Pool_0 00:09:25.446 element at address: 0x20000085b100 with size: 0.016113 MiB 00:09:25.446 associated memzone info: size: 0.015991 MiB name: RG_ring_3_116066 00:09:25.446 element at address: 0x20002826f240 with size: 0.002441 MiB 00:09:25.446 associated memzone info: size: 0.002319 MiB name: RG_MP_Session_Pool 00:09:25.446 element at address: 0x2000004ffc40 with size: 0.000305 MiB 00:09:25.446 associated memzone info: size: 0.000183 MiB name: MP_msgpool_116066 00:09:25.446 element at address: 0x2000008ffa00 with size: 0.000305 MiB 00:09:25.446 associated memzone info: size: 0.000183 MiB name: MP_fsdev_io_116066 00:09:25.446 element at address: 0x20000085af00 with size: 0.000305 MiB 00:09:25.446 associated memzone info: size: 0.000183 MiB name: MP_bdev_io_116066 00:09:25.446 element at address: 0x20002826fd00 with size: 0.000305 MiB 00:09:25.446 associated memzone info: size: 0.000183 MiB name: MP_Session_Pool 00:09:25.446 17:25:07 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@25 -- # trap - SIGINT SIGTERM EXIT 00:09:25.446 17:25:07 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@26 -- # killprocess 116066 00:09:25.446 17:25:07 dpdk_mem_utility -- common/autotest_common.sh@954 -- # '[' -z 116066 ']' 00:09:25.446 17:25:07 dpdk_mem_utility -- common/autotest_common.sh@958 -- # kill -0 116066 00:09:25.446 17:25:07 dpdk_mem_utility -- common/autotest_common.sh@959 -- # uname 00:09:25.446 17:25:07 dpdk_mem_utility -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:25.446 17:25:07 dpdk_mem_utility -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 116066 00:09:25.446 17:25:07 dpdk_mem_utility -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:25.446 17:25:07 dpdk_mem_utility -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:25.446 17:25:07 dpdk_mem_utility -- common/autotest_common.sh@972 -- # echo 'killing process with pid 116066' 00:09:25.446 killing process with pid 116066 00:09:25.446 17:25:07 dpdk_mem_utility -- common/autotest_common.sh@973 -- # kill 116066 00:09:25.446 17:25:07 dpdk_mem_utility -- common/autotest_common.sh@978 -- # wait 116066 00:09:26.013 00:09:26.013 real 0m1.128s 00:09:26.013 user 0m1.089s 00:09:26.013 sys 0m0.433s 00:09:26.013 17:25:07 dpdk_mem_utility -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:26.013 17:25:07 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:09:26.013 ************************************ 00:09:26.013 END TEST dpdk_mem_utility 00:09:26.013 ************************************ 00:09:26.013 17:25:07 -- spdk/autotest.sh@168 -- # run_test event /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/event.sh 00:09:26.013 17:25:07 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:09:26.013 17:25:07 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:26.013 17:25:07 -- common/autotest_common.sh@10 -- # set +x 00:09:26.013 ************************************ 00:09:26.013 START TEST event 00:09:26.013 ************************************ 00:09:26.013 17:25:07 event -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/event.sh 00:09:26.013 * Looking for test storage... 00:09:26.013 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event 00:09:26.013 17:25:07 event -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:09:26.013 17:25:07 event -- common/autotest_common.sh@1711 -- # lcov --version 00:09:26.013 17:25:07 event -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:09:26.013 17:25:07 event -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:09:26.013 17:25:07 event -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:09:26.013 17:25:07 event -- scripts/common.sh@333 -- # local ver1 ver1_l 00:09:26.013 17:25:07 event -- scripts/common.sh@334 -- # local ver2 ver2_l 00:09:26.013 17:25:07 event -- scripts/common.sh@336 -- # IFS=.-: 00:09:26.013 17:25:07 event -- scripts/common.sh@336 -- # read -ra ver1 00:09:26.013 17:25:07 event -- scripts/common.sh@337 -- # IFS=.-: 00:09:26.013 17:25:07 event -- scripts/common.sh@337 -- # read -ra ver2 00:09:26.013 17:25:07 event -- scripts/common.sh@338 -- # local 'op=<' 00:09:26.013 17:25:07 event -- scripts/common.sh@340 -- # ver1_l=2 00:09:26.013 17:25:07 event -- scripts/common.sh@341 -- # ver2_l=1 00:09:26.013 17:25:07 event -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:09:26.013 17:25:07 event -- scripts/common.sh@344 -- # case "$op" in 00:09:26.013 17:25:07 event -- scripts/common.sh@345 -- # : 1 00:09:26.013 17:25:07 event -- scripts/common.sh@364 -- # (( v = 0 )) 00:09:26.013 17:25:07 event -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:09:26.013 17:25:07 event -- scripts/common.sh@365 -- # decimal 1 00:09:26.013 17:25:07 event -- scripts/common.sh@353 -- # local d=1 00:09:26.013 17:25:07 event -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:09:26.013 17:25:07 event -- scripts/common.sh@355 -- # echo 1 00:09:26.013 17:25:07 event -- scripts/common.sh@365 -- # ver1[v]=1 00:09:26.013 17:25:07 event -- scripts/common.sh@366 -- # decimal 2 00:09:26.013 17:25:07 event -- scripts/common.sh@353 -- # local d=2 00:09:26.013 17:25:07 event -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:09:26.013 17:25:07 event -- scripts/common.sh@355 -- # echo 2 00:09:26.013 17:25:07 event -- scripts/common.sh@366 -- # ver2[v]=2 00:09:26.013 17:25:07 event -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:09:26.013 17:25:07 event -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:09:26.013 17:25:07 event -- scripts/common.sh@368 -- # return 0 00:09:26.013 17:25:07 event -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:09:26.013 17:25:07 event -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:09:26.013 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:26.013 --rc genhtml_branch_coverage=1 00:09:26.013 --rc genhtml_function_coverage=1 00:09:26.013 --rc genhtml_legend=1 00:09:26.013 --rc geninfo_all_blocks=1 00:09:26.013 --rc geninfo_unexecuted_blocks=1 00:09:26.013 00:09:26.013 ' 00:09:26.013 17:25:07 event -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:09:26.013 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:26.013 --rc genhtml_branch_coverage=1 00:09:26.013 --rc genhtml_function_coverage=1 00:09:26.013 --rc genhtml_legend=1 00:09:26.013 --rc geninfo_all_blocks=1 00:09:26.013 --rc geninfo_unexecuted_blocks=1 00:09:26.013 00:09:26.013 ' 00:09:26.013 17:25:07 event -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:09:26.014 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:26.014 --rc genhtml_branch_coverage=1 00:09:26.014 --rc genhtml_function_coverage=1 00:09:26.014 --rc genhtml_legend=1 00:09:26.014 --rc geninfo_all_blocks=1 00:09:26.014 --rc geninfo_unexecuted_blocks=1 00:09:26.014 00:09:26.014 ' 00:09:26.014 17:25:07 event -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:09:26.014 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:26.014 --rc genhtml_branch_coverage=1 00:09:26.014 --rc genhtml_function_coverage=1 00:09:26.014 --rc genhtml_legend=1 00:09:26.014 --rc geninfo_all_blocks=1 00:09:26.014 --rc geninfo_unexecuted_blocks=1 00:09:26.014 00:09:26.014 ' 00:09:26.014 17:25:07 event -- event/event.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/bdev/nbd_common.sh 00:09:26.014 17:25:07 event -- bdev/nbd_common.sh@6 -- # set -e 00:09:26.014 17:25:07 event -- event/event.sh@45 -- # run_test event_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:09:26.014 17:25:07 event -- common/autotest_common.sh@1105 -- # '[' 6 -le 1 ']' 00:09:26.014 17:25:07 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:26.014 17:25:07 event -- common/autotest_common.sh@10 -- # set +x 00:09:26.272 ************************************ 00:09:26.272 START TEST event_perf 00:09:26.272 ************************************ 00:09:26.272 17:25:07 event.event_perf -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:09:26.272 Running I/O for 1 seconds...[2024-12-06 17:25:07.884066] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:09:26.272 [2024-12-06 17:25:07.884125] [ DPDK EAL parameters: event_perf --no-shconf -c 0xF --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid116262 ] 00:09:26.272 [2024-12-06 17:25:07.951369] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:09:26.272 [2024-12-06 17:25:08.003089] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:09:26.272 [2024-12-06 17:25:08.003154] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:09:26.272 [2024-12-06 17:25:08.003217] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:09:26.272 [2024-12-06 17:25:08.003220] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:27.691 Running I/O for 1 seconds... 00:09:27.691 lcore 0: 226824 00:09:27.691 lcore 1: 226822 00:09:27.692 lcore 2: 226822 00:09:27.692 lcore 3: 226822 00:09:27.692 done. 00:09:27.692 00:09:27.692 real 0m1.179s 00:09:27.692 user 0m4.098s 00:09:27.692 sys 0m0.077s 00:09:27.692 17:25:09 event.event_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:27.692 17:25:09 event.event_perf -- common/autotest_common.sh@10 -- # set +x 00:09:27.692 ************************************ 00:09:27.692 END TEST event_perf 00:09:27.692 ************************************ 00:09:27.692 17:25:09 event -- event/event.sh@46 -- # run_test event_reactor /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/reactor/reactor -t 1 00:09:27.692 17:25:09 event -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:09:27.692 17:25:09 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:27.692 17:25:09 event -- common/autotest_common.sh@10 -- # set +x 00:09:27.692 ************************************ 00:09:27.692 START TEST event_reactor 00:09:27.692 ************************************ 00:09:27.692 17:25:09 event.event_reactor -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/reactor/reactor -t 1 00:09:27.692 [2024-12-06 17:25:09.114895] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:09:27.692 [2024-12-06 17:25:09.114957] [ DPDK EAL parameters: reactor --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid116420 ] 00:09:27.692 [2024-12-06 17:25:09.181072] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:27.692 [2024-12-06 17:25:09.228287] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:28.625 test_start 00:09:28.625 oneshot 00:09:28.625 tick 100 00:09:28.625 tick 100 00:09:28.625 tick 250 00:09:28.625 tick 100 00:09:28.625 tick 100 00:09:28.625 tick 250 00:09:28.625 tick 500 00:09:28.625 tick 100 00:09:28.625 tick 100 00:09:28.625 tick 100 00:09:28.625 tick 250 00:09:28.625 tick 100 00:09:28.625 tick 100 00:09:28.625 test_end 00:09:28.625 00:09:28.625 real 0m1.172s 00:09:28.625 user 0m1.099s 00:09:28.625 sys 0m0.068s 00:09:28.625 17:25:10 event.event_reactor -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:28.625 17:25:10 event.event_reactor -- common/autotest_common.sh@10 -- # set +x 00:09:28.625 ************************************ 00:09:28.625 END TEST event_reactor 00:09:28.625 ************************************ 00:09:28.625 17:25:10 event -- event/event.sh@47 -- # run_test event_reactor_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/reactor_perf/reactor_perf -t 1 00:09:28.625 17:25:10 event -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:09:28.625 17:25:10 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:28.625 17:25:10 event -- common/autotest_common.sh@10 -- # set +x 00:09:28.625 ************************************ 00:09:28.625 START TEST event_reactor_perf 00:09:28.625 ************************************ 00:09:28.625 17:25:10 event.event_reactor_perf -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/reactor_perf/reactor_perf -t 1 00:09:28.625 [2024-12-06 17:25:10.337809] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:09:28.625 [2024-12-06 17:25:10.337875] [ DPDK EAL parameters: reactor_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid116591 ] 00:09:28.625 [2024-12-06 17:25:10.405850] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:28.625 [2024-12-06 17:25:10.455364] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:30.016 test_start 00:09:30.016 test_end 00:09:30.016 Performance: 442643 events per second 00:09:30.016 00:09:30.016 real 0m1.175s 00:09:30.016 user 0m1.098s 00:09:30.016 sys 0m0.073s 00:09:30.016 17:25:11 event.event_reactor_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:30.016 17:25:11 event.event_reactor_perf -- common/autotest_common.sh@10 -- # set +x 00:09:30.016 ************************************ 00:09:30.016 END TEST event_reactor_perf 00:09:30.016 ************************************ 00:09:30.016 17:25:11 event -- event/event.sh@49 -- # uname -s 00:09:30.016 17:25:11 event -- event/event.sh@49 -- # '[' Linux = Linux ']' 00:09:30.016 17:25:11 event -- event/event.sh@50 -- # run_test event_scheduler /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/scheduler/scheduler.sh 00:09:30.016 17:25:11 event -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:09:30.016 17:25:11 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:30.016 17:25:11 event -- common/autotest_common.sh@10 -- # set +x 00:09:30.016 ************************************ 00:09:30.016 START TEST event_scheduler 00:09:30.016 ************************************ 00:09:30.016 17:25:11 event.event_scheduler -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/scheduler/scheduler.sh 00:09:30.016 * Looking for test storage... 00:09:30.016 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/scheduler 00:09:30.016 17:25:11 event.event_scheduler -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:09:30.016 17:25:11 event.event_scheduler -- common/autotest_common.sh@1711 -- # lcov --version 00:09:30.016 17:25:11 event.event_scheduler -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:09:30.016 17:25:11 event.event_scheduler -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:09:30.016 17:25:11 event.event_scheduler -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:09:30.016 17:25:11 event.event_scheduler -- scripts/common.sh@333 -- # local ver1 ver1_l 00:09:30.016 17:25:11 event.event_scheduler -- scripts/common.sh@334 -- # local ver2 ver2_l 00:09:30.016 17:25:11 event.event_scheduler -- scripts/common.sh@336 -- # IFS=.-: 00:09:30.016 17:25:11 event.event_scheduler -- scripts/common.sh@336 -- # read -ra ver1 00:09:30.016 17:25:11 event.event_scheduler -- scripts/common.sh@337 -- # IFS=.-: 00:09:30.016 17:25:11 event.event_scheduler -- scripts/common.sh@337 -- # read -ra ver2 00:09:30.016 17:25:11 event.event_scheduler -- scripts/common.sh@338 -- # local 'op=<' 00:09:30.016 17:25:11 event.event_scheduler -- scripts/common.sh@340 -- # ver1_l=2 00:09:30.016 17:25:11 event.event_scheduler -- scripts/common.sh@341 -- # ver2_l=1 00:09:30.016 17:25:11 event.event_scheduler -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:09:30.016 17:25:11 event.event_scheduler -- scripts/common.sh@344 -- # case "$op" in 00:09:30.016 17:25:11 event.event_scheduler -- scripts/common.sh@345 -- # : 1 00:09:30.016 17:25:11 event.event_scheduler -- scripts/common.sh@364 -- # (( v = 0 )) 00:09:30.016 17:25:11 event.event_scheduler -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:09:30.016 17:25:11 event.event_scheduler -- scripts/common.sh@365 -- # decimal 1 00:09:30.016 17:25:11 event.event_scheduler -- scripts/common.sh@353 -- # local d=1 00:09:30.016 17:25:11 event.event_scheduler -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:09:30.016 17:25:11 event.event_scheduler -- scripts/common.sh@355 -- # echo 1 00:09:30.016 17:25:11 event.event_scheduler -- scripts/common.sh@365 -- # ver1[v]=1 00:09:30.016 17:25:11 event.event_scheduler -- scripts/common.sh@366 -- # decimal 2 00:09:30.016 17:25:11 event.event_scheduler -- scripts/common.sh@353 -- # local d=2 00:09:30.016 17:25:11 event.event_scheduler -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:09:30.016 17:25:11 event.event_scheduler -- scripts/common.sh@355 -- # echo 2 00:09:30.016 17:25:11 event.event_scheduler -- scripts/common.sh@366 -- # ver2[v]=2 00:09:30.016 17:25:11 event.event_scheduler -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:09:30.016 17:25:11 event.event_scheduler -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:09:30.016 17:25:11 event.event_scheduler -- scripts/common.sh@368 -- # return 0 00:09:30.016 17:25:11 event.event_scheduler -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:09:30.016 17:25:11 event.event_scheduler -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:09:30.016 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:30.016 --rc genhtml_branch_coverage=1 00:09:30.016 --rc genhtml_function_coverage=1 00:09:30.016 --rc genhtml_legend=1 00:09:30.016 --rc geninfo_all_blocks=1 00:09:30.016 --rc geninfo_unexecuted_blocks=1 00:09:30.016 00:09:30.016 ' 00:09:30.016 17:25:11 event.event_scheduler -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:09:30.016 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:30.016 --rc genhtml_branch_coverage=1 00:09:30.016 --rc genhtml_function_coverage=1 00:09:30.016 --rc genhtml_legend=1 00:09:30.016 --rc geninfo_all_blocks=1 00:09:30.016 --rc geninfo_unexecuted_blocks=1 00:09:30.016 00:09:30.016 ' 00:09:30.016 17:25:11 event.event_scheduler -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:09:30.016 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:30.016 --rc genhtml_branch_coverage=1 00:09:30.016 --rc genhtml_function_coverage=1 00:09:30.016 --rc genhtml_legend=1 00:09:30.016 --rc geninfo_all_blocks=1 00:09:30.016 --rc geninfo_unexecuted_blocks=1 00:09:30.016 00:09:30.016 ' 00:09:30.016 17:25:11 event.event_scheduler -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:09:30.016 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:30.016 --rc genhtml_branch_coverage=1 00:09:30.016 --rc genhtml_function_coverage=1 00:09:30.016 --rc genhtml_legend=1 00:09:30.016 --rc geninfo_all_blocks=1 00:09:30.016 --rc geninfo_unexecuted_blocks=1 00:09:30.016 00:09:30.016 ' 00:09:30.016 17:25:11 event.event_scheduler -- scheduler/scheduler.sh@29 -- # rpc=rpc_cmd 00:09:30.016 17:25:11 event.event_scheduler -- scheduler/scheduler.sh@35 -- # scheduler_pid=116886 00:09:30.016 17:25:11 event.event_scheduler -- scheduler/scheduler.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/scheduler/scheduler -m 0xF -p 0x2 --wait-for-rpc -f 00:09:30.017 17:25:11 event.event_scheduler -- scheduler/scheduler.sh@36 -- # trap 'killprocess $scheduler_pid; exit 1' SIGINT SIGTERM EXIT 00:09:30.017 17:25:11 event.event_scheduler -- scheduler/scheduler.sh@37 -- # waitforlisten 116886 00:09:30.017 17:25:11 event.event_scheduler -- common/autotest_common.sh@835 -- # '[' -z 116886 ']' 00:09:30.017 17:25:11 event.event_scheduler -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:30.017 17:25:11 event.event_scheduler -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:30.017 17:25:11 event.event_scheduler -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:30.017 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:30.017 17:25:11 event.event_scheduler -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:30.017 17:25:11 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:09:30.017 [2024-12-06 17:25:11.737315] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:09:30.017 [2024-12-06 17:25:11.737396] [ DPDK EAL parameters: scheduler --no-shconf -c 0xF --main-lcore=2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid116886 ] 00:09:30.017 [2024-12-06 17:25:11.803987] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:09:30.276 [2024-12-06 17:25:11.854107] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:30.276 [2024-12-06 17:25:11.854169] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:09:30.276 [2024-12-06 17:25:11.854239] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:09:30.276 [2024-12-06 17:25:11.854247] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:09:30.276 17:25:11 event.event_scheduler -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:30.276 17:25:11 event.event_scheduler -- common/autotest_common.sh@868 -- # return 0 00:09:30.276 17:25:11 event.event_scheduler -- scheduler/scheduler.sh@39 -- # rpc_cmd framework_set_scheduler dynamic 00:09:30.276 17:25:11 event.event_scheduler -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.276 17:25:11 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:09:30.276 [2024-12-06 17:25:11.967312] dpdk_governor.c: 178:_init: *ERROR*: App core mask contains some but not all of a set of SMT siblings 00:09:30.276 [2024-12-06 17:25:11.967340] scheduler_dynamic.c: 280:init: *NOTICE*: Unable to initialize dpdk governor 00:09:30.276 [2024-12-06 17:25:11.967373] scheduler_dynamic.c: 427:set_opts: *NOTICE*: Setting scheduler load limit to 20 00:09:30.276 [2024-12-06 17:25:11.967390] scheduler_dynamic.c: 429:set_opts: *NOTICE*: Setting scheduler core limit to 80 00:09:30.276 [2024-12-06 17:25:11.967401] scheduler_dynamic.c: 431:set_opts: *NOTICE*: Setting scheduler core busy to 95 00:09:30.276 17:25:11 event.event_scheduler -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.276 17:25:11 event.event_scheduler -- scheduler/scheduler.sh@40 -- # rpc_cmd framework_start_init 00:09:30.276 17:25:11 event.event_scheduler -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.276 17:25:11 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:09:30.276 [2024-12-06 17:25:12.066243] scheduler.c: 382:test_start: *NOTICE*: Scheduler test application started. 00:09:30.276 17:25:12 event.event_scheduler -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.276 17:25:12 event.event_scheduler -- scheduler/scheduler.sh@43 -- # run_test scheduler_create_thread scheduler_create_thread 00:09:30.276 17:25:12 event.event_scheduler -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:09:30.276 17:25:12 event.event_scheduler -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:30.276 17:25:12 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:09:30.276 ************************************ 00:09:30.276 START TEST scheduler_create_thread 00:09:30.276 ************************************ 00:09:30.276 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1129 -- # scheduler_create_thread 00:09:30.276 17:25:12 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@12 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x1 -a 100 00:09:30.276 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.276 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:09:30.276 2 00:09:30.276 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.276 17:25:12 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@13 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x2 -a 100 00:09:30.276 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.276 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:09:30.535 3 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@14 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x4 -a 100 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:09:30.535 4 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@15 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x8 -a 100 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:09:30.535 5 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@16 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x1 -a 0 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:09:30.535 6 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@17 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x2 -a 0 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:09:30.535 7 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@18 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x4 -a 0 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:09:30.535 8 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@19 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x8 -a 0 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:09:30.535 9 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@21 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n one_third_active -a 30 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:09:30.535 10 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n half_active -a 0 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # thread_id=11 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@23 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_set_active 11 50 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n deleted -a 100 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # thread_id=12 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@26 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_delete 12 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:30.535 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:09:31.099 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:31.099 00:09:31.099 real 0m0.592s 00:09:31.099 user 0m0.012s 00:09:31.099 sys 0m0.002s 00:09:31.099 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:31.099 17:25:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:09:31.099 ************************************ 00:09:31.099 END TEST scheduler_create_thread 00:09:31.099 ************************************ 00:09:31.099 17:25:12 event.event_scheduler -- scheduler/scheduler.sh@45 -- # trap - SIGINT SIGTERM EXIT 00:09:31.099 17:25:12 event.event_scheduler -- scheduler/scheduler.sh@46 -- # killprocess 116886 00:09:31.099 17:25:12 event.event_scheduler -- common/autotest_common.sh@954 -- # '[' -z 116886 ']' 00:09:31.099 17:25:12 event.event_scheduler -- common/autotest_common.sh@958 -- # kill -0 116886 00:09:31.099 17:25:12 event.event_scheduler -- common/autotest_common.sh@959 -- # uname 00:09:31.099 17:25:12 event.event_scheduler -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:31.099 17:25:12 event.event_scheduler -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 116886 00:09:31.099 17:25:12 event.event_scheduler -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:09:31.099 17:25:12 event.event_scheduler -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:09:31.099 17:25:12 event.event_scheduler -- common/autotest_common.sh@972 -- # echo 'killing process with pid 116886' 00:09:31.099 killing process with pid 116886 00:09:31.099 17:25:12 event.event_scheduler -- common/autotest_common.sh@973 -- # kill 116886 00:09:31.099 17:25:12 event.event_scheduler -- common/autotest_common.sh@978 -- # wait 116886 00:09:31.357 [2024-12-06 17:25:13.170372] scheduler.c: 360:test_shutdown: *NOTICE*: Scheduler test application stopped. 00:09:31.616 00:09:31.616 real 0m1.818s 00:09:31.616 user 0m2.497s 00:09:31.616 sys 0m0.361s 00:09:31.616 17:25:13 event.event_scheduler -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:31.616 17:25:13 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:09:31.616 ************************************ 00:09:31.616 END TEST event_scheduler 00:09:31.616 ************************************ 00:09:31.616 17:25:13 event -- event/event.sh@51 -- # modprobe -n nbd 00:09:31.616 17:25:13 event -- event/event.sh@52 -- # run_test app_repeat app_repeat_test 00:09:31.616 17:25:13 event -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:09:31.616 17:25:13 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:31.616 17:25:13 event -- common/autotest_common.sh@10 -- # set +x 00:09:31.616 ************************************ 00:09:31.616 START TEST app_repeat 00:09:31.616 ************************************ 00:09:31.616 17:25:13 event.app_repeat -- common/autotest_common.sh@1129 -- # app_repeat_test 00:09:31.616 17:25:13 event.app_repeat -- event/event.sh@12 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:09:31.616 17:25:13 event.app_repeat -- event/event.sh@13 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:09:31.616 17:25:13 event.app_repeat -- event/event.sh@13 -- # local nbd_list 00:09:31.616 17:25:13 event.app_repeat -- event/event.sh@14 -- # bdev_list=('Malloc0' 'Malloc1') 00:09:31.616 17:25:13 event.app_repeat -- event/event.sh@14 -- # local bdev_list 00:09:31.616 17:25:13 event.app_repeat -- event/event.sh@15 -- # local repeat_times=4 00:09:31.616 17:25:13 event.app_repeat -- event/event.sh@17 -- # modprobe nbd 00:09:31.616 17:25:13 event.app_repeat -- event/event.sh@19 -- # repeat_pid=117078 00:09:31.616 17:25:13 event.app_repeat -- event/event.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/app_repeat/app_repeat -r /var/tmp/spdk-nbd.sock -m 0x3 -t 4 00:09:31.616 17:25:13 event.app_repeat -- event/event.sh@20 -- # trap 'killprocess $repeat_pid; exit 1' SIGINT SIGTERM EXIT 00:09:31.616 17:25:13 event.app_repeat -- event/event.sh@21 -- # echo 'Process app_repeat pid: 117078' 00:09:31.616 Process app_repeat pid: 117078 00:09:31.616 17:25:13 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:09:31.616 17:25:13 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 0' 00:09:31.616 spdk_app_start Round 0 00:09:31.616 17:25:13 event.app_repeat -- event/event.sh@25 -- # waitforlisten 117078 /var/tmp/spdk-nbd.sock 00:09:31.616 17:25:13 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 117078 ']' 00:09:31.616 17:25:13 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:09:31.616 17:25:13 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:31.616 17:25:13 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:09:31.616 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:09:31.616 17:25:13 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:31.616 17:25:13 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:09:31.616 [2024-12-06 17:25:13.445298] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:09:31.616 [2024-12-06 17:25:13.445361] [ DPDK EAL parameters: app_repeat --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid117078 ] 00:09:31.875 [2024-12-06 17:25:13.511359] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:09:31.875 [2024-12-06 17:25:13.556479] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:09:31.875 [2024-12-06 17:25:13.556482] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:31.875 17:25:13 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:31.875 17:25:13 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:09:31.875 17:25:13 event.app_repeat -- event/event.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:09:32.144 Malloc0 00:09:32.144 17:25:13 event.app_repeat -- event/event.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:09:32.713 Malloc1 00:09:32.713 17:25:14 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:09:32.713 17:25:14 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:09:32.713 17:25:14 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:09:32.713 17:25:14 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:09:32.713 17:25:14 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:09:32.713 17:25:14 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:09:32.713 17:25:14 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:09:32.713 17:25:14 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:09:32.713 17:25:14 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:09:32.713 17:25:14 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:09:32.713 17:25:14 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:09:32.713 17:25:14 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:09:32.713 17:25:14 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:09:32.713 17:25:14 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:09:32.713 17:25:14 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:09:32.713 17:25:14 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:09:32.972 /dev/nbd0 00:09:32.972 17:25:14 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:09:32.972 17:25:14 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:09:32.972 17:25:14 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:09:32.972 17:25:14 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:09:32.972 17:25:14 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:09:32.972 17:25:14 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:09:32.972 17:25:14 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:09:32.972 17:25:14 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:09:32.972 17:25:14 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:09:32.972 17:25:14 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:09:32.972 17:25:14 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:09:32.972 1+0 records in 00:09:32.972 1+0 records out 00:09:32.972 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000222604 s, 18.4 MB/s 00:09:32.972 17:25:14 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:09:32.972 17:25:14 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:09:32.972 17:25:14 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:09:32.972 17:25:14 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:09:32.972 17:25:14 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:09:32.972 17:25:14 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:09:32.972 17:25:14 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:09:32.972 17:25:14 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:09:33.230 /dev/nbd1 00:09:33.230 17:25:14 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:09:33.230 17:25:14 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:09:33.230 17:25:14 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:09:33.230 17:25:14 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:09:33.230 17:25:14 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:09:33.230 17:25:14 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:09:33.230 17:25:14 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:09:33.230 17:25:14 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:09:33.230 17:25:14 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:09:33.230 17:25:14 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:09:33.230 17:25:14 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:09:33.230 1+0 records in 00:09:33.230 1+0 records out 00:09:33.230 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000208952 s, 19.6 MB/s 00:09:33.230 17:25:14 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:09:33.230 17:25:14 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:09:33.230 17:25:14 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:09:33.230 17:25:14 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:09:33.230 17:25:14 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:09:33.230 17:25:14 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:09:33.230 17:25:14 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:09:33.230 17:25:14 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:09:33.230 17:25:14 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:09:33.230 17:25:14 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:09:33.488 { 00:09:33.488 "nbd_device": "/dev/nbd0", 00:09:33.488 "bdev_name": "Malloc0" 00:09:33.488 }, 00:09:33.488 { 00:09:33.488 "nbd_device": "/dev/nbd1", 00:09:33.488 "bdev_name": "Malloc1" 00:09:33.488 } 00:09:33.488 ]' 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:09:33.488 { 00:09:33.488 "nbd_device": "/dev/nbd0", 00:09:33.488 "bdev_name": "Malloc0" 00:09:33.488 }, 00:09:33.488 { 00:09:33.488 "nbd_device": "/dev/nbd1", 00:09:33.488 "bdev_name": "Malloc1" 00:09:33.488 } 00:09:33.488 ]' 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:09:33.488 /dev/nbd1' 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:09:33.488 /dev/nbd1' 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest bs=4096 count=256 00:09:33.488 256+0 records in 00:09:33.488 256+0 records out 00:09:33.488 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00383577 s, 273 MB/s 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:09:33.488 256+0 records in 00:09:33.488 256+0 records out 00:09:33.488 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0199139 s, 52.7 MB/s 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:09:33.488 256+0 records in 00:09:33.488 256+0 records out 00:09:33.488 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0216525 s, 48.4 MB/s 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd0 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd1 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:33.488 17:25:15 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:09:34.054 17:25:15 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:09:34.054 17:25:15 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:09:34.054 17:25:15 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:09:34.054 17:25:15 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:34.054 17:25:15 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:34.054 17:25:15 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:09:34.055 17:25:15 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:09:34.055 17:25:15 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:09:34.055 17:25:15 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:34.055 17:25:15 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:09:34.313 17:25:15 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:09:34.313 17:25:15 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:09:34.313 17:25:15 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:09:34.313 17:25:15 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:34.313 17:25:15 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:34.313 17:25:15 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:09:34.313 17:25:15 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:09:34.313 17:25:15 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:09:34.313 17:25:15 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:09:34.313 17:25:15 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:09:34.313 17:25:15 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:09:34.572 17:25:16 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:09:34.572 17:25:16 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:09:34.572 17:25:16 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:09:34.572 17:25:16 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:09:34.572 17:25:16 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:09:34.572 17:25:16 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:09:34.572 17:25:16 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:09:34.572 17:25:16 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:09:34.572 17:25:16 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:09:34.572 17:25:16 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:09:34.572 17:25:16 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:09:34.572 17:25:16 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:09:34.572 17:25:16 event.app_repeat -- event/event.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:09:34.831 17:25:16 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:09:35.090 [2024-12-06 17:25:16.702734] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:09:35.091 [2024-12-06 17:25:16.746405] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:09:35.091 [2024-12-06 17:25:16.746405] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:35.091 [2024-12-06 17:25:16.801784] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:09:35.091 [2024-12-06 17:25:16.801848] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:09:38.386 17:25:19 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:09:38.386 17:25:19 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 1' 00:09:38.386 spdk_app_start Round 1 00:09:38.386 17:25:19 event.app_repeat -- event/event.sh@25 -- # waitforlisten 117078 /var/tmp/spdk-nbd.sock 00:09:38.386 17:25:19 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 117078 ']' 00:09:38.386 17:25:19 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:09:38.386 17:25:19 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:38.386 17:25:19 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:09:38.386 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:09:38.386 17:25:19 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:38.386 17:25:19 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:09:38.386 17:25:19 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:38.386 17:25:19 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:09:38.386 17:25:19 event.app_repeat -- event/event.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:09:38.386 Malloc0 00:09:38.386 17:25:20 event.app_repeat -- event/event.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:09:38.645 Malloc1 00:09:38.645 17:25:20 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:09:38.645 17:25:20 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:09:38.645 17:25:20 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:09:38.645 17:25:20 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:09:38.645 17:25:20 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:09:38.645 17:25:20 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:09:38.645 17:25:20 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:09:38.645 17:25:20 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:09:38.645 17:25:20 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:09:38.645 17:25:20 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:09:38.645 17:25:20 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:09:38.645 17:25:20 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:09:38.645 17:25:20 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:09:38.645 17:25:20 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:09:38.645 17:25:20 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:09:38.645 17:25:20 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:09:38.905 /dev/nbd0 00:09:38.905 17:25:20 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:09:38.905 17:25:20 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:09:38.905 17:25:20 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:09:38.905 17:25:20 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:09:38.905 17:25:20 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:09:38.905 17:25:20 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:09:38.905 17:25:20 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:09:38.905 17:25:20 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:09:38.905 17:25:20 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:09:38.905 17:25:20 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:09:38.905 17:25:20 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:09:38.905 1+0 records in 00:09:38.905 1+0 records out 00:09:38.905 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000224469 s, 18.2 MB/s 00:09:38.905 17:25:20 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:09:38.905 17:25:20 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:09:38.905 17:25:20 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:09:38.905 17:25:20 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:09:38.905 17:25:20 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:09:38.905 17:25:20 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:09:38.905 17:25:20 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:09:38.905 17:25:20 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:09:39.164 /dev/nbd1 00:09:39.422 17:25:21 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:09:39.422 17:25:21 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:09:39.422 17:25:21 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:09:39.422 17:25:21 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:09:39.422 17:25:21 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:09:39.422 17:25:21 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:09:39.422 17:25:21 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:09:39.423 17:25:21 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:09:39.423 17:25:21 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:09:39.423 17:25:21 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:09:39.423 17:25:21 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:09:39.423 1+0 records in 00:09:39.423 1+0 records out 00:09:39.423 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000214435 s, 19.1 MB/s 00:09:39.423 17:25:21 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:09:39.423 17:25:21 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:09:39.423 17:25:21 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:09:39.423 17:25:21 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:09:39.423 17:25:21 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:09:39.423 17:25:21 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:09:39.423 17:25:21 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:09:39.423 17:25:21 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:09:39.423 17:25:21 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:09:39.423 17:25:21 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:09:39.682 17:25:21 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:09:39.682 { 00:09:39.682 "nbd_device": "/dev/nbd0", 00:09:39.682 "bdev_name": "Malloc0" 00:09:39.682 }, 00:09:39.682 { 00:09:39.682 "nbd_device": "/dev/nbd1", 00:09:39.682 "bdev_name": "Malloc1" 00:09:39.682 } 00:09:39.682 ]' 00:09:39.682 17:25:21 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:09:39.682 { 00:09:39.682 "nbd_device": "/dev/nbd0", 00:09:39.682 "bdev_name": "Malloc0" 00:09:39.682 }, 00:09:39.682 { 00:09:39.682 "nbd_device": "/dev/nbd1", 00:09:39.682 "bdev_name": "Malloc1" 00:09:39.682 } 00:09:39.682 ]' 00:09:39.682 17:25:21 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:09:39.682 17:25:21 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:09:39.682 /dev/nbd1' 00:09:39.682 17:25:21 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:09:39.682 /dev/nbd1' 00:09:39.682 17:25:21 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:09:39.682 17:25:21 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:09:39.682 17:25:21 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:09:39.682 17:25:21 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:09:39.682 17:25:21 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:09:39.682 17:25:21 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:09:39.682 17:25:21 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:09:39.682 17:25:21 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:09:39.682 17:25:21 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:09:39.682 17:25:21 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:09:39.682 17:25:21 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:09:39.682 17:25:21 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest bs=4096 count=256 00:09:39.682 256+0 records in 00:09:39.682 256+0 records out 00:09:39.682 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00426781 s, 246 MB/s 00:09:39.682 17:25:21 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:09:39.682 17:25:21 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:09:39.682 256+0 records in 00:09:39.682 256+0 records out 00:09:39.683 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0203086 s, 51.6 MB/s 00:09:39.683 17:25:21 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:09:39.683 17:25:21 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:09:39.683 256+0 records in 00:09:39.683 256+0 records out 00:09:39.683 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0227266 s, 46.1 MB/s 00:09:39.683 17:25:21 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:09:39.683 17:25:21 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:09:39.683 17:25:21 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:09:39.683 17:25:21 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:09:39.683 17:25:21 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:09:39.683 17:25:21 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:09:39.683 17:25:21 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:09:39.683 17:25:21 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:09:39.683 17:25:21 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd0 00:09:39.683 17:25:21 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:09:39.683 17:25:21 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd1 00:09:39.683 17:25:21 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:09:39.683 17:25:21 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:09:39.683 17:25:21 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:09:39.683 17:25:21 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:09:39.683 17:25:21 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:09:39.683 17:25:21 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:09:39.683 17:25:21 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:39.683 17:25:21 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:09:39.941 17:25:21 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:09:39.941 17:25:21 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:09:39.941 17:25:21 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:09:39.941 17:25:21 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:39.941 17:25:21 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:39.941 17:25:21 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:09:39.941 17:25:21 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:09:39.941 17:25:21 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:09:39.941 17:25:21 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:39.941 17:25:21 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:09:40.200 17:25:21 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:09:40.200 17:25:21 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:09:40.200 17:25:21 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:09:40.200 17:25:21 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:40.200 17:25:21 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:40.200 17:25:21 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:09:40.200 17:25:21 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:09:40.200 17:25:21 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:09:40.200 17:25:21 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:09:40.200 17:25:21 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:09:40.200 17:25:21 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:09:40.459 17:25:22 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:09:40.459 17:25:22 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:09:40.459 17:25:22 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:09:40.718 17:25:22 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:09:40.718 17:25:22 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:09:40.718 17:25:22 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:09:40.718 17:25:22 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:09:40.718 17:25:22 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:09:40.718 17:25:22 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:09:40.718 17:25:22 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:09:40.718 17:25:22 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:09:40.718 17:25:22 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:09:40.718 17:25:22 event.app_repeat -- event/event.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:09:40.977 17:25:22 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:09:40.977 [2024-12-06 17:25:22.781396] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:09:41.236 [2024-12-06 17:25:22.827362] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:41.236 [2024-12-06 17:25:22.827362] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:09:41.236 [2024-12-06 17:25:22.886667] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:09:41.236 [2024-12-06 17:25:22.886759] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:09:44.518 17:25:25 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:09:44.518 17:25:25 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 2' 00:09:44.518 spdk_app_start Round 2 00:09:44.518 17:25:25 event.app_repeat -- event/event.sh@25 -- # waitforlisten 117078 /var/tmp/spdk-nbd.sock 00:09:44.518 17:25:25 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 117078 ']' 00:09:44.518 17:25:25 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:09:44.518 17:25:25 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:44.518 17:25:25 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:09:44.518 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:09:44.518 17:25:25 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:44.518 17:25:25 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:09:44.518 17:25:25 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:44.518 17:25:25 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:09:44.518 17:25:25 event.app_repeat -- event/event.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:09:44.518 Malloc0 00:09:44.518 17:25:26 event.app_repeat -- event/event.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:09:44.784 Malloc1 00:09:44.784 17:25:26 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:09:44.784 17:25:26 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:09:44.784 17:25:26 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:09:44.784 17:25:26 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:09:44.784 17:25:26 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:09:44.784 17:25:26 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:09:44.784 17:25:26 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:09:44.784 17:25:26 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:09:44.784 17:25:26 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:09:44.784 17:25:26 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:09:44.784 17:25:26 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:09:44.784 17:25:26 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:09:44.784 17:25:26 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:09:44.784 17:25:26 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:09:44.784 17:25:26 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:09:44.784 17:25:26 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:09:45.042 /dev/nbd0 00:09:45.042 17:25:26 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:09:45.042 17:25:26 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:09:45.042 17:25:26 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd0 00:09:45.042 17:25:26 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:09:45.042 17:25:26 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:09:45.042 17:25:26 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:09:45.042 17:25:26 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd0 /proc/partitions 00:09:45.042 17:25:26 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:09:45.042 17:25:26 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:09:45.042 17:25:26 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:09:45.042 17:25:26 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:09:45.042 1+0 records in 00:09:45.042 1+0 records out 00:09:45.042 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000263768 s, 15.5 MB/s 00:09:45.042 17:25:26 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:09:45.042 17:25:26 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:09:45.042 17:25:26 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:09:45.042 17:25:26 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:09:45.042 17:25:26 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:09:45.042 17:25:26 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:09:45.042 17:25:26 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:09:45.042 17:25:26 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:09:45.301 /dev/nbd1 00:09:45.301 17:25:27 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:09:45.301 17:25:27 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:09:45.301 17:25:27 event.app_repeat -- common/autotest_common.sh@872 -- # local nbd_name=nbd1 00:09:45.301 17:25:27 event.app_repeat -- common/autotest_common.sh@873 -- # local i 00:09:45.301 17:25:27 event.app_repeat -- common/autotest_common.sh@875 -- # (( i = 1 )) 00:09:45.301 17:25:27 event.app_repeat -- common/autotest_common.sh@875 -- # (( i <= 20 )) 00:09:45.301 17:25:27 event.app_repeat -- common/autotest_common.sh@876 -- # grep -q -w nbd1 /proc/partitions 00:09:45.301 17:25:27 event.app_repeat -- common/autotest_common.sh@877 -- # break 00:09:45.301 17:25:27 event.app_repeat -- common/autotest_common.sh@888 -- # (( i = 1 )) 00:09:45.301 17:25:27 event.app_repeat -- common/autotest_common.sh@888 -- # (( i <= 20 )) 00:09:45.301 17:25:27 event.app_repeat -- common/autotest_common.sh@889 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:09:45.301 1+0 records in 00:09:45.301 1+0 records out 00:09:45.301 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000164488 s, 24.9 MB/s 00:09:45.301 17:25:27 event.app_repeat -- common/autotest_common.sh@890 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:09:45.301 17:25:27 event.app_repeat -- common/autotest_common.sh@890 -- # size=4096 00:09:45.301 17:25:27 event.app_repeat -- common/autotest_common.sh@891 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:09:45.301 17:25:27 event.app_repeat -- common/autotest_common.sh@892 -- # '[' 4096 '!=' 0 ']' 00:09:45.301 17:25:27 event.app_repeat -- common/autotest_common.sh@893 -- # return 0 00:09:45.301 17:25:27 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:09:45.301 17:25:27 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:09:45.301 17:25:27 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:09:45.301 17:25:27 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:09:45.301 17:25:27 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:09:45.559 17:25:27 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:09:45.559 { 00:09:45.559 "nbd_device": "/dev/nbd0", 00:09:45.559 "bdev_name": "Malloc0" 00:09:45.559 }, 00:09:45.559 { 00:09:45.559 "nbd_device": "/dev/nbd1", 00:09:45.559 "bdev_name": "Malloc1" 00:09:45.559 } 00:09:45.559 ]' 00:09:45.559 17:25:27 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:09:45.559 { 00:09:45.559 "nbd_device": "/dev/nbd0", 00:09:45.559 "bdev_name": "Malloc0" 00:09:45.559 }, 00:09:45.559 { 00:09:45.559 "nbd_device": "/dev/nbd1", 00:09:45.559 "bdev_name": "Malloc1" 00:09:45.559 } 00:09:45.559 ]' 00:09:45.559 17:25:27 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:09:45.817 /dev/nbd1' 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:09:45.817 /dev/nbd1' 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest bs=4096 count=256 00:09:45.817 256+0 records in 00:09:45.817 256+0 records out 00:09:45.817 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0050155 s, 209 MB/s 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:09:45.817 256+0 records in 00:09:45.817 256+0 records out 00:09:45.817 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0199208 s, 52.6 MB/s 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:09:45.817 256+0 records in 00:09:45.817 256+0 records out 00:09:45.817 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.021777 s, 48.2 MB/s 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd0 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd1 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:45.817 17:25:27 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:09:46.075 17:25:27 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:09:46.075 17:25:27 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:09:46.075 17:25:27 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:09:46.075 17:25:27 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:46.075 17:25:27 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:46.075 17:25:27 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:09:46.075 17:25:27 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:09:46.075 17:25:27 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:09:46.075 17:25:27 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:09:46.075 17:25:27 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:09:46.332 17:25:28 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:09:46.332 17:25:28 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:09:46.332 17:25:28 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:09:46.332 17:25:28 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:09:46.332 17:25:28 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:09:46.332 17:25:28 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:09:46.332 17:25:28 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:09:46.332 17:25:28 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:09:46.332 17:25:28 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:09:46.332 17:25:28 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:09:46.332 17:25:28 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:09:46.589 17:25:28 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:09:46.589 17:25:28 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:09:46.589 17:25:28 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:09:46.589 17:25:28 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:09:46.589 17:25:28 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:09:46.589 17:25:28 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:09:46.589 17:25:28 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:09:46.589 17:25:28 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:09:46.589 17:25:28 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:09:46.589 17:25:28 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:09:46.589 17:25:28 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:09:46.589 17:25:28 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:09:46.589 17:25:28 event.app_repeat -- event/event.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:09:46.847 17:25:28 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:09:47.104 [2024-12-06 17:25:28.866935] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:09:47.104 [2024-12-06 17:25:28.908955] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:09:47.104 [2024-12-06 17:25:28.908958] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:47.363 [2024-12-06 17:25:28.967297] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:09:47.363 [2024-12-06 17:25:28.967359] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:09:49.891 17:25:31 event.app_repeat -- event/event.sh@38 -- # waitforlisten 117078 /var/tmp/spdk-nbd.sock 00:09:49.891 17:25:31 event.app_repeat -- common/autotest_common.sh@835 -- # '[' -z 117078 ']' 00:09:49.891 17:25:31 event.app_repeat -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:09:49.891 17:25:31 event.app_repeat -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:49.891 17:25:31 event.app_repeat -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:09:49.891 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:09:49.891 17:25:31 event.app_repeat -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:49.891 17:25:31 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:09:50.148 17:25:31 event.app_repeat -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:50.148 17:25:31 event.app_repeat -- common/autotest_common.sh@868 -- # return 0 00:09:50.148 17:25:31 event.app_repeat -- event/event.sh@39 -- # killprocess 117078 00:09:50.148 17:25:31 event.app_repeat -- common/autotest_common.sh@954 -- # '[' -z 117078 ']' 00:09:50.148 17:25:31 event.app_repeat -- common/autotest_common.sh@958 -- # kill -0 117078 00:09:50.148 17:25:31 event.app_repeat -- common/autotest_common.sh@959 -- # uname 00:09:50.148 17:25:31 event.app_repeat -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:50.148 17:25:31 event.app_repeat -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 117078 00:09:50.148 17:25:31 event.app_repeat -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:50.148 17:25:31 event.app_repeat -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:50.148 17:25:31 event.app_repeat -- common/autotest_common.sh@972 -- # echo 'killing process with pid 117078' 00:09:50.148 killing process with pid 117078 00:09:50.148 17:25:31 event.app_repeat -- common/autotest_common.sh@973 -- # kill 117078 00:09:50.148 17:25:31 event.app_repeat -- common/autotest_common.sh@978 -- # wait 117078 00:09:50.408 spdk_app_start is called in Round 0. 00:09:50.408 Shutdown signal received, stop current app iteration 00:09:50.408 Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 reinitialization... 00:09:50.408 spdk_app_start is called in Round 1. 00:09:50.408 Shutdown signal received, stop current app iteration 00:09:50.408 Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 reinitialization... 00:09:50.408 spdk_app_start is called in Round 2. 00:09:50.408 Shutdown signal received, stop current app iteration 00:09:50.408 Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 reinitialization... 00:09:50.408 spdk_app_start is called in Round 3. 00:09:50.408 Shutdown signal received, stop current app iteration 00:09:50.408 17:25:32 event.app_repeat -- event/event.sh@40 -- # trap - SIGINT SIGTERM EXIT 00:09:50.408 17:25:32 event.app_repeat -- event/event.sh@42 -- # return 0 00:09:50.408 00:09:50.408 real 0m18.741s 00:09:50.408 user 0m41.522s 00:09:50.408 sys 0m3.250s 00:09:50.408 17:25:32 event.app_repeat -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:50.408 17:25:32 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:09:50.408 ************************************ 00:09:50.408 END TEST app_repeat 00:09:50.408 ************************************ 00:09:50.408 17:25:32 event -- event/event.sh@54 -- # (( SPDK_TEST_CRYPTO == 0 )) 00:09:50.408 17:25:32 event -- event/event.sh@55 -- # run_test cpu_locks /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/cpu_locks.sh 00:09:50.408 17:25:32 event -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:09:50.408 17:25:32 event -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:50.408 17:25:32 event -- common/autotest_common.sh@10 -- # set +x 00:09:50.408 ************************************ 00:09:50.408 START TEST cpu_locks 00:09:50.408 ************************************ 00:09:50.409 17:25:32 event.cpu_locks -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/cpu_locks.sh 00:09:50.668 * Looking for test storage... 00:09:50.668 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event 00:09:50.668 17:25:32 event.cpu_locks -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:09:50.668 17:25:32 event.cpu_locks -- common/autotest_common.sh@1711 -- # lcov --version 00:09:50.668 17:25:32 event.cpu_locks -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:09:50.668 17:25:32 event.cpu_locks -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:09:50.668 17:25:32 event.cpu_locks -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:09:50.668 17:25:32 event.cpu_locks -- scripts/common.sh@333 -- # local ver1 ver1_l 00:09:50.668 17:25:32 event.cpu_locks -- scripts/common.sh@334 -- # local ver2 ver2_l 00:09:50.668 17:25:32 event.cpu_locks -- scripts/common.sh@336 -- # IFS=.-: 00:09:50.668 17:25:32 event.cpu_locks -- scripts/common.sh@336 -- # read -ra ver1 00:09:50.668 17:25:32 event.cpu_locks -- scripts/common.sh@337 -- # IFS=.-: 00:09:50.668 17:25:32 event.cpu_locks -- scripts/common.sh@337 -- # read -ra ver2 00:09:50.668 17:25:32 event.cpu_locks -- scripts/common.sh@338 -- # local 'op=<' 00:09:50.668 17:25:32 event.cpu_locks -- scripts/common.sh@340 -- # ver1_l=2 00:09:50.668 17:25:32 event.cpu_locks -- scripts/common.sh@341 -- # ver2_l=1 00:09:50.668 17:25:32 event.cpu_locks -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:09:50.668 17:25:32 event.cpu_locks -- scripts/common.sh@344 -- # case "$op" in 00:09:50.668 17:25:32 event.cpu_locks -- scripts/common.sh@345 -- # : 1 00:09:50.668 17:25:32 event.cpu_locks -- scripts/common.sh@364 -- # (( v = 0 )) 00:09:50.668 17:25:32 event.cpu_locks -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:09:50.668 17:25:32 event.cpu_locks -- scripts/common.sh@365 -- # decimal 1 00:09:50.668 17:25:32 event.cpu_locks -- scripts/common.sh@353 -- # local d=1 00:09:50.668 17:25:32 event.cpu_locks -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:09:50.668 17:25:32 event.cpu_locks -- scripts/common.sh@355 -- # echo 1 00:09:50.668 17:25:32 event.cpu_locks -- scripts/common.sh@365 -- # ver1[v]=1 00:09:50.668 17:25:32 event.cpu_locks -- scripts/common.sh@366 -- # decimal 2 00:09:50.668 17:25:32 event.cpu_locks -- scripts/common.sh@353 -- # local d=2 00:09:50.668 17:25:32 event.cpu_locks -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:09:50.668 17:25:32 event.cpu_locks -- scripts/common.sh@355 -- # echo 2 00:09:50.668 17:25:32 event.cpu_locks -- scripts/common.sh@366 -- # ver2[v]=2 00:09:50.668 17:25:32 event.cpu_locks -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:09:50.668 17:25:32 event.cpu_locks -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:09:50.668 17:25:32 event.cpu_locks -- scripts/common.sh@368 -- # return 0 00:09:50.668 17:25:32 event.cpu_locks -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:09:50.668 17:25:32 event.cpu_locks -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:09:50.668 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:50.668 --rc genhtml_branch_coverage=1 00:09:50.668 --rc genhtml_function_coverage=1 00:09:50.668 --rc genhtml_legend=1 00:09:50.668 --rc geninfo_all_blocks=1 00:09:50.668 --rc geninfo_unexecuted_blocks=1 00:09:50.668 00:09:50.668 ' 00:09:50.668 17:25:32 event.cpu_locks -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:09:50.668 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:50.668 --rc genhtml_branch_coverage=1 00:09:50.668 --rc genhtml_function_coverage=1 00:09:50.668 --rc genhtml_legend=1 00:09:50.668 --rc geninfo_all_blocks=1 00:09:50.668 --rc geninfo_unexecuted_blocks=1 00:09:50.668 00:09:50.668 ' 00:09:50.668 17:25:32 event.cpu_locks -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:09:50.668 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:50.668 --rc genhtml_branch_coverage=1 00:09:50.668 --rc genhtml_function_coverage=1 00:09:50.668 --rc genhtml_legend=1 00:09:50.668 --rc geninfo_all_blocks=1 00:09:50.668 --rc geninfo_unexecuted_blocks=1 00:09:50.668 00:09:50.668 ' 00:09:50.668 17:25:32 event.cpu_locks -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:09:50.668 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:50.668 --rc genhtml_branch_coverage=1 00:09:50.668 --rc genhtml_function_coverage=1 00:09:50.668 --rc genhtml_legend=1 00:09:50.668 --rc geninfo_all_blocks=1 00:09:50.668 --rc geninfo_unexecuted_blocks=1 00:09:50.668 00:09:50.668 ' 00:09:50.668 17:25:32 event.cpu_locks -- event/cpu_locks.sh@11 -- # rpc_sock1=/var/tmp/spdk.sock 00:09:50.668 17:25:32 event.cpu_locks -- event/cpu_locks.sh@12 -- # rpc_sock2=/var/tmp/spdk2.sock 00:09:50.668 17:25:32 event.cpu_locks -- event/cpu_locks.sh@164 -- # trap cleanup EXIT SIGTERM SIGINT 00:09:50.668 17:25:32 event.cpu_locks -- event/cpu_locks.sh@166 -- # run_test default_locks default_locks 00:09:50.668 17:25:32 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:09:50.668 17:25:32 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:50.668 17:25:32 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:09:50.668 ************************************ 00:09:50.668 START TEST default_locks 00:09:50.668 ************************************ 00:09:50.668 17:25:32 event.cpu_locks.default_locks -- common/autotest_common.sh@1129 -- # default_locks 00:09:50.668 17:25:32 event.cpu_locks.default_locks -- event/cpu_locks.sh@46 -- # spdk_tgt_pid=119570 00:09:50.669 17:25:32 event.cpu_locks.default_locks -- event/cpu_locks.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:09:50.669 17:25:32 event.cpu_locks.default_locks -- event/cpu_locks.sh@47 -- # waitforlisten 119570 00:09:50.669 17:25:32 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # '[' -z 119570 ']' 00:09:50.669 17:25:32 event.cpu_locks.default_locks -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:50.669 17:25:32 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:50.669 17:25:32 event.cpu_locks.default_locks -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:50.669 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:50.669 17:25:32 event.cpu_locks.default_locks -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:50.669 17:25:32 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:09:50.669 [2024-12-06 17:25:32.439820] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:09:50.669 [2024-12-06 17:25:32.439899] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid119570 ] 00:09:50.928 [2024-12-06 17:25:32.510348] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:50.928 [2024-12-06 17:25:32.554509] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:51.186 17:25:32 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:51.186 17:25:32 event.cpu_locks.default_locks -- common/autotest_common.sh@868 -- # return 0 00:09:51.186 17:25:32 event.cpu_locks.default_locks -- event/cpu_locks.sh@49 -- # locks_exist 119570 00:09:51.186 17:25:32 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # lslocks -p 119570 00:09:51.186 17:25:32 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:09:51.444 lslocks: write error 00:09:51.444 17:25:33 event.cpu_locks.default_locks -- event/cpu_locks.sh@50 -- # killprocess 119570 00:09:51.444 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@954 -- # '[' -z 119570 ']' 00:09:51.444 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@958 -- # kill -0 119570 00:09:51.444 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@959 -- # uname 00:09:51.444 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:51.444 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 119570 00:09:51.444 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:51.444 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:51.444 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@972 -- # echo 'killing process with pid 119570' 00:09:51.444 killing process with pid 119570 00:09:51.444 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@973 -- # kill 119570 00:09:51.444 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@978 -- # wait 119570 00:09:52.012 17:25:33 event.cpu_locks.default_locks -- event/cpu_locks.sh@52 -- # NOT waitforlisten 119570 00:09:52.012 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@652 -- # local es=0 00:09:52.012 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@654 -- # valid_exec_arg waitforlisten 119570 00:09:52.012 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@640 -- # local arg=waitforlisten 00:09:52.012 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:52.012 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@644 -- # type -t waitforlisten 00:09:52.012 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:52.012 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@655 -- # waitforlisten 119570 00:09:52.012 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # '[' -z 119570 ']' 00:09:52.012 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:52.012 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:52.012 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:52.012 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:52.012 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:52.012 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:09:52.012 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 850: kill: (119570) - No such process 00:09:52.012 ERROR: process (pid: 119570) is no longer running 00:09:52.012 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:52.012 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@868 -- # return 1 00:09:52.012 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@655 -- # es=1 00:09:52.012 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:09:52.012 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:09:52.012 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:09:52.012 17:25:33 event.cpu_locks.default_locks -- event/cpu_locks.sh@54 -- # no_locks 00:09:52.012 17:25:33 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # lock_files=() 00:09:52.012 17:25:33 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # local lock_files 00:09:52.012 17:25:33 event.cpu_locks.default_locks -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:09:52.012 00:09:52.012 real 0m1.209s 00:09:52.012 user 0m1.161s 00:09:52.012 sys 0m0.532s 00:09:52.012 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:52.012 17:25:33 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:09:52.012 ************************************ 00:09:52.012 END TEST default_locks 00:09:52.012 ************************************ 00:09:52.012 17:25:33 event.cpu_locks -- event/cpu_locks.sh@167 -- # run_test default_locks_via_rpc default_locks_via_rpc 00:09:52.012 17:25:33 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:09:52.012 17:25:33 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:52.012 17:25:33 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:09:52.012 ************************************ 00:09:52.012 START TEST default_locks_via_rpc 00:09:52.012 ************************************ 00:09:52.012 17:25:33 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1129 -- # default_locks_via_rpc 00:09:52.012 17:25:33 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@62 -- # spdk_tgt_pid=119732 00:09:52.012 17:25:33 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:09:52.012 17:25:33 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@63 -- # waitforlisten 119732 00:09:52.012 17:25:33 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 119732 ']' 00:09:52.012 17:25:33 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:52.012 17:25:33 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:52.012 17:25:33 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:52.012 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:52.012 17:25:33 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:52.012 17:25:33 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:09:52.012 [2024-12-06 17:25:33.704636] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:09:52.012 [2024-12-06 17:25:33.704746] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid119732 ] 00:09:52.012 [2024-12-06 17:25:33.768414] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:52.012 [2024-12-06 17:25:33.810169] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:52.271 17:25:34 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:52.271 17:25:34 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:09:52.271 17:25:34 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@65 -- # rpc_cmd framework_disable_cpumask_locks 00:09:52.271 17:25:34 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.271 17:25:34 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:09:52.271 17:25:34 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.271 17:25:34 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@67 -- # no_locks 00:09:52.271 17:25:34 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # lock_files=() 00:09:52.271 17:25:34 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # local lock_files 00:09:52.271 17:25:34 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:09:52.271 17:25:34 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@69 -- # rpc_cmd framework_enable_cpumask_locks 00:09:52.271 17:25:34 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:09:52.271 17:25:34 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:09:52.271 17:25:34 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:09:52.271 17:25:34 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@71 -- # locks_exist 119732 00:09:52.271 17:25:34 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # lslocks -p 119732 00:09:52.271 17:25:34 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:09:52.529 17:25:34 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@73 -- # killprocess 119732 00:09:52.529 17:25:34 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@954 -- # '[' -z 119732 ']' 00:09:52.529 17:25:34 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@958 -- # kill -0 119732 00:09:52.529 17:25:34 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@959 -- # uname 00:09:52.529 17:25:34 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:52.529 17:25:34 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 119732 00:09:52.788 17:25:34 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:52.788 17:25:34 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:52.788 17:25:34 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 119732' 00:09:52.788 killing process with pid 119732 00:09:52.788 17:25:34 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@973 -- # kill 119732 00:09:52.788 17:25:34 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@978 -- # wait 119732 00:09:53.049 00:09:53.049 real 0m1.123s 00:09:53.049 user 0m1.094s 00:09:53.049 sys 0m0.495s 00:09:53.049 17:25:34 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:53.049 17:25:34 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:09:53.049 ************************************ 00:09:53.049 END TEST default_locks_via_rpc 00:09:53.049 ************************************ 00:09:53.049 17:25:34 event.cpu_locks -- event/cpu_locks.sh@168 -- # run_test non_locking_app_on_locked_coremask non_locking_app_on_locked_coremask 00:09:53.049 17:25:34 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:09:53.049 17:25:34 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:53.049 17:25:34 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:09:53.049 ************************************ 00:09:53.049 START TEST non_locking_app_on_locked_coremask 00:09:53.049 ************************************ 00:09:53.049 17:25:34 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1129 -- # non_locking_app_on_locked_coremask 00:09:53.049 17:25:34 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@80 -- # spdk_tgt_pid=119898 00:09:53.049 17:25:34 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:09:53.049 17:25:34 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@81 -- # waitforlisten 119898 /var/tmp/spdk.sock 00:09:53.049 17:25:34 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 119898 ']' 00:09:53.049 17:25:34 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:53.049 17:25:34 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:53.049 17:25:34 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:53.049 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:53.049 17:25:34 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:53.049 17:25:34 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:09:53.049 [2024-12-06 17:25:34.878438] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:09:53.049 [2024-12-06 17:25:34.878525] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid119898 ] 00:09:53.309 [2024-12-06 17:25:34.944288] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:53.309 [2024-12-06 17:25:34.991073] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:53.567 17:25:35 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:53.567 17:25:35 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 0 00:09:53.567 17:25:35 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@84 -- # spdk_tgt_pid2=120021 00:09:53.567 17:25:35 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks -r /var/tmp/spdk2.sock 00:09:53.567 17:25:35 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@85 -- # waitforlisten 120021 /var/tmp/spdk2.sock 00:09:53.567 17:25:35 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 120021 ']' 00:09:53.567 17:25:35 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:09:53.567 17:25:35 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:53.567 17:25:35 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:09:53.567 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:09:53.567 17:25:35 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:53.567 17:25:35 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:09:53.567 [2024-12-06 17:25:35.303306] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:09:53.567 [2024-12-06 17:25:35.303382] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid120021 ] 00:09:53.825 [2024-12-06 17:25:35.407278] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:09:53.825 [2024-12-06 17:25:35.407321] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:53.825 [2024-12-06 17:25:35.503555] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:54.393 17:25:35 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:54.393 17:25:35 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 0 00:09:54.393 17:25:35 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@87 -- # locks_exist 119898 00:09:54.393 17:25:35 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 119898 00:09:54.393 17:25:35 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:09:54.652 lslocks: write error 00:09:54.652 17:25:36 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@89 -- # killprocess 119898 00:09:54.652 17:25:36 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # '[' -z 119898 ']' 00:09:54.652 17:25:36 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # kill -0 119898 00:09:54.652 17:25:36 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # uname 00:09:54.652 17:25:36 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:54.652 17:25:36 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 119898 00:09:54.652 17:25:36 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:54.652 17:25:36 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:54.652 17:25:36 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 119898' 00:09:54.652 killing process with pid 119898 00:09:54.652 17:25:36 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@973 -- # kill 119898 00:09:54.652 17:25:36 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@978 -- # wait 119898 00:09:55.590 17:25:37 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@90 -- # killprocess 120021 00:09:55.590 17:25:37 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # '[' -z 120021 ']' 00:09:55.590 17:25:37 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # kill -0 120021 00:09:55.590 17:25:37 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # uname 00:09:55.590 17:25:37 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:55.590 17:25:37 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 120021 00:09:55.590 17:25:37 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:55.590 17:25:37 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:55.590 17:25:37 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 120021' 00:09:55.590 killing process with pid 120021 00:09:55.590 17:25:37 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@973 -- # kill 120021 00:09:55.590 17:25:37 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@978 -- # wait 120021 00:09:55.850 00:09:55.850 real 0m2.807s 00:09:55.850 user 0m2.820s 00:09:55.850 sys 0m1.002s 00:09:55.850 17:25:37 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:55.850 17:25:37 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:09:55.850 ************************************ 00:09:55.850 END TEST non_locking_app_on_locked_coremask 00:09:55.850 ************************************ 00:09:55.850 17:25:37 event.cpu_locks -- event/cpu_locks.sh@169 -- # run_test locking_app_on_unlocked_coremask locking_app_on_unlocked_coremask 00:09:55.850 17:25:37 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:09:55.850 17:25:37 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:55.850 17:25:37 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:09:55.850 ************************************ 00:09:55.850 START TEST locking_app_on_unlocked_coremask 00:09:55.850 ************************************ 00:09:55.850 17:25:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1129 -- # locking_app_on_unlocked_coremask 00:09:55.850 17:25:37 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@98 -- # spdk_tgt_pid=120319 00:09:55.850 17:25:37 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@97 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks 00:09:55.850 17:25:37 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@99 -- # waitforlisten 120319 /var/tmp/spdk.sock 00:09:55.850 17:25:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # '[' -z 120319 ']' 00:09:55.850 17:25:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:55.850 17:25:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:55.850 17:25:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:55.850 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:55.850 17:25:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:55.850 17:25:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:09:56.110 [2024-12-06 17:25:37.739775] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:09:56.110 [2024-12-06 17:25:37.739856] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid120319 ] 00:09:56.110 [2024-12-06 17:25:37.807016] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:09:56.110 [2024-12-06 17:25:37.807048] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:56.110 [2024-12-06 17:25:37.851888] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:56.368 17:25:38 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:56.368 17:25:38 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@868 -- # return 0 00:09:56.368 17:25:38 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@102 -- # spdk_tgt_pid2=120333 00:09:56.368 17:25:38 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@103 -- # waitforlisten 120333 /var/tmp/spdk2.sock 00:09:56.369 17:25:38 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@101 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:09:56.369 17:25:38 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # '[' -z 120333 ']' 00:09:56.369 17:25:38 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:09:56.369 17:25:38 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:56.369 17:25:38 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:09:56.369 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:09:56.369 17:25:38 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:56.369 17:25:38 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:09:56.369 [2024-12-06 17:25:38.158276] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:09:56.369 [2024-12-06 17:25:38.158360] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid120333 ] 00:09:56.627 [2024-12-06 17:25:38.255403] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:56.627 [2024-12-06 17:25:38.349545] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:57.195 17:25:38 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:57.195 17:25:38 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@868 -- # return 0 00:09:57.195 17:25:38 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@105 -- # locks_exist 120333 00:09:57.195 17:25:38 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 120333 00:09:57.195 17:25:38 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:09:57.453 lslocks: write error 00:09:57.453 17:25:39 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@107 -- # killprocess 120319 00:09:57.453 17:25:39 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # '[' -z 120319 ']' 00:09:57.453 17:25:39 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@958 -- # kill -0 120319 00:09:57.453 17:25:39 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # uname 00:09:57.453 17:25:39 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:57.453 17:25:39 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 120319 00:09:57.713 17:25:39 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:57.713 17:25:39 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:57.713 17:25:39 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 120319' 00:09:57.713 killing process with pid 120319 00:09:57.713 17:25:39 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@973 -- # kill 120319 00:09:57.713 17:25:39 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@978 -- # wait 120319 00:09:58.281 17:25:40 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@108 -- # killprocess 120333 00:09:58.281 17:25:40 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # '[' -z 120333 ']' 00:09:58.281 17:25:40 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@958 -- # kill -0 120333 00:09:58.281 17:25:40 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # uname 00:09:58.281 17:25:40 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:09:58.281 17:25:40 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 120333 00:09:58.281 17:25:40 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:09:58.281 17:25:40 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:09:58.281 17:25:40 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 120333' 00:09:58.281 killing process with pid 120333 00:09:58.281 17:25:40 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@973 -- # kill 120333 00:09:58.281 17:25:40 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@978 -- # wait 120333 00:09:58.848 00:09:58.848 real 0m2.782s 00:09:58.848 user 0m2.804s 00:09:58.848 sys 0m0.976s 00:09:58.848 17:25:40 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:09:58.848 17:25:40 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:09:58.848 ************************************ 00:09:58.848 END TEST locking_app_on_unlocked_coremask 00:09:58.848 ************************************ 00:09:58.848 17:25:40 event.cpu_locks -- event/cpu_locks.sh@170 -- # run_test locking_app_on_locked_coremask locking_app_on_locked_coremask 00:09:58.848 17:25:40 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:09:58.848 17:25:40 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:09:58.848 17:25:40 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:09:58.848 ************************************ 00:09:58.848 START TEST locking_app_on_locked_coremask 00:09:58.848 ************************************ 00:09:58.848 17:25:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1129 -- # locking_app_on_locked_coremask 00:09:58.848 17:25:40 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@115 -- # spdk_tgt_pid=120631 00:09:58.848 17:25:40 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@114 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:09:58.848 17:25:40 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@116 -- # waitforlisten 120631 /var/tmp/spdk.sock 00:09:58.848 17:25:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 120631 ']' 00:09:58.848 17:25:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:58.848 17:25:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:58.848 17:25:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:58.848 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:58.848 17:25:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:58.848 17:25:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:09:58.848 [2024-12-06 17:25:40.577431] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:09:58.848 [2024-12-06 17:25:40.577526] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid120631 ] 00:09:58.848 [2024-12-06 17:25:40.644698] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:09:59.108 [2024-12-06 17:25:40.693957] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:09:59.367 17:25:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:59.367 17:25:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 0 00:09:59.367 17:25:40 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@119 -- # spdk_tgt_pid2=120760 00:09:59.367 17:25:40 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@118 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:09:59.367 17:25:40 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@120 -- # NOT waitforlisten 120760 /var/tmp/spdk2.sock 00:09:59.367 17:25:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@652 -- # local es=0 00:09:59.367 17:25:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@654 -- # valid_exec_arg waitforlisten 120760 /var/tmp/spdk2.sock 00:09:59.367 17:25:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@640 -- # local arg=waitforlisten 00:09:59.367 17:25:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:59.367 17:25:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@644 -- # type -t waitforlisten 00:09:59.367 17:25:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:09:59.367 17:25:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@655 -- # waitforlisten 120760 /var/tmp/spdk2.sock 00:09:59.367 17:25:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # '[' -z 120760 ']' 00:09:59.367 17:25:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:09:59.367 17:25:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:09:59.367 17:25:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:09:59.367 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:09:59.367 17:25:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:09:59.367 17:25:40 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:09:59.367 [2024-12-06 17:25:41.006365] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:09:59.367 [2024-12-06 17:25:41.006453] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid120760 ] 00:09:59.367 [2024-12-06 17:25:41.106030] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 0, probably process 120631 has claimed it. 00:09:59.367 [2024-12-06 17:25:41.106099] app.c: 912:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:09:59.933 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 850: kill: (120760) - No such process 00:09:59.933 ERROR: process (pid: 120760) is no longer running 00:09:59.933 17:25:41 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:09:59.933 17:25:41 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@868 -- # return 1 00:09:59.933 17:25:41 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@655 -- # es=1 00:09:59.933 17:25:41 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:09:59.933 17:25:41 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:09:59.933 17:25:41 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:09:59.934 17:25:41 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@122 -- # locks_exist 120631 00:09:59.934 17:25:41 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 120631 00:09:59.934 17:25:41 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:10:00.192 lslocks: write error 00:10:00.192 17:25:42 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@124 -- # killprocess 120631 00:10:00.192 17:25:42 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # '[' -z 120631 ']' 00:10:00.192 17:25:42 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # kill -0 120631 00:10:00.192 17:25:42 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # uname 00:10:00.192 17:25:42 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:00.192 17:25:42 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 120631 00:10:00.451 17:25:42 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:00.451 17:25:42 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:00.451 17:25:42 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 120631' 00:10:00.451 killing process with pid 120631 00:10:00.451 17:25:42 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@973 -- # kill 120631 00:10:00.451 17:25:42 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@978 -- # wait 120631 00:10:00.712 00:10:00.712 real 0m1.912s 00:10:00.712 user 0m2.122s 00:10:00.712 sys 0m0.627s 00:10:00.712 17:25:42 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:00.712 17:25:42 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:10:00.712 ************************************ 00:10:00.712 END TEST locking_app_on_locked_coremask 00:10:00.712 ************************************ 00:10:00.712 17:25:42 event.cpu_locks -- event/cpu_locks.sh@171 -- # run_test locking_overlapped_coremask locking_overlapped_coremask 00:10:00.712 17:25:42 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:10:00.712 17:25:42 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:00.712 17:25:42 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:10:00.712 ************************************ 00:10:00.712 START TEST locking_overlapped_coremask 00:10:00.712 ************************************ 00:10:00.712 17:25:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1129 -- # locking_overlapped_coremask 00:10:00.712 17:25:42 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@132 -- # spdk_tgt_pid=120929 00:10:00.712 17:25:42 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@131 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x7 00:10:00.712 17:25:42 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@133 -- # waitforlisten 120929 /var/tmp/spdk.sock 00:10:00.712 17:25:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # '[' -z 120929 ']' 00:10:00.712 17:25:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:00.712 17:25:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:00.712 17:25:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:00.712 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:00.712 17:25:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:00.712 17:25:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:10:00.712 [2024-12-06 17:25:42.543170] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:10:00.712 [2024-12-06 17:25:42.543280] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid120929 ] 00:10:00.973 [2024-12-06 17:25:42.607501] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:10:00.973 [2024-12-06 17:25:42.651355] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:10:00.973 [2024-12-06 17:25:42.651461] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:10:00.973 [2024-12-06 17:25:42.651469] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:01.232 17:25:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:01.233 17:25:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@868 -- # return 0 00:10:01.233 17:25:42 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@136 -- # spdk_tgt_pid2=121022 00:10:01.233 17:25:42 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@135 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock 00:10:01.233 17:25:42 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@137 -- # NOT waitforlisten 121022 /var/tmp/spdk2.sock 00:10:01.233 17:25:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@652 -- # local es=0 00:10:01.233 17:25:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@654 -- # valid_exec_arg waitforlisten 121022 /var/tmp/spdk2.sock 00:10:01.233 17:25:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@640 -- # local arg=waitforlisten 00:10:01.233 17:25:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:01.233 17:25:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@644 -- # type -t waitforlisten 00:10:01.233 17:25:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:01.233 17:25:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@655 -- # waitforlisten 121022 /var/tmp/spdk2.sock 00:10:01.233 17:25:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # '[' -z 121022 ']' 00:10:01.233 17:25:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:10:01.233 17:25:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:01.233 17:25:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:10:01.233 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:10:01.233 17:25:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:01.233 17:25:42 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:10:01.233 [2024-12-06 17:25:42.966816] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:10:01.233 [2024-12-06 17:25:42.966906] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid121022 ] 00:10:01.492 [2024-12-06 17:25:43.072059] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 120929 has claimed it. 00:10:01.492 [2024-12-06 17:25:43.072127] app.c: 912:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:10:02.060 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 850: kill: (121022) - No such process 00:10:02.060 ERROR: process (pid: 121022) is no longer running 00:10:02.060 17:25:43 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:02.060 17:25:43 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@868 -- # return 1 00:10:02.060 17:25:43 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@655 -- # es=1 00:10:02.060 17:25:43 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:10:02.060 17:25:43 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:10:02.060 17:25:43 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:10:02.060 17:25:43 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@139 -- # check_remaining_locks 00:10:02.060 17:25:43 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:10:02.060 17:25:43 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:10:02.060 17:25:43 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:10:02.060 17:25:43 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@141 -- # killprocess 120929 00:10:02.060 17:25:43 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@954 -- # '[' -z 120929 ']' 00:10:02.060 17:25:43 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@958 -- # kill -0 120929 00:10:02.060 17:25:43 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@959 -- # uname 00:10:02.060 17:25:43 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:02.060 17:25:43 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 120929 00:10:02.060 17:25:43 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:02.060 17:25:43 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:02.060 17:25:43 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@972 -- # echo 'killing process with pid 120929' 00:10:02.060 killing process with pid 120929 00:10:02.060 17:25:43 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@973 -- # kill 120929 00:10:02.060 17:25:43 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@978 -- # wait 120929 00:10:02.319 00:10:02.319 real 0m1.621s 00:10:02.319 user 0m4.586s 00:10:02.319 sys 0m0.461s 00:10:02.319 17:25:44 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:02.319 17:25:44 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:10:02.319 ************************************ 00:10:02.319 END TEST locking_overlapped_coremask 00:10:02.319 ************************************ 00:10:02.319 17:25:44 event.cpu_locks -- event/cpu_locks.sh@172 -- # run_test locking_overlapped_coremask_via_rpc locking_overlapped_coremask_via_rpc 00:10:02.319 17:25:44 event.cpu_locks -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:10:02.319 17:25:44 event.cpu_locks -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:02.319 17:25:44 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:10:02.578 ************************************ 00:10:02.578 START TEST locking_overlapped_coremask_via_rpc 00:10:02.578 ************************************ 00:10:02.578 17:25:44 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1129 -- # locking_overlapped_coremask_via_rpc 00:10:02.578 17:25:44 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@148 -- # spdk_tgt_pid=121217 00:10:02.578 17:25:44 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@147 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x7 --disable-cpumask-locks 00:10:02.578 17:25:44 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@149 -- # waitforlisten 121217 /var/tmp/spdk.sock 00:10:02.578 17:25:44 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 121217 ']' 00:10:02.578 17:25:44 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:02.578 17:25:44 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:02.578 17:25:44 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:02.578 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:02.578 17:25:44 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:02.578 17:25:44 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:10:02.578 [2024-12-06 17:25:44.216282] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:10:02.578 [2024-12-06 17:25:44.216369] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid121217 ] 00:10:02.578 [2024-12-06 17:25:44.283800] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:10:02.578 [2024-12-06 17:25:44.283837] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:10:02.578 [2024-12-06 17:25:44.333113] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:10:02.578 [2024-12-06 17:25:44.333176] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:10:02.578 [2024-12-06 17:25:44.333180] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:02.836 17:25:44 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:02.836 17:25:44 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:10:02.836 17:25:44 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@152 -- # spdk_tgt_pid2=121234 00:10:02.836 17:25:44 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@153 -- # waitforlisten 121234 /var/tmp/spdk2.sock 00:10:02.836 17:25:44 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@151 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock --disable-cpumask-locks 00:10:02.836 17:25:44 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 121234 ']' 00:10:02.836 17:25:44 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:10:02.836 17:25:44 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:02.836 17:25:44 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:10:02.836 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:10:02.836 17:25:44 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:02.836 17:25:44 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:10:02.836 [2024-12-06 17:25:44.647288] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:10:02.837 [2024-12-06 17:25:44.647372] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid121234 ] 00:10:03.100 [2024-12-06 17:25:44.752551] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:10:03.100 [2024-12-06 17:25:44.752594] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:10:03.100 [2024-12-06 17:25:44.849819] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:10:03.100 [2024-12-06 17:25:44.849878] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:10:03.100 [2024-12-06 17:25:44.849880] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:10:04.034 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:04.034 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:10:04.034 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@155 -- # rpc_cmd framework_enable_cpumask_locks 00:10:04.034 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.034 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:10:04.034 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:04.034 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@156 -- # NOT rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:10:04.034 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@652 -- # local es=0 00:10:04.034 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:10:04.034 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:10:04.034 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:04.034 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:10:04.034 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:04.034 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@655 -- # rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:10:04.034 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:04.034 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:10:04.034 [2024-12-06 17:25:45.636764] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 121217 has claimed it. 00:10:04.034 request: 00:10:04.034 { 00:10:04.034 "method": "framework_enable_cpumask_locks", 00:10:04.035 "req_id": 1 00:10:04.035 } 00:10:04.035 Got JSON-RPC error response 00:10:04.035 response: 00:10:04.035 { 00:10:04.035 "code": -32603, 00:10:04.035 "message": "Failed to claim CPU core: 2" 00:10:04.035 } 00:10:04.035 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:10:04.035 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@655 -- # es=1 00:10:04.035 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:10:04.035 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:10:04.035 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:10:04.035 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@158 -- # waitforlisten 121217 /var/tmp/spdk.sock 00:10:04.035 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 121217 ']' 00:10:04.035 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:04.035 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:04.035 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:04.035 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:04.035 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:04.035 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:10:04.292 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:04.292 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:10:04.292 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@159 -- # waitforlisten 121234 /var/tmp/spdk2.sock 00:10:04.292 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # '[' -z 121234 ']' 00:10:04.292 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk2.sock 00:10:04.292 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:04.292 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:10:04.292 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:10:04.292 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:04.292 17:25:45 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:10:04.552 17:25:46 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:04.552 17:25:46 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@868 -- # return 0 00:10:04.552 17:25:46 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@161 -- # check_remaining_locks 00:10:04.552 17:25:46 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:10:04.552 17:25:46 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:10:04.552 17:25:46 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:10:04.552 00:10:04.552 real 0m2.023s 00:10:04.552 user 0m1.109s 00:10:04.552 sys 0m0.189s 00:10:04.552 17:25:46 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:04.552 17:25:46 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:10:04.552 ************************************ 00:10:04.552 END TEST locking_overlapped_coremask_via_rpc 00:10:04.552 ************************************ 00:10:04.552 17:25:46 event.cpu_locks -- event/cpu_locks.sh@174 -- # cleanup 00:10:04.552 17:25:46 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 121217 ]] 00:10:04.552 17:25:46 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 121217 00:10:04.552 17:25:46 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 121217 ']' 00:10:04.552 17:25:46 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 121217 00:10:04.552 17:25:46 event.cpu_locks -- common/autotest_common.sh@959 -- # uname 00:10:04.552 17:25:46 event.cpu_locks -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:04.552 17:25:46 event.cpu_locks -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 121217 00:10:04.552 17:25:46 event.cpu_locks -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:04.552 17:25:46 event.cpu_locks -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:04.552 17:25:46 event.cpu_locks -- common/autotest_common.sh@972 -- # echo 'killing process with pid 121217' 00:10:04.552 killing process with pid 121217 00:10:04.552 17:25:46 event.cpu_locks -- common/autotest_common.sh@973 -- # kill 121217 00:10:04.552 17:25:46 event.cpu_locks -- common/autotest_common.sh@978 -- # wait 121217 00:10:04.811 17:25:46 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 121234 ]] 00:10:04.811 17:25:46 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 121234 00:10:04.811 17:25:46 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 121234 ']' 00:10:04.811 17:25:46 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 121234 00:10:04.811 17:25:46 event.cpu_locks -- common/autotest_common.sh@959 -- # uname 00:10:04.811 17:25:46 event.cpu_locks -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:04.811 17:25:46 event.cpu_locks -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 121234 00:10:05.072 17:25:46 event.cpu_locks -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:10:05.072 17:25:46 event.cpu_locks -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:10:05.072 17:25:46 event.cpu_locks -- common/autotest_common.sh@972 -- # echo 'killing process with pid 121234' 00:10:05.072 killing process with pid 121234 00:10:05.072 17:25:46 event.cpu_locks -- common/autotest_common.sh@973 -- # kill 121234 00:10:05.072 17:25:46 event.cpu_locks -- common/autotest_common.sh@978 -- # wait 121234 00:10:05.332 17:25:47 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:10:05.332 17:25:47 event.cpu_locks -- event/cpu_locks.sh@1 -- # cleanup 00:10:05.332 17:25:47 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 121217 ]] 00:10:05.332 17:25:47 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 121217 00:10:05.332 17:25:47 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 121217 ']' 00:10:05.332 17:25:47 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 121217 00:10:05.332 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 958: kill: (121217) - No such process 00:10:05.332 17:25:47 event.cpu_locks -- common/autotest_common.sh@981 -- # echo 'Process with pid 121217 is not found' 00:10:05.332 Process with pid 121217 is not found 00:10:05.332 17:25:47 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 121234 ]] 00:10:05.332 17:25:47 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 121234 00:10:05.332 17:25:47 event.cpu_locks -- common/autotest_common.sh@954 -- # '[' -z 121234 ']' 00:10:05.332 17:25:47 event.cpu_locks -- common/autotest_common.sh@958 -- # kill -0 121234 00:10:05.332 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 958: kill: (121234) - No such process 00:10:05.332 17:25:47 event.cpu_locks -- common/autotest_common.sh@981 -- # echo 'Process with pid 121234 is not found' 00:10:05.332 Process with pid 121234 is not found 00:10:05.332 17:25:47 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:10:05.332 00:10:05.332 real 0m14.855s 00:10:05.332 user 0m27.129s 00:10:05.332 sys 0m5.229s 00:10:05.332 17:25:47 event.cpu_locks -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:05.332 17:25:47 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:10:05.332 ************************************ 00:10:05.332 END TEST cpu_locks 00:10:05.332 ************************************ 00:10:05.332 00:10:05.332 real 0m39.400s 00:10:05.332 user 1m17.659s 00:10:05.332 sys 0m9.328s 00:10:05.332 17:25:47 event -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:05.332 17:25:47 event -- common/autotest_common.sh@10 -- # set +x 00:10:05.332 ************************************ 00:10:05.332 END TEST event 00:10:05.332 ************************************ 00:10:05.332 17:25:47 -- spdk/autotest.sh@169 -- # run_test thread /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/thread.sh 00:10:05.332 17:25:47 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:10:05.332 17:25:47 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:05.332 17:25:47 -- common/autotest_common.sh@10 -- # set +x 00:10:05.332 ************************************ 00:10:05.332 START TEST thread 00:10:05.332 ************************************ 00:10:05.332 17:25:47 thread -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/thread.sh 00:10:05.591 * Looking for test storage... 00:10:05.591 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread 00:10:05.591 17:25:47 thread -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:10:05.591 17:25:47 thread -- common/autotest_common.sh@1711 -- # lcov --version 00:10:05.591 17:25:47 thread -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:10:05.591 17:25:47 thread -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:10:05.591 17:25:47 thread -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:10:05.591 17:25:47 thread -- scripts/common.sh@333 -- # local ver1 ver1_l 00:10:05.591 17:25:47 thread -- scripts/common.sh@334 -- # local ver2 ver2_l 00:10:05.591 17:25:47 thread -- scripts/common.sh@336 -- # IFS=.-: 00:10:05.591 17:25:47 thread -- scripts/common.sh@336 -- # read -ra ver1 00:10:05.591 17:25:47 thread -- scripts/common.sh@337 -- # IFS=.-: 00:10:05.591 17:25:47 thread -- scripts/common.sh@337 -- # read -ra ver2 00:10:05.591 17:25:47 thread -- scripts/common.sh@338 -- # local 'op=<' 00:10:05.591 17:25:47 thread -- scripts/common.sh@340 -- # ver1_l=2 00:10:05.591 17:25:47 thread -- scripts/common.sh@341 -- # ver2_l=1 00:10:05.591 17:25:47 thread -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:10:05.591 17:25:47 thread -- scripts/common.sh@344 -- # case "$op" in 00:10:05.591 17:25:47 thread -- scripts/common.sh@345 -- # : 1 00:10:05.591 17:25:47 thread -- scripts/common.sh@364 -- # (( v = 0 )) 00:10:05.591 17:25:47 thread -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:10:05.591 17:25:47 thread -- scripts/common.sh@365 -- # decimal 1 00:10:05.591 17:25:47 thread -- scripts/common.sh@353 -- # local d=1 00:10:05.591 17:25:47 thread -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:10:05.591 17:25:47 thread -- scripts/common.sh@355 -- # echo 1 00:10:05.591 17:25:47 thread -- scripts/common.sh@365 -- # ver1[v]=1 00:10:05.591 17:25:47 thread -- scripts/common.sh@366 -- # decimal 2 00:10:05.591 17:25:47 thread -- scripts/common.sh@353 -- # local d=2 00:10:05.591 17:25:47 thread -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:10:05.591 17:25:47 thread -- scripts/common.sh@355 -- # echo 2 00:10:05.591 17:25:47 thread -- scripts/common.sh@366 -- # ver2[v]=2 00:10:05.591 17:25:47 thread -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:10:05.591 17:25:47 thread -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:10:05.591 17:25:47 thread -- scripts/common.sh@368 -- # return 0 00:10:05.591 17:25:47 thread -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:10:05.591 17:25:47 thread -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:10:05.591 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:05.591 --rc genhtml_branch_coverage=1 00:10:05.591 --rc genhtml_function_coverage=1 00:10:05.591 --rc genhtml_legend=1 00:10:05.591 --rc geninfo_all_blocks=1 00:10:05.591 --rc geninfo_unexecuted_blocks=1 00:10:05.591 00:10:05.591 ' 00:10:05.591 17:25:47 thread -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:10:05.591 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:05.591 --rc genhtml_branch_coverage=1 00:10:05.591 --rc genhtml_function_coverage=1 00:10:05.591 --rc genhtml_legend=1 00:10:05.591 --rc geninfo_all_blocks=1 00:10:05.591 --rc geninfo_unexecuted_blocks=1 00:10:05.591 00:10:05.591 ' 00:10:05.591 17:25:47 thread -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:10:05.591 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:05.591 --rc genhtml_branch_coverage=1 00:10:05.591 --rc genhtml_function_coverage=1 00:10:05.591 --rc genhtml_legend=1 00:10:05.591 --rc geninfo_all_blocks=1 00:10:05.591 --rc geninfo_unexecuted_blocks=1 00:10:05.591 00:10:05.591 ' 00:10:05.591 17:25:47 thread -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:10:05.591 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:05.591 --rc genhtml_branch_coverage=1 00:10:05.591 --rc genhtml_function_coverage=1 00:10:05.591 --rc genhtml_legend=1 00:10:05.591 --rc geninfo_all_blocks=1 00:10:05.591 --rc geninfo_unexecuted_blocks=1 00:10:05.591 00:10:05.591 ' 00:10:05.591 17:25:47 thread -- thread/thread.sh@11 -- # run_test thread_poller_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:10:05.591 17:25:47 thread -- common/autotest_common.sh@1105 -- # '[' 8 -le 1 ']' 00:10:05.591 17:25:47 thread -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:05.591 17:25:47 thread -- common/autotest_common.sh@10 -- # set +x 00:10:05.591 ************************************ 00:10:05.591 START TEST thread_poller_perf 00:10:05.591 ************************************ 00:10:05.591 17:25:47 thread.thread_poller_perf -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:10:05.591 [2024-12-06 17:25:47.332321] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:10:05.591 [2024-12-06 17:25:47.332389] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid121611 ] 00:10:05.591 [2024-12-06 17:25:47.401539] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:05.849 [2024-12-06 17:25:47.449407] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:05.849 Running 1000 pollers for 1 seconds with 1 microseconds period. 00:10:06.782 [2024-12-06T16:25:48.621Z] ====================================== 00:10:06.782 [2024-12-06T16:25:48.621Z] busy:2709543762 (cyc) 00:10:06.782 [2024-12-06T16:25:48.621Z] total_run_count: 368000 00:10:06.782 [2024-12-06T16:25:48.621Z] tsc_hz: 2700000000 (cyc) 00:10:06.782 [2024-12-06T16:25:48.621Z] ====================================== 00:10:06.782 [2024-12-06T16:25:48.621Z] poller_cost: 7362 (cyc), 2726 (nsec) 00:10:06.782 00:10:06.782 real 0m1.181s 00:10:06.782 user 0m1.106s 00:10:06.782 sys 0m0.071s 00:10:06.782 17:25:48 thread.thread_poller_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:06.782 17:25:48 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:10:06.782 ************************************ 00:10:06.782 END TEST thread_poller_perf 00:10:06.782 ************************************ 00:10:06.782 17:25:48 thread -- thread/thread.sh@12 -- # run_test thread_poller_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:10:06.782 17:25:48 thread -- common/autotest_common.sh@1105 -- # '[' 8 -le 1 ']' 00:10:06.782 17:25:48 thread -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:06.782 17:25:48 thread -- common/autotest_common.sh@10 -- # set +x 00:10:06.782 ************************************ 00:10:06.782 START TEST thread_poller_perf 00:10:06.782 ************************************ 00:10:06.782 17:25:48 thread.thread_poller_perf -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:10:06.782 [2024-12-06 17:25:48.566587] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:10:06.782 [2024-12-06 17:25:48.566653] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid121832 ] 00:10:07.041 [2024-12-06 17:25:48.634019] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:07.041 [2024-12-06 17:25:48.678823] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:07.041 Running 1000 pollers for 1 seconds with 0 microseconds period. 00:10:07.980 [2024-12-06T16:25:49.819Z] ====================================== 00:10:07.980 [2024-12-06T16:25:49.819Z] busy:2702087178 (cyc) 00:10:07.980 [2024-12-06T16:25:49.819Z] total_run_count: 4487000 00:10:07.980 [2024-12-06T16:25:49.819Z] tsc_hz: 2700000000 (cyc) 00:10:07.980 [2024-12-06T16:25:49.819Z] ====================================== 00:10:07.980 [2024-12-06T16:25:49.819Z] poller_cost: 602 (cyc), 222 (nsec) 00:10:07.980 00:10:07.980 real 0m1.169s 00:10:07.980 user 0m1.087s 00:10:07.980 sys 0m0.077s 00:10:07.980 17:25:49 thread.thread_poller_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:07.980 17:25:49 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:10:07.980 ************************************ 00:10:07.980 END TEST thread_poller_perf 00:10:07.980 ************************************ 00:10:07.980 17:25:49 thread -- thread/thread.sh@17 -- # [[ y != \y ]] 00:10:07.980 00:10:07.980 real 0m2.594s 00:10:07.980 user 0m2.326s 00:10:07.980 sys 0m0.275s 00:10:07.980 17:25:49 thread -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:07.980 17:25:49 thread -- common/autotest_common.sh@10 -- # set +x 00:10:07.980 ************************************ 00:10:07.980 END TEST thread 00:10:07.980 ************************************ 00:10:07.980 17:25:49 -- spdk/autotest.sh@171 -- # [[ 0 -eq 1 ]] 00:10:07.980 17:25:49 -- spdk/autotest.sh@176 -- # run_test app_cmdline /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/cmdline.sh 00:10:07.980 17:25:49 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:10:07.980 17:25:49 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:07.980 17:25:49 -- common/autotest_common.sh@10 -- # set +x 00:10:07.980 ************************************ 00:10:07.980 START TEST app_cmdline 00:10:07.980 ************************************ 00:10:07.980 17:25:49 app_cmdline -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/cmdline.sh 00:10:08.240 * Looking for test storage... 00:10:08.240 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app 00:10:08.240 17:25:49 app_cmdline -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:10:08.240 17:25:49 app_cmdline -- common/autotest_common.sh@1711 -- # lcov --version 00:10:08.240 17:25:49 app_cmdline -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:10:08.240 17:25:49 app_cmdline -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:10:08.241 17:25:49 app_cmdline -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:10:08.241 17:25:49 app_cmdline -- scripts/common.sh@333 -- # local ver1 ver1_l 00:10:08.241 17:25:49 app_cmdline -- scripts/common.sh@334 -- # local ver2 ver2_l 00:10:08.241 17:25:49 app_cmdline -- scripts/common.sh@336 -- # IFS=.-: 00:10:08.241 17:25:49 app_cmdline -- scripts/common.sh@336 -- # read -ra ver1 00:10:08.241 17:25:49 app_cmdline -- scripts/common.sh@337 -- # IFS=.-: 00:10:08.241 17:25:49 app_cmdline -- scripts/common.sh@337 -- # read -ra ver2 00:10:08.241 17:25:49 app_cmdline -- scripts/common.sh@338 -- # local 'op=<' 00:10:08.241 17:25:49 app_cmdline -- scripts/common.sh@340 -- # ver1_l=2 00:10:08.241 17:25:49 app_cmdline -- scripts/common.sh@341 -- # ver2_l=1 00:10:08.241 17:25:49 app_cmdline -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:10:08.241 17:25:49 app_cmdline -- scripts/common.sh@344 -- # case "$op" in 00:10:08.241 17:25:49 app_cmdline -- scripts/common.sh@345 -- # : 1 00:10:08.241 17:25:49 app_cmdline -- scripts/common.sh@364 -- # (( v = 0 )) 00:10:08.241 17:25:49 app_cmdline -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:10:08.241 17:25:49 app_cmdline -- scripts/common.sh@365 -- # decimal 1 00:10:08.241 17:25:49 app_cmdline -- scripts/common.sh@353 -- # local d=1 00:10:08.241 17:25:49 app_cmdline -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:10:08.241 17:25:49 app_cmdline -- scripts/common.sh@355 -- # echo 1 00:10:08.241 17:25:49 app_cmdline -- scripts/common.sh@365 -- # ver1[v]=1 00:10:08.241 17:25:49 app_cmdline -- scripts/common.sh@366 -- # decimal 2 00:10:08.241 17:25:49 app_cmdline -- scripts/common.sh@353 -- # local d=2 00:10:08.241 17:25:49 app_cmdline -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:10:08.241 17:25:49 app_cmdline -- scripts/common.sh@355 -- # echo 2 00:10:08.241 17:25:49 app_cmdline -- scripts/common.sh@366 -- # ver2[v]=2 00:10:08.241 17:25:49 app_cmdline -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:10:08.241 17:25:49 app_cmdline -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:10:08.241 17:25:49 app_cmdline -- scripts/common.sh@368 -- # return 0 00:10:08.241 17:25:49 app_cmdline -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:10:08.241 17:25:49 app_cmdline -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:10:08.241 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:08.241 --rc genhtml_branch_coverage=1 00:10:08.241 --rc genhtml_function_coverage=1 00:10:08.241 --rc genhtml_legend=1 00:10:08.241 --rc geninfo_all_blocks=1 00:10:08.241 --rc geninfo_unexecuted_blocks=1 00:10:08.241 00:10:08.241 ' 00:10:08.241 17:25:49 app_cmdline -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:10:08.241 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:08.241 --rc genhtml_branch_coverage=1 00:10:08.241 --rc genhtml_function_coverage=1 00:10:08.241 --rc genhtml_legend=1 00:10:08.241 --rc geninfo_all_blocks=1 00:10:08.241 --rc geninfo_unexecuted_blocks=1 00:10:08.241 00:10:08.241 ' 00:10:08.241 17:25:49 app_cmdline -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:10:08.241 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:08.241 --rc genhtml_branch_coverage=1 00:10:08.241 --rc genhtml_function_coverage=1 00:10:08.241 --rc genhtml_legend=1 00:10:08.241 --rc geninfo_all_blocks=1 00:10:08.241 --rc geninfo_unexecuted_blocks=1 00:10:08.241 00:10:08.241 ' 00:10:08.241 17:25:49 app_cmdline -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:10:08.241 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:08.241 --rc genhtml_branch_coverage=1 00:10:08.241 --rc genhtml_function_coverage=1 00:10:08.241 --rc genhtml_legend=1 00:10:08.241 --rc geninfo_all_blocks=1 00:10:08.241 --rc geninfo_unexecuted_blocks=1 00:10:08.241 00:10:08.241 ' 00:10:08.241 17:25:49 app_cmdline -- app/cmdline.sh@14 -- # trap 'killprocess $spdk_tgt_pid' EXIT 00:10:08.241 17:25:49 app_cmdline -- app/cmdline.sh@17 -- # spdk_tgt_pid=122087 00:10:08.241 17:25:49 app_cmdline -- app/cmdline.sh@16 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --rpcs-allowed spdk_get_version,rpc_get_methods 00:10:08.241 17:25:49 app_cmdline -- app/cmdline.sh@18 -- # waitforlisten 122087 00:10:08.241 17:25:49 app_cmdline -- common/autotest_common.sh@835 -- # '[' -z 122087 ']' 00:10:08.241 17:25:49 app_cmdline -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:08.241 17:25:49 app_cmdline -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:08.241 17:25:49 app_cmdline -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:08.241 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:08.241 17:25:49 app_cmdline -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:08.241 17:25:49 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:10:08.241 [2024-12-06 17:25:49.999096] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:10:08.241 [2024-12-06 17:25:49.999177] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid122087 ] 00:10:08.241 [2024-12-06 17:25:50.072285] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:08.501 [2024-12-06 17:25:50.122245] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:08.764 17:25:50 app_cmdline -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:08.764 17:25:50 app_cmdline -- common/autotest_common.sh@868 -- # return 0 00:10:08.764 17:25:50 app_cmdline -- app/cmdline.sh@20 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py spdk_get_version 00:10:09.023 { 00:10:09.023 "version": "SPDK v25.01-pre git sha1 a5e6ecf28", 00:10:09.023 "fields": { 00:10:09.023 "major": 25, 00:10:09.023 "minor": 1, 00:10:09.023 "patch": 0, 00:10:09.023 "suffix": "-pre", 00:10:09.023 "commit": "a5e6ecf28" 00:10:09.023 } 00:10:09.023 } 00:10:09.023 17:25:50 app_cmdline -- app/cmdline.sh@22 -- # expected_methods=() 00:10:09.023 17:25:50 app_cmdline -- app/cmdline.sh@23 -- # expected_methods+=("rpc_get_methods") 00:10:09.023 17:25:50 app_cmdline -- app/cmdline.sh@24 -- # expected_methods+=("spdk_get_version") 00:10:09.023 17:25:50 app_cmdline -- app/cmdline.sh@26 -- # methods=($(rpc_cmd rpc_get_methods | jq -r ".[]" | sort)) 00:10:09.023 17:25:50 app_cmdline -- app/cmdline.sh@26 -- # rpc_cmd rpc_get_methods 00:10:09.023 17:25:50 app_cmdline -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:09.023 17:25:50 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:10:09.023 17:25:50 app_cmdline -- app/cmdline.sh@26 -- # jq -r '.[]' 00:10:09.023 17:25:50 app_cmdline -- app/cmdline.sh@26 -- # sort 00:10:09.023 17:25:50 app_cmdline -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:09.023 17:25:50 app_cmdline -- app/cmdline.sh@27 -- # (( 2 == 2 )) 00:10:09.023 17:25:50 app_cmdline -- app/cmdline.sh@28 -- # [[ rpc_get_methods spdk_get_version == \r\p\c\_\g\e\t\_\m\e\t\h\o\d\s\ \s\p\d\k\_\g\e\t\_\v\e\r\s\i\o\n ]] 00:10:09.023 17:25:50 app_cmdline -- app/cmdline.sh@30 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:10:09.023 17:25:50 app_cmdline -- common/autotest_common.sh@652 -- # local es=0 00:10:09.023 17:25:50 app_cmdline -- common/autotest_common.sh@654 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:10:09.023 17:25:50 app_cmdline -- common/autotest_common.sh@640 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:10:09.023 17:25:50 app_cmdline -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:09.023 17:25:50 app_cmdline -- common/autotest_common.sh@644 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:10:09.023 17:25:50 app_cmdline -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:09.023 17:25:50 app_cmdline -- common/autotest_common.sh@646 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:10:09.023 17:25:50 app_cmdline -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:10:09.023 17:25:50 app_cmdline -- common/autotest_common.sh@646 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:10:09.023 17:25:50 app_cmdline -- common/autotest_common.sh@646 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:10:09.023 17:25:50 app_cmdline -- common/autotest_common.sh@655 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:10:09.282 request: 00:10:09.282 { 00:10:09.282 "method": "env_dpdk_get_mem_stats", 00:10:09.282 "req_id": 1 00:10:09.282 } 00:10:09.282 Got JSON-RPC error response 00:10:09.282 response: 00:10:09.282 { 00:10:09.282 "code": -32601, 00:10:09.282 "message": "Method not found" 00:10:09.282 } 00:10:09.282 17:25:51 app_cmdline -- common/autotest_common.sh@655 -- # es=1 00:10:09.282 17:25:51 app_cmdline -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:10:09.282 17:25:51 app_cmdline -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:10:09.282 17:25:51 app_cmdline -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:10:09.282 17:25:51 app_cmdline -- app/cmdline.sh@1 -- # killprocess 122087 00:10:09.282 17:25:51 app_cmdline -- common/autotest_common.sh@954 -- # '[' -z 122087 ']' 00:10:09.282 17:25:51 app_cmdline -- common/autotest_common.sh@958 -- # kill -0 122087 00:10:09.282 17:25:51 app_cmdline -- common/autotest_common.sh@959 -- # uname 00:10:09.282 17:25:51 app_cmdline -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:09.282 17:25:51 app_cmdline -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 122087 00:10:09.282 17:25:51 app_cmdline -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:10:09.282 17:25:51 app_cmdline -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:10:09.282 17:25:51 app_cmdline -- common/autotest_common.sh@972 -- # echo 'killing process with pid 122087' 00:10:09.282 killing process with pid 122087 00:10:09.282 17:25:51 app_cmdline -- common/autotest_common.sh@973 -- # kill 122087 00:10:09.282 17:25:51 app_cmdline -- common/autotest_common.sh@978 -- # wait 122087 00:10:09.851 00:10:09.851 real 0m1.621s 00:10:09.851 user 0m2.005s 00:10:09.851 sys 0m0.528s 00:10:09.851 17:25:51 app_cmdline -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:09.851 17:25:51 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:10:09.851 ************************************ 00:10:09.851 END TEST app_cmdline 00:10:09.851 ************************************ 00:10:09.851 17:25:51 -- spdk/autotest.sh@177 -- # run_test version /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/version.sh 00:10:09.851 17:25:51 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:10:09.851 17:25:51 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:09.851 17:25:51 -- common/autotest_common.sh@10 -- # set +x 00:10:09.851 ************************************ 00:10:09.851 START TEST version 00:10:09.851 ************************************ 00:10:09.851 17:25:51 version -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/version.sh 00:10:09.851 * Looking for test storage... 00:10:09.851 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app 00:10:09.851 17:25:51 version -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:10:09.851 17:25:51 version -- common/autotest_common.sh@1711 -- # lcov --version 00:10:09.851 17:25:51 version -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:10:09.851 17:25:51 version -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:10:09.851 17:25:51 version -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:10:09.851 17:25:51 version -- scripts/common.sh@333 -- # local ver1 ver1_l 00:10:09.851 17:25:51 version -- scripts/common.sh@334 -- # local ver2 ver2_l 00:10:09.851 17:25:51 version -- scripts/common.sh@336 -- # IFS=.-: 00:10:09.851 17:25:51 version -- scripts/common.sh@336 -- # read -ra ver1 00:10:09.851 17:25:51 version -- scripts/common.sh@337 -- # IFS=.-: 00:10:09.851 17:25:51 version -- scripts/common.sh@337 -- # read -ra ver2 00:10:09.851 17:25:51 version -- scripts/common.sh@338 -- # local 'op=<' 00:10:09.851 17:25:51 version -- scripts/common.sh@340 -- # ver1_l=2 00:10:09.851 17:25:51 version -- scripts/common.sh@341 -- # ver2_l=1 00:10:09.851 17:25:51 version -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:10:09.851 17:25:51 version -- scripts/common.sh@344 -- # case "$op" in 00:10:09.851 17:25:51 version -- scripts/common.sh@345 -- # : 1 00:10:09.851 17:25:51 version -- scripts/common.sh@364 -- # (( v = 0 )) 00:10:09.851 17:25:51 version -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:10:09.851 17:25:51 version -- scripts/common.sh@365 -- # decimal 1 00:10:09.851 17:25:51 version -- scripts/common.sh@353 -- # local d=1 00:10:09.851 17:25:51 version -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:10:09.851 17:25:51 version -- scripts/common.sh@355 -- # echo 1 00:10:09.851 17:25:51 version -- scripts/common.sh@365 -- # ver1[v]=1 00:10:09.851 17:25:51 version -- scripts/common.sh@366 -- # decimal 2 00:10:09.851 17:25:51 version -- scripts/common.sh@353 -- # local d=2 00:10:09.851 17:25:51 version -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:10:09.851 17:25:51 version -- scripts/common.sh@355 -- # echo 2 00:10:09.851 17:25:51 version -- scripts/common.sh@366 -- # ver2[v]=2 00:10:09.851 17:25:51 version -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:10:09.851 17:25:51 version -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:10:09.851 17:25:51 version -- scripts/common.sh@368 -- # return 0 00:10:09.851 17:25:51 version -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:10:09.851 17:25:51 version -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:10:09.851 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:09.851 --rc genhtml_branch_coverage=1 00:10:09.851 --rc genhtml_function_coverage=1 00:10:09.851 --rc genhtml_legend=1 00:10:09.851 --rc geninfo_all_blocks=1 00:10:09.851 --rc geninfo_unexecuted_blocks=1 00:10:09.851 00:10:09.851 ' 00:10:09.851 17:25:51 version -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:10:09.851 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:09.851 --rc genhtml_branch_coverage=1 00:10:09.851 --rc genhtml_function_coverage=1 00:10:09.851 --rc genhtml_legend=1 00:10:09.851 --rc geninfo_all_blocks=1 00:10:09.851 --rc geninfo_unexecuted_blocks=1 00:10:09.851 00:10:09.851 ' 00:10:09.851 17:25:51 version -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:10:09.851 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:09.851 --rc genhtml_branch_coverage=1 00:10:09.851 --rc genhtml_function_coverage=1 00:10:09.851 --rc genhtml_legend=1 00:10:09.851 --rc geninfo_all_blocks=1 00:10:09.851 --rc geninfo_unexecuted_blocks=1 00:10:09.851 00:10:09.851 ' 00:10:09.851 17:25:51 version -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:10:09.851 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:09.851 --rc genhtml_branch_coverage=1 00:10:09.851 --rc genhtml_function_coverage=1 00:10:09.851 --rc genhtml_legend=1 00:10:09.851 --rc geninfo_all_blocks=1 00:10:09.851 --rc geninfo_unexecuted_blocks=1 00:10:09.851 00:10:09.851 ' 00:10:09.851 17:25:51 version -- app/version.sh@17 -- # get_header_version major 00:10:09.851 17:25:51 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MAJOR[[:space:]]+' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/version.h 00:10:09.851 17:25:51 version -- app/version.sh@14 -- # cut -f2 00:10:09.851 17:25:51 version -- app/version.sh@14 -- # tr -d '"' 00:10:09.851 17:25:51 version -- app/version.sh@17 -- # major=25 00:10:09.851 17:25:51 version -- app/version.sh@18 -- # get_header_version minor 00:10:09.851 17:25:51 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MINOR[[:space:]]+' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/version.h 00:10:09.851 17:25:51 version -- app/version.sh@14 -- # cut -f2 00:10:09.851 17:25:51 version -- app/version.sh@14 -- # tr -d '"' 00:10:09.851 17:25:51 version -- app/version.sh@18 -- # minor=1 00:10:09.851 17:25:51 version -- app/version.sh@19 -- # get_header_version patch 00:10:09.851 17:25:51 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_PATCH[[:space:]]+' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/version.h 00:10:09.851 17:25:51 version -- app/version.sh@14 -- # cut -f2 00:10:09.851 17:25:51 version -- app/version.sh@14 -- # tr -d '"' 00:10:09.851 17:25:51 version -- app/version.sh@19 -- # patch=0 00:10:09.851 17:25:51 version -- app/version.sh@20 -- # get_header_version suffix 00:10:09.851 17:25:51 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_SUFFIX[[:space:]]+' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/version.h 00:10:09.851 17:25:51 version -- app/version.sh@14 -- # cut -f2 00:10:09.851 17:25:51 version -- app/version.sh@14 -- # tr -d '"' 00:10:09.851 17:25:51 version -- app/version.sh@20 -- # suffix=-pre 00:10:09.851 17:25:51 version -- app/version.sh@22 -- # version=25.1 00:10:09.851 17:25:51 version -- app/version.sh@25 -- # (( patch != 0 )) 00:10:09.851 17:25:51 version -- app/version.sh@28 -- # version=25.1rc0 00:10:09.851 17:25:51 version -- app/version.sh@30 -- # PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python 00:10:09.851 17:25:51 version -- app/version.sh@30 -- # python3 -c 'import spdk; print(spdk.__version__)' 00:10:09.851 17:25:51 version -- app/version.sh@30 -- # py_version=25.1rc0 00:10:09.851 17:25:51 version -- app/version.sh@31 -- # [[ 25.1rc0 == \2\5\.\1\r\c\0 ]] 00:10:09.851 00:10:09.851 real 0m0.209s 00:10:09.851 user 0m0.140s 00:10:09.851 sys 0m0.094s 00:10:09.851 17:25:51 version -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:09.851 17:25:51 version -- common/autotest_common.sh@10 -- # set +x 00:10:09.851 ************************************ 00:10:09.851 END TEST version 00:10:09.851 ************************************ 00:10:10.111 17:25:51 -- spdk/autotest.sh@179 -- # '[' 0 -eq 1 ']' 00:10:10.111 17:25:51 -- spdk/autotest.sh@188 -- # [[ 0 -eq 1 ]] 00:10:10.111 17:25:51 -- spdk/autotest.sh@194 -- # uname -s 00:10:10.111 17:25:51 -- spdk/autotest.sh@194 -- # [[ Linux == Linux ]] 00:10:10.111 17:25:51 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:10:10.111 17:25:51 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:10:10.111 17:25:51 -- spdk/autotest.sh@207 -- # '[' 0 -eq 1 ']' 00:10:10.111 17:25:51 -- spdk/autotest.sh@256 -- # '[' 0 -eq 1 ']' 00:10:10.111 17:25:51 -- spdk/autotest.sh@260 -- # timing_exit lib 00:10:10.111 17:25:51 -- common/autotest_common.sh@732 -- # xtrace_disable 00:10:10.111 17:25:51 -- common/autotest_common.sh@10 -- # set +x 00:10:10.111 17:25:51 -- spdk/autotest.sh@262 -- # '[' 0 -eq 1 ']' 00:10:10.111 17:25:51 -- spdk/autotest.sh@267 -- # '[' 0 -eq 1 ']' 00:10:10.111 17:25:51 -- spdk/autotest.sh@276 -- # '[' 1 -eq 1 ']' 00:10:10.111 17:25:51 -- spdk/autotest.sh@277 -- # export NET_TYPE 00:10:10.111 17:25:51 -- spdk/autotest.sh@280 -- # '[' tcp = rdma ']' 00:10:10.111 17:25:51 -- spdk/autotest.sh@283 -- # '[' tcp = tcp ']' 00:10:10.111 17:25:51 -- spdk/autotest.sh@284 -- # run_test nvmf_tcp /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf.sh --transport=tcp 00:10:10.111 17:25:51 -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:10:10.111 17:25:51 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:10.111 17:25:51 -- common/autotest_common.sh@10 -- # set +x 00:10:10.111 ************************************ 00:10:10.111 START TEST nvmf_tcp 00:10:10.111 ************************************ 00:10:10.111 17:25:51 nvmf_tcp -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf.sh --transport=tcp 00:10:10.111 * Looking for test storage... 00:10:10.111 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf 00:10:10.111 17:25:51 nvmf_tcp -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:10:10.111 17:25:51 nvmf_tcp -- common/autotest_common.sh@1711 -- # lcov --version 00:10:10.111 17:25:51 nvmf_tcp -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:10:10.111 17:25:51 nvmf_tcp -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:10:10.111 17:25:51 nvmf_tcp -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:10:10.111 17:25:51 nvmf_tcp -- scripts/common.sh@333 -- # local ver1 ver1_l 00:10:10.111 17:25:51 nvmf_tcp -- scripts/common.sh@334 -- # local ver2 ver2_l 00:10:10.111 17:25:51 nvmf_tcp -- scripts/common.sh@336 -- # IFS=.-: 00:10:10.111 17:25:51 nvmf_tcp -- scripts/common.sh@336 -- # read -ra ver1 00:10:10.111 17:25:51 nvmf_tcp -- scripts/common.sh@337 -- # IFS=.-: 00:10:10.111 17:25:51 nvmf_tcp -- scripts/common.sh@337 -- # read -ra ver2 00:10:10.111 17:25:51 nvmf_tcp -- scripts/common.sh@338 -- # local 'op=<' 00:10:10.111 17:25:51 nvmf_tcp -- scripts/common.sh@340 -- # ver1_l=2 00:10:10.111 17:25:51 nvmf_tcp -- scripts/common.sh@341 -- # ver2_l=1 00:10:10.111 17:25:51 nvmf_tcp -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:10:10.111 17:25:51 nvmf_tcp -- scripts/common.sh@344 -- # case "$op" in 00:10:10.111 17:25:51 nvmf_tcp -- scripts/common.sh@345 -- # : 1 00:10:10.111 17:25:51 nvmf_tcp -- scripts/common.sh@364 -- # (( v = 0 )) 00:10:10.111 17:25:51 nvmf_tcp -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:10:10.111 17:25:51 nvmf_tcp -- scripts/common.sh@365 -- # decimal 1 00:10:10.111 17:25:51 nvmf_tcp -- scripts/common.sh@353 -- # local d=1 00:10:10.111 17:25:51 nvmf_tcp -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:10:10.111 17:25:51 nvmf_tcp -- scripts/common.sh@355 -- # echo 1 00:10:10.111 17:25:51 nvmf_tcp -- scripts/common.sh@365 -- # ver1[v]=1 00:10:10.111 17:25:51 nvmf_tcp -- scripts/common.sh@366 -- # decimal 2 00:10:10.111 17:25:51 nvmf_tcp -- scripts/common.sh@353 -- # local d=2 00:10:10.111 17:25:51 nvmf_tcp -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:10:10.111 17:25:51 nvmf_tcp -- scripts/common.sh@355 -- # echo 2 00:10:10.111 17:25:51 nvmf_tcp -- scripts/common.sh@366 -- # ver2[v]=2 00:10:10.111 17:25:51 nvmf_tcp -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:10:10.111 17:25:51 nvmf_tcp -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:10:10.111 17:25:51 nvmf_tcp -- scripts/common.sh@368 -- # return 0 00:10:10.111 17:25:51 nvmf_tcp -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:10:10.111 17:25:51 nvmf_tcp -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:10:10.111 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:10.111 --rc genhtml_branch_coverage=1 00:10:10.111 --rc genhtml_function_coverage=1 00:10:10.111 --rc genhtml_legend=1 00:10:10.111 --rc geninfo_all_blocks=1 00:10:10.111 --rc geninfo_unexecuted_blocks=1 00:10:10.111 00:10:10.111 ' 00:10:10.111 17:25:51 nvmf_tcp -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:10:10.111 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:10.111 --rc genhtml_branch_coverage=1 00:10:10.111 --rc genhtml_function_coverage=1 00:10:10.111 --rc genhtml_legend=1 00:10:10.111 --rc geninfo_all_blocks=1 00:10:10.111 --rc geninfo_unexecuted_blocks=1 00:10:10.111 00:10:10.111 ' 00:10:10.111 17:25:51 nvmf_tcp -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:10:10.111 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:10.111 --rc genhtml_branch_coverage=1 00:10:10.111 --rc genhtml_function_coverage=1 00:10:10.111 --rc genhtml_legend=1 00:10:10.111 --rc geninfo_all_blocks=1 00:10:10.111 --rc geninfo_unexecuted_blocks=1 00:10:10.111 00:10:10.111 ' 00:10:10.111 17:25:51 nvmf_tcp -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:10:10.111 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:10.111 --rc genhtml_branch_coverage=1 00:10:10.111 --rc genhtml_function_coverage=1 00:10:10.111 --rc genhtml_legend=1 00:10:10.111 --rc geninfo_all_blocks=1 00:10:10.111 --rc geninfo_unexecuted_blocks=1 00:10:10.111 00:10:10.111 ' 00:10:10.111 17:25:51 nvmf_tcp -- nvmf/nvmf.sh@10 -- # uname -s 00:10:10.111 17:25:51 nvmf_tcp -- nvmf/nvmf.sh@10 -- # '[' '!' Linux = Linux ']' 00:10:10.111 17:25:51 nvmf_tcp -- nvmf/nvmf.sh@14 -- # run_test nvmf_target_core /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_target_core.sh --transport=tcp 00:10:10.111 17:25:51 nvmf_tcp -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:10:10.111 17:25:51 nvmf_tcp -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:10.111 17:25:51 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:10:10.111 ************************************ 00:10:10.111 START TEST nvmf_target_core 00:10:10.111 ************************************ 00:10:10.112 17:25:51 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_target_core.sh --transport=tcp 00:10:10.371 * Looking for test storage... 00:10:10.371 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf 00:10:10.371 17:25:51 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:10:10.371 17:25:51 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1711 -- # lcov --version 00:10:10.371 17:25:51 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@333 -- # local ver1 ver1_l 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@334 -- # local ver2 ver2_l 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@336 -- # IFS=.-: 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@336 -- # read -ra ver1 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@337 -- # IFS=.-: 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@337 -- # read -ra ver2 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@338 -- # local 'op=<' 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@340 -- # ver1_l=2 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@341 -- # ver2_l=1 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@344 -- # case "$op" in 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@345 -- # : 1 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@364 -- # (( v = 0 )) 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@365 -- # decimal 1 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@353 -- # local d=1 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@355 -- # echo 1 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@365 -- # ver1[v]=1 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@366 -- # decimal 2 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@353 -- # local d=2 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@355 -- # echo 2 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@366 -- # ver2[v]=2 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@368 -- # return 0 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:10:10.371 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:10.371 --rc genhtml_branch_coverage=1 00:10:10.371 --rc genhtml_function_coverage=1 00:10:10.371 --rc genhtml_legend=1 00:10:10.371 --rc geninfo_all_blocks=1 00:10:10.371 --rc geninfo_unexecuted_blocks=1 00:10:10.371 00:10:10.371 ' 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:10:10.371 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:10.371 --rc genhtml_branch_coverage=1 00:10:10.371 --rc genhtml_function_coverage=1 00:10:10.371 --rc genhtml_legend=1 00:10:10.371 --rc geninfo_all_blocks=1 00:10:10.371 --rc geninfo_unexecuted_blocks=1 00:10:10.371 00:10:10.371 ' 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:10:10.371 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:10.371 --rc genhtml_branch_coverage=1 00:10:10.371 --rc genhtml_function_coverage=1 00:10:10.371 --rc genhtml_legend=1 00:10:10.371 --rc geninfo_all_blocks=1 00:10:10.371 --rc geninfo_unexecuted_blocks=1 00:10:10.371 00:10:10.371 ' 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:10:10.371 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:10.371 --rc genhtml_branch_coverage=1 00:10:10.371 --rc genhtml_function_coverage=1 00:10:10.371 --rc genhtml_legend=1 00:10:10.371 --rc geninfo_all_blocks=1 00:10:10.371 --rc geninfo_unexecuted_blocks=1 00:10:10.371 00:10:10.371 ' 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@10 -- # uname -s 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@10 -- # '[' '!' Linux = Linux ']' 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@14 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@7 -- # uname -s 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@15 -- # shopt -s extglob 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- paths/export.sh@5 -- # export PATH 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@51 -- # : 0 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:10:10.371 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@55 -- # have_pci_nics=0 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@16 -- # trap 'exit 1' SIGINT SIGTERM EXIT 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@18 -- # TEST_ARGS=("$@") 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@20 -- # [[ 0 -eq 0 ]] 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@21 -- # run_test nvmf_abort /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort.sh --transport=tcp 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:10.371 17:25:52 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:10:10.371 ************************************ 00:10:10.371 START TEST nvmf_abort 00:10:10.371 ************************************ 00:10:10.372 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort.sh --transport=tcp 00:10:10.372 * Looking for test storage... 00:10:10.372 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:10.372 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:10:10.372 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1711 -- # lcov --version 00:10:10.372 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@333 -- # local ver1 ver1_l 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@334 -- # local ver2 ver2_l 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@336 -- # IFS=.-: 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@336 -- # read -ra ver1 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@337 -- # IFS=.-: 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@337 -- # read -ra ver2 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@338 -- # local 'op=<' 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@340 -- # ver1_l=2 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@341 -- # ver2_l=1 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@344 -- # case "$op" in 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@345 -- # : 1 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@364 -- # (( v = 0 )) 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@365 -- # decimal 1 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@353 -- # local d=1 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@355 -- # echo 1 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@365 -- # ver1[v]=1 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@366 -- # decimal 2 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@353 -- # local d=2 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@355 -- # echo 2 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@366 -- # ver2[v]=2 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@368 -- # return 0 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:10:10.634 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:10.634 --rc genhtml_branch_coverage=1 00:10:10.634 --rc genhtml_function_coverage=1 00:10:10.634 --rc genhtml_legend=1 00:10:10.634 --rc geninfo_all_blocks=1 00:10:10.634 --rc geninfo_unexecuted_blocks=1 00:10:10.634 00:10:10.634 ' 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:10:10.634 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:10.634 --rc genhtml_branch_coverage=1 00:10:10.634 --rc genhtml_function_coverage=1 00:10:10.634 --rc genhtml_legend=1 00:10:10.634 --rc geninfo_all_blocks=1 00:10:10.634 --rc geninfo_unexecuted_blocks=1 00:10:10.634 00:10:10.634 ' 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:10:10.634 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:10.634 --rc genhtml_branch_coverage=1 00:10:10.634 --rc genhtml_function_coverage=1 00:10:10.634 --rc genhtml_legend=1 00:10:10.634 --rc geninfo_all_blocks=1 00:10:10.634 --rc geninfo_unexecuted_blocks=1 00:10:10.634 00:10:10.634 ' 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:10:10.634 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:10.634 --rc genhtml_branch_coverage=1 00:10:10.634 --rc genhtml_function_coverage=1 00:10:10.634 --rc genhtml_legend=1 00:10:10.634 --rc geninfo_all_blocks=1 00:10:10.634 --rc geninfo_unexecuted_blocks=1 00:10:10.634 00:10:10.634 ' 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@7 -- # uname -s 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:10:10.634 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@15 -- # shopt -s extglob 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- paths/export.sh@5 -- # export PATH 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@51 -- # : 0 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:10:10.635 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@55 -- # have_pci_nics=0 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@11 -- # MALLOC_BDEV_SIZE=64 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@12 -- # MALLOC_BLOCK_SIZE=4096 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@14 -- # nvmftestinit 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@476 -- # prepare_net_devs 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@438 -- # local -g is_hw=no 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@440 -- # remove_spdk_ns 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@309 -- # xtrace_disable 00:10:10.635 17:25:52 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@315 -- # pci_devs=() 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@315 -- # local -a pci_devs 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@316 -- # pci_net_devs=() 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@317 -- # pci_drivers=() 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@317 -- # local -A pci_drivers 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@319 -- # net_devs=() 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@319 -- # local -ga net_devs 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@320 -- # e810=() 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@320 -- # local -ga e810 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@321 -- # x722=() 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@321 -- # local -ga x722 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@322 -- # mlx=() 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@322 -- # local -ga mlx 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:10:12.546 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:10:12.546 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@418 -- # [[ up == up ]] 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:10:12.546 Found net devices under 0000:0a:00.0: cvl_0_0 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:12.546 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:10:12.547 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:10:12.547 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@418 -- # [[ up == up ]] 00:10:12.547 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:10:12.547 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:12.547 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:10:12.547 Found net devices under 0000:0a:00.1: cvl_0_1 00:10:12.547 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:10:12.547 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:10:12.547 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@442 -- # is_hw=yes 00:10:12.547 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:10:12.547 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:10:12.547 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:10:12.547 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:10:12.547 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:10:12.547 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:10:12.547 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:10:12.547 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:10:12.547 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:10:12.547 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:10:12.547 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:10:12.547 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:10:12.547 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:10:12.547 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:10:12.547 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:10:12.547 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:10:12.547 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:10:12.547 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:10:12.807 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:10:12.807 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.273 ms 00:10:12.807 00:10:12.807 --- 10.0.0.2 ping statistics --- 00:10:12.807 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:12.807 rtt min/avg/max/mdev = 0.273/0.273/0.273/0.000 ms 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:10:12.807 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:10:12.807 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.141 ms 00:10:12.807 00:10:12.807 --- 10.0.0.1 ping statistics --- 00:10:12.807 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:12.807 rtt min/avg/max/mdev = 0.141/0.141/0.141/0.000 ms 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@450 -- # return 0 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@15 -- # nvmfappstart -m 0xE 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@726 -- # xtrace_disable 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@509 -- # nvmfpid=124181 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@510 -- # waitforlisten 124181 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@835 -- # '[' -z 124181 ']' 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:12.807 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:12.807 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:10:12.807 [2024-12-06 17:25:54.571419] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:10:12.807 [2024-12-06 17:25:54.571489] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:12.807 [2024-12-06 17:25:54.640425] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:10:13.067 [2024-12-06 17:25:54.687499] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:10:13.067 [2024-12-06 17:25:54.687551] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:10:13.067 [2024-12-06 17:25:54.687574] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:10:13.067 [2024-12-06 17:25:54.687585] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:10:13.067 [2024-12-06 17:25:54.687594] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:10:13.067 [2024-12-06 17:25:54.688937] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:10:13.067 [2024-12-06 17:25:54.689074] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:10:13.067 [2024-12-06 17:25:54.689078] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@868 -- # return 0 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@732 -- # xtrace_disable 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -a 256 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:10:13.067 [2024-12-06 17:25:54.831296] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@20 -- # rpc_cmd bdev_malloc_create 64 4096 -b Malloc0 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:10:13.067 Malloc0 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@21 -- # rpc_cmd bdev_delay_create -b Malloc0 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:10:13.067 Delay0 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@25 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 Delay0 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@26 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.067 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:10:13.327 [2024-12-06 17:25:54.907030] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:10:13.327 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.327 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:10:13.327 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:13.327 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:10:13.327 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:13.327 17:25:54 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -c 0x1 -t 1 -l warning -q 128 00:10:13.327 [2024-12-06 17:25:55.022479] nvme_fabric.c: 295:nvme_fabric_discover_probe: *WARNING*: Skipping unsupported current discovery service or discovery service referral 00:10:15.862 Initializing NVMe Controllers 00:10:15.862 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode0 00:10:15.862 controller IO queue size 128 less than required 00:10:15.862 Consider using lower queue depth or small IO size because IO requests may be queued at the NVMe driver. 00:10:15.862 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 0 00:10:15.862 Initialization complete. Launching workers. 00:10:15.862 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 I/O completed: 123, failed: 29106 00:10:15.862 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) abort submitted 29167, failed to submit 62 00:10:15.862 success 29110, unsuccessful 57, failed 0 00:10:15.862 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@34 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:10:15.862 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:10:15.862 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:10:15.862 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:10:15.862 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@36 -- # trap - SIGINT SIGTERM EXIT 00:10:15.862 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@38 -- # nvmftestfini 00:10:15.862 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@516 -- # nvmfcleanup 00:10:15.862 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@121 -- # sync 00:10:15.862 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:10:15.862 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@124 -- # set +e 00:10:15.862 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@125 -- # for i in {1..20} 00:10:15.862 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:10:15.862 rmmod nvme_tcp 00:10:15.862 rmmod nvme_fabrics 00:10:15.862 rmmod nvme_keyring 00:10:15.862 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:10:15.862 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@128 -- # set -e 00:10:15.862 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@129 -- # return 0 00:10:15.863 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@517 -- # '[' -n 124181 ']' 00:10:15.863 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@518 -- # killprocess 124181 00:10:15.863 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@954 -- # '[' -z 124181 ']' 00:10:15.863 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@958 -- # kill -0 124181 00:10:15.863 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@959 -- # uname 00:10:15.863 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:10:15.863 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 124181 00:10:15.863 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:10:15.863 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:10:15.863 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@972 -- # echo 'killing process with pid 124181' 00:10:15.863 killing process with pid 124181 00:10:15.863 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@973 -- # kill 124181 00:10:15.863 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@978 -- # wait 124181 00:10:15.863 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:10:15.863 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:10:15.863 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:10:15.863 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@297 -- # iptr 00:10:15.863 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@791 -- # iptables-save 00:10:15.863 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:10:15.863 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@791 -- # iptables-restore 00:10:15.863 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:10:15.863 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@302 -- # remove_spdk_ns 00:10:15.863 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:15.863 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:10:15.863 17:25:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:17.779 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:10:17.779 00:10:17.779 real 0m7.388s 00:10:17.779 user 0m10.851s 00:10:17.779 sys 0m2.379s 00:10:17.779 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1130 -- # xtrace_disable 00:10:17.779 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:10:17.779 ************************************ 00:10:17.779 END TEST nvmf_abort 00:10:17.779 ************************************ 00:10:17.779 17:25:59 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@22 -- # run_test nvmf_ns_hotplug_stress /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh --transport=tcp 00:10:17.779 17:25:59 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:10:17.779 17:25:59 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1111 -- # xtrace_disable 00:10:17.779 17:25:59 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:10:17.779 ************************************ 00:10:17.779 START TEST nvmf_ns_hotplug_stress 00:10:17.779 ************************************ 00:10:17.779 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh --transport=tcp 00:10:17.779 * Looking for test storage... 00:10:17.779 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:17.779 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:10:17.779 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1711 -- # lcov --version 00:10:17.779 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@333 -- # local ver1 ver1_l 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@334 -- # local ver2 ver2_l 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@336 -- # IFS=.-: 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@336 -- # read -ra ver1 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@337 -- # IFS=.-: 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@337 -- # read -ra ver2 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@338 -- # local 'op=<' 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@340 -- # ver1_l=2 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@341 -- # ver2_l=1 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@344 -- # case "$op" in 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@345 -- # : 1 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@364 -- # (( v = 0 )) 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@365 -- # decimal 1 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@353 -- # local d=1 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@355 -- # echo 1 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@365 -- # ver1[v]=1 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@366 -- # decimal 2 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@353 -- # local d=2 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@355 -- # echo 2 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@366 -- # ver2[v]=2 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@368 -- # return 0 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:10:18.039 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:18.039 --rc genhtml_branch_coverage=1 00:10:18.039 --rc genhtml_function_coverage=1 00:10:18.039 --rc genhtml_legend=1 00:10:18.039 --rc geninfo_all_blocks=1 00:10:18.039 --rc geninfo_unexecuted_blocks=1 00:10:18.039 00:10:18.039 ' 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:10:18.039 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:18.039 --rc genhtml_branch_coverage=1 00:10:18.039 --rc genhtml_function_coverage=1 00:10:18.039 --rc genhtml_legend=1 00:10:18.039 --rc geninfo_all_blocks=1 00:10:18.039 --rc geninfo_unexecuted_blocks=1 00:10:18.039 00:10:18.039 ' 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:10:18.039 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:18.039 --rc genhtml_branch_coverage=1 00:10:18.039 --rc genhtml_function_coverage=1 00:10:18.039 --rc genhtml_legend=1 00:10:18.039 --rc geninfo_all_blocks=1 00:10:18.039 --rc geninfo_unexecuted_blocks=1 00:10:18.039 00:10:18.039 ' 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:10:18.039 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:18.039 --rc genhtml_branch_coverage=1 00:10:18.039 --rc genhtml_function_coverage=1 00:10:18.039 --rc genhtml_legend=1 00:10:18.039 --rc geninfo_all_blocks=1 00:10:18.039 --rc geninfo_unexecuted_blocks=1 00:10:18.039 00:10:18.039 ' 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@7 -- # uname -s 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@15 -- # shopt -s extglob 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:10:18.039 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- paths/export.sh@5 -- # export PATH 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@51 -- # : 0 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:10:18.040 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@55 -- # have_pci_nics=0 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@11 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@22 -- # nvmftestinit 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@476 -- # prepare_net_devs 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@438 -- # local -g is_hw=no 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@440 -- # remove_spdk_ns 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@309 -- # xtrace_disable 00:10:18.040 17:25:59 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@315 -- # pci_devs=() 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@315 -- # local -a pci_devs 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@316 -- # pci_net_devs=() 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@317 -- # pci_drivers=() 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@317 -- # local -A pci_drivers 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@319 -- # net_devs=() 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@319 -- # local -ga net_devs 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@320 -- # e810=() 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@320 -- # local -ga e810 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@321 -- # x722=() 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@321 -- # local -ga x722 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@322 -- # mlx=() 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@322 -- # local -ga mlx 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:10:20.576 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:10:20.576 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:10:20.577 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@418 -- # [[ up == up ]] 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:10:20.577 Found net devices under 0000:0a:00.0: cvl_0_0 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@418 -- # [[ up == up ]] 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:10:20.577 Found net devices under 0000:0a:00.1: cvl_0_1 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@442 -- # is_hw=yes 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:10:20.577 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:10:20.577 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.352 ms 00:10:20.577 00:10:20.577 --- 10.0.0.2 ping statistics --- 00:10:20.577 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:20.577 rtt min/avg/max/mdev = 0.352/0.352/0.352/0.000 ms 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:10:20.577 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:10:20.577 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.172 ms 00:10:20.577 00:10:20.577 --- 10.0.0.1 ping statistics --- 00:10:20.577 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:20.577 rtt min/avg/max/mdev = 0.172/0.172/0.172/0.000 ms 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@450 -- # return 0 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:10:20.577 17:26:01 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:10:20.577 17:26:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@23 -- # nvmfappstart -m 0xE 00:10:20.577 17:26:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:10:20.577 17:26:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@726 -- # xtrace_disable 00:10:20.577 17:26:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:10:20.577 17:26:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@509 -- # nvmfpid=126417 00:10:20.577 17:26:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:10:20.577 17:26:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@510 -- # waitforlisten 126417 00:10:20.577 17:26:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@835 -- # '[' -z 126417 ']' 00:10:20.577 17:26:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:20.577 17:26:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@840 -- # local max_retries=100 00:10:20.577 17:26:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:20.577 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:20.577 17:26:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@844 -- # xtrace_disable 00:10:20.577 17:26:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:10:20.577 [2024-12-06 17:26:02.062857] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:10:20.577 [2024-12-06 17:26:02.062946] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:20.577 [2024-12-06 17:26:02.141259] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:10:20.577 [2024-12-06 17:26:02.188459] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:10:20.577 [2024-12-06 17:26:02.188529] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:10:20.577 [2024-12-06 17:26:02.188542] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:10:20.577 [2024-12-06 17:26:02.188564] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:10:20.577 [2024-12-06 17:26:02.188574] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:10:20.577 [2024-12-06 17:26:02.190010] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:10:20.577 [2024-12-06 17:26:02.190031] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:10:20.577 [2024-12-06 17:26:02.190034] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:10:20.577 17:26:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:10:20.577 17:26:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@868 -- # return 0 00:10:20.577 17:26:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:10:20.578 17:26:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@732 -- # xtrace_disable 00:10:20.578 17:26:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:10:20.578 17:26:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:10:20.578 17:26:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@25 -- # null_size=1000 00:10:20.578 17:26:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:10:20.835 [2024-12-06 17:26:02.589168] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:10:20.835 17:26:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:10:21.093 17:26:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:10:21.350 [2024-12-06 17:26:03.127899] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:10:21.350 17:26:03 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:10:21.608 17:26:03 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 512 -b Malloc0 00:10:22.174 Malloc0 00:10:22.174 17:26:03 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_delay_create -b Malloc0 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:10:22.174 Delay0 00:10:22.174 17:26:03 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:22.431 17:26:04 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create NULL1 1000 512 00:10:22.994 NULL1 00:10:22.994 17:26:04 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 NULL1 00:10:22.994 17:26:04 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@42 -- # PERF_PID=126844 00:10:22.994 17:26:04 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0x1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 30 -q 128 -w randread -o 512 -Q 1000 00:10:22.994 17:26:04 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:22.994 17:26:04 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:24.361 Read completed with error (sct=0, sc=11) 00:10:24.361 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:24.361 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:10:24.361 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:10:24.361 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:10:24.618 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:10:24.618 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:10:24.618 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:10:24.618 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:10:24.619 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1001 00:10:24.619 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1001 00:10:24.877 true 00:10:24.877 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:24.877 17:26:06 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:25.812 17:26:07 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:25.812 17:26:07 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1002 00:10:25.812 17:26:07 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1002 00:10:26.071 true 00:10:26.329 17:26:07 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:26.329 17:26:07 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:26.587 17:26:08 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:26.845 17:26:08 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1003 00:10:26.845 17:26:08 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1003 00:10:27.103 true 00:10:27.103 17:26:08 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:27.103 17:26:08 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:27.361 17:26:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:27.620 17:26:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1004 00:10:27.620 17:26:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1004 00:10:27.879 true 00:10:27.879 17:26:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:27.879 17:26:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:28.831 17:26:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:28.831 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:10:29.090 17:26:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1005 00:10:29.090 17:26:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1005 00:10:29.348 true 00:10:29.348 17:26:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:29.348 17:26:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:29.606 17:26:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:29.864 17:26:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1006 00:10:29.864 17:26:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1006 00:10:30.122 true 00:10:30.122 17:26:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:30.122 17:26:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:31.057 17:26:12 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:31.057 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:10:31.057 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:10:31.057 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:10:31.315 17:26:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1007 00:10:31.315 17:26:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1007 00:10:31.573 true 00:10:31.573 17:26:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:31.573 17:26:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:31.832 17:26:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:32.090 17:26:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1008 00:10:32.090 17:26:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1008 00:10:32.349 true 00:10:32.349 17:26:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:32.349 17:26:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:33.283 17:26:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:33.283 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:10:33.540 17:26:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1009 00:10:33.540 17:26:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1009 00:10:33.797 true 00:10:33.797 17:26:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:33.797 17:26:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:34.054 17:26:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:34.311 17:26:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1010 00:10:34.311 17:26:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1010 00:10:34.570 true 00:10:34.570 17:26:16 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:34.570 17:26:16 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:35.504 17:26:17 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:35.504 17:26:17 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1011 00:10:35.504 17:26:17 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1011 00:10:35.761 true 00:10:35.761 17:26:17 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:35.761 17:26:17 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:36.019 17:26:17 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:36.584 17:26:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1012 00:10:36.584 17:26:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1012 00:10:36.584 true 00:10:36.584 17:26:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:36.584 17:26:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:36.842 17:26:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:37.101 17:26:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1013 00:10:37.101 17:26:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1013 00:10:37.667 true 00:10:37.667 17:26:19 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:37.667 17:26:19 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:38.623 17:26:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:38.880 17:26:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1014 00:10:38.880 17:26:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1014 00:10:39.138 true 00:10:39.138 17:26:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:39.138 17:26:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:39.396 17:26:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:39.653 17:26:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1015 00:10:39.653 17:26:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1015 00:10:39.910 true 00:10:39.910 17:26:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:39.910 17:26:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:40.167 17:26:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:40.425 17:26:22 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1016 00:10:40.425 17:26:22 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1016 00:10:40.682 true 00:10:40.682 17:26:22 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:40.682 17:26:22 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:41.615 17:26:23 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:41.872 17:26:23 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1017 00:10:41.872 17:26:23 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1017 00:10:42.130 true 00:10:42.130 17:26:23 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:42.130 17:26:23 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:42.387 17:26:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:42.952 17:26:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1018 00:10:42.952 17:26:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1018 00:10:42.952 true 00:10:42.952 17:26:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:42.952 17:26:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:43.211 17:26:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:43.774 17:26:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1019 00:10:43.774 17:26:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1019 00:10:43.774 true 00:10:43.774 17:26:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:43.774 17:26:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:44.707 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:10:44.707 17:26:26 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:44.707 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:10:44.707 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:10:44.965 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:10:44.965 17:26:26 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1020 00:10:44.965 17:26:26 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1020 00:10:45.222 true 00:10:45.222 17:26:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:45.222 17:26:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:45.480 17:26:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:45.737 17:26:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1021 00:10:45.737 17:26:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1021 00:10:45.996 true 00:10:45.996 17:26:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:45.996 17:26:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:46.930 17:26:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:47.188 17:26:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1022 00:10:47.188 17:26:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1022 00:10:47.446 true 00:10:47.446 17:26:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:47.446 17:26:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:47.705 17:26:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:47.963 17:26:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1023 00:10:47.963 17:26:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1023 00:10:48.221 true 00:10:48.221 17:26:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:48.221 17:26:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:48.786 17:26:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:48.786 17:26:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1024 00:10:48.786 17:26:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1024 00:10:49.045 true 00:10:49.303 17:26:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:49.303 17:26:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:50.238 17:26:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:50.496 17:26:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1025 00:10:50.496 17:26:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1025 00:10:50.754 true 00:10:50.754 17:26:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:50.754 17:26:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:51.013 17:26:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:51.271 17:26:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1026 00:10:51.271 17:26:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1026 00:10:51.528 true 00:10:51.528 17:26:33 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:51.528 17:26:33 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:51.785 17:26:33 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:52.043 17:26:33 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1027 00:10:52.043 17:26:33 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1027 00:10:52.301 true 00:10:52.302 17:26:34 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:52.302 17:26:34 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:53.237 17:26:35 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:53.495 Initializing NVMe Controllers 00:10:53.495 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:10:53.495 Controller IO queue size 128, less than required. 00:10:53.495 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:10:53.495 Controller IO queue size 128, less than required. 00:10:53.495 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:10:53.495 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:10:53.495 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:10:53.495 Initialization complete. Launching workers. 00:10:53.495 ======================================================== 00:10:53.495 Latency(us) 00:10:53.495 Device Information : IOPS MiB/s Average min max 00:10:53.495 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 609.60 0.30 94273.16 3403.45 1102520.89 00:10:53.495 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 9233.44 4.51 13864.06 3245.47 492070.44 00:10:53.495 ======================================================== 00:10:53.495 Total : 9843.04 4.81 18843.93 3245.47 1102520.89 00:10:53.495 00:10:53.495 17:26:35 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1028 00:10:53.495 17:26:35 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1028 00:10:53.753 true 00:10:54.010 17:26:35 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 126844 00:10:54.010 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh: line 44: kill: (126844) - No such process 00:10:54.011 17:26:35 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@53 -- # wait 126844 00:10:54.011 17:26:35 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:54.268 17:26:35 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:10:54.526 17:26:36 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@58 -- # nthreads=8 00:10:54.526 17:26:36 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@58 -- # pids=() 00:10:54.526 17:26:36 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i = 0 )) 00:10:54.526 17:26:36 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:10:54.526 17:26:36 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null0 100 4096 00:10:54.784 null0 00:10:54.784 17:26:36 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:10:54.784 17:26:36 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:10:54.784 17:26:36 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null1 100 4096 00:10:55.042 null1 00:10:55.042 17:26:36 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:10:55.042 17:26:36 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:10:55.042 17:26:36 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null2 100 4096 00:10:55.300 null2 00:10:55.300 17:26:36 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:10:55.300 17:26:36 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:10:55.300 17:26:36 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null3 100 4096 00:10:55.558 null3 00:10:55.558 17:26:37 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:10:55.558 17:26:37 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:10:55.558 17:26:37 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null4 100 4096 00:10:55.816 null4 00:10:55.816 17:26:37 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:10:55.816 17:26:37 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:10:55.816 17:26:37 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null5 100 4096 00:10:56.075 null5 00:10:56.075 17:26:37 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:10:56.075 17:26:37 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:10:56.075 17:26:37 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null6 100 4096 00:10:56.333 null6 00:10:56.333 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:10:56.333 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:10:56.333 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null7 100 4096 00:10:56.592 null7 00:10:56.592 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i = 0 )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 1 null0 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=1 bdev=null0 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 2 null1 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=2 bdev=null1 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 3 null2 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=3 bdev=null2 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 4 null3 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=4 bdev=null3 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 5 null4 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=5 bdev=null4 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 6 null5 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=6 bdev=null5 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 7 null6 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=7 bdev=null6 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 8 null7 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=8 bdev=null7 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@66 -- # wait 130925 130926 130927 130930 130932 130934 130936 130938 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:56.593 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:10:56.852 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:10:56.852 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:10:56.852 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:10:56.852 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:56.852 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:10:56.852 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:10:56.852 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:10:56.852 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:10:57.133 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:57.133 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:57.133 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:10:57.133 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:57.133 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:57.133 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:10:57.133 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:57.133 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:57.133 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:10:57.133 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:57.133 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:57.133 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:10:57.133 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:57.133 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:57.133 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:57.133 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:10:57.133 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:57.133 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:10:57.133 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:57.133 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:57.133 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:10:57.133 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:57.133 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:57.133 17:26:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:10:57.392 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:10:57.650 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:10:57.650 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:10:57.650 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:10:57.650 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:57.650 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:10:57.650 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:10:57.650 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:10:57.907 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:57.907 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:57.907 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:10:57.907 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:57.907 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:57.907 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:10:57.907 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:57.907 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:57.907 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:10:57.907 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:57.907 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:57.907 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:10:57.907 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:57.907 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:57.907 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:10:57.908 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:57.908 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:57.908 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:10:57.908 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:57.908 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:57.908 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:10:57.908 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:57.908 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:57.908 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:10:58.165 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:10:58.165 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:10:58.165 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:10:58.165 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:58.165 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:10:58.165 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:10:58.165 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:10:58.165 17:26:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:10:58.424 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:58.424 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:58.424 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:10:58.424 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:58.424 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:58.424 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:10:58.424 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:58.424 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:58.424 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:10:58.424 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:58.424 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:58.424 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:10:58.424 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:58.424 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:58.424 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:10:58.424 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:58.425 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:58.425 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:10:58.425 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:58.425 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:58.425 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:10:58.425 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:58.425 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:58.425 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:10:58.683 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:10:58.683 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:10:58.683 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:10:58.683 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:58.683 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:10:58.683 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:10:58.683 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:10:58.683 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:10:58.940 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:58.940 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:58.940 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:10:58.940 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:58.940 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:58.940 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:10:58.940 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:58.940 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:58.940 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:10:58.940 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:58.940 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:58.940 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:10:58.940 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:58.940 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:58.940 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:10:58.940 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:58.940 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:58.940 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:10:58.940 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:58.940 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:58.940 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:10:58.940 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:58.940 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:58.940 17:26:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:10:59.197 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:10:59.454 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:10:59.454 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:10:59.454 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:10:59.454 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:10:59.454 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:59.454 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:10:59.454 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:10:59.711 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:59.711 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:59.711 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:10:59.711 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:59.711 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:59.711 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:10:59.711 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:59.711 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:59.711 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:10:59.711 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:59.711 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:59.711 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:10:59.711 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:59.711 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:59.711 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:10:59.711 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:59.711 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:59.711 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:59.711 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:59.711 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:10:59.711 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:10:59.711 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:10:59.711 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:10:59.711 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:10:59.969 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:10:59.969 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:59.969 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:10:59.969 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:10:59.969 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:10:59.969 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:10:59.969 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:10:59.969 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:11:00.227 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:00.227 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:00.227 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:11:00.227 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:00.227 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:00.227 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:11:00.227 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:00.227 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:00.227 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:11:00.227 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:00.227 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:00.227 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:11:00.227 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:00.227 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:00.227 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:11:00.227 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:00.227 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:00.227 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:00.227 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:00.227 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:11:00.227 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:11:00.227 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:00.227 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:00.227 17:26:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:11:00.485 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:11:00.486 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:11:00.486 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:11:00.486 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:11:00.486 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:11:00.486 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:11:00.486 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:11:00.486 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:11:00.743 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:00.743 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:00.743 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:11:00.743 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:00.743 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:00.743 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:11:00.743 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:00.743 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:00.743 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:11:00.743 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:00.743 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:00.743 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:11:00.743 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:00.744 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:00.744 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:11:00.744 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:00.744 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:00.744 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:11:00.744 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:00.744 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:00.744 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:11:00.744 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:00.744 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:00.744 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:11:01.309 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:11:01.309 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:11:01.309 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:11:01.309 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:11:01.309 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:11:01.309 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:11:01.309 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:11:01.309 17:26:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:11:01.567 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:01.567 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:01.567 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:11:01.567 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:01.567 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:01.567 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:11:01.567 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:01.567 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:01.567 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:11:01.567 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:01.567 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:01.567 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:11:01.567 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:01.567 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:01.567 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:11:01.567 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:01.567 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:01.567 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:11:01.567 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:01.567 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:01.567 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:11:01.567 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:01.567 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:01.567 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:11:01.826 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:11:01.826 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:11:01.826 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:11:01.826 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:11:01.826 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:11:01.826 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:11:01.826 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:11:01.826 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:11:02.085 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:02.085 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:02.085 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:11:02.085 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:02.085 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:02.085 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:11:02.085 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:02.085 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:02.085 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:11:02.085 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:02.085 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:02.085 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:11:02.085 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:02.085 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:02.085 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:11:02.085 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:02.085 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:02.085 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:11:02.085 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:02.085 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:02.085 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:11:02.085 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:02.085 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:02.085 17:26:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:11:02.344 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:11:02.344 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:11:02.344 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:11:02.344 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:11:02.344 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:11:02.344 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:11:02.344 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:11:02.344 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:11:02.604 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:02.604 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:02.604 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:02.604 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:02.604 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:02.604 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:02.604 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:02.604 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:02.604 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:02.604 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:02.604 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:02.604 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:02.604 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:02.604 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:02.604 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:11:02.604 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:11:02.604 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@68 -- # trap - SIGINT SIGTERM EXIT 00:11:02.604 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@70 -- # nvmftestfini 00:11:02.604 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@516 -- # nvmfcleanup 00:11:02.604 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@121 -- # sync 00:11:02.604 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:11:02.604 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@124 -- # set +e 00:11:02.604 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@125 -- # for i in {1..20} 00:11:02.604 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:11:02.604 rmmod nvme_tcp 00:11:02.604 rmmod nvme_fabrics 00:11:02.864 rmmod nvme_keyring 00:11:02.864 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:11:02.864 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@128 -- # set -e 00:11:02.864 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@129 -- # return 0 00:11:02.864 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@517 -- # '[' -n 126417 ']' 00:11:02.864 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@518 -- # killprocess 126417 00:11:02.864 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@954 -- # '[' -z 126417 ']' 00:11:02.864 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@958 -- # kill -0 126417 00:11:02.864 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@959 -- # uname 00:11:02.864 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:02.864 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 126417 00:11:02.864 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:11:02.864 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:11:02.864 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@972 -- # echo 'killing process with pid 126417' 00:11:02.864 killing process with pid 126417 00:11:02.864 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@973 -- # kill 126417 00:11:02.864 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@978 -- # wait 126417 00:11:03.124 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:11:03.124 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:11:03.124 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:11:03.124 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@297 -- # iptr 00:11:03.124 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@791 -- # iptables-save 00:11:03.124 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:11:03.124 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@791 -- # iptables-restore 00:11:03.124 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:11:03.124 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@302 -- # remove_spdk_ns 00:11:03.124 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:03.125 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:11:03.125 17:26:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:05.037 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:11:05.037 00:11:05.037 real 0m47.229s 00:11:05.037 user 3m40.175s 00:11:05.037 sys 0m15.478s 00:11:05.037 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:05.037 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:11:05.037 ************************************ 00:11:05.037 END TEST nvmf_ns_hotplug_stress 00:11:05.037 ************************************ 00:11:05.037 17:26:46 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@23 -- # run_test nvmf_delete_subsystem /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh --transport=tcp 00:11:05.037 17:26:46 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:11:05.037 17:26:46 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:05.037 17:26:46 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:11:05.037 ************************************ 00:11:05.037 START TEST nvmf_delete_subsystem 00:11:05.037 ************************************ 00:11:05.037 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh --transport=tcp 00:11:05.297 * Looking for test storage... 00:11:05.297 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:11:05.297 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:11:05.297 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1711 -- # lcov --version 00:11:05.297 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:11:05.297 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:11:05.297 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:11:05.297 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@333 -- # local ver1 ver1_l 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@334 -- # local ver2 ver2_l 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@336 -- # IFS=.-: 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@336 -- # read -ra ver1 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@337 -- # IFS=.-: 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@337 -- # read -ra ver2 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@338 -- # local 'op=<' 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@340 -- # ver1_l=2 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@341 -- # ver2_l=1 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@344 -- # case "$op" in 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@345 -- # : 1 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@364 -- # (( v = 0 )) 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@365 -- # decimal 1 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@353 -- # local d=1 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@355 -- # echo 1 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@365 -- # ver1[v]=1 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@366 -- # decimal 2 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@353 -- # local d=2 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@355 -- # echo 2 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@366 -- # ver2[v]=2 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@368 -- # return 0 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:11:05.298 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:05.298 --rc genhtml_branch_coverage=1 00:11:05.298 --rc genhtml_function_coverage=1 00:11:05.298 --rc genhtml_legend=1 00:11:05.298 --rc geninfo_all_blocks=1 00:11:05.298 --rc geninfo_unexecuted_blocks=1 00:11:05.298 00:11:05.298 ' 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:11:05.298 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:05.298 --rc genhtml_branch_coverage=1 00:11:05.298 --rc genhtml_function_coverage=1 00:11:05.298 --rc genhtml_legend=1 00:11:05.298 --rc geninfo_all_blocks=1 00:11:05.298 --rc geninfo_unexecuted_blocks=1 00:11:05.298 00:11:05.298 ' 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:11:05.298 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:05.298 --rc genhtml_branch_coverage=1 00:11:05.298 --rc genhtml_function_coverage=1 00:11:05.298 --rc genhtml_legend=1 00:11:05.298 --rc geninfo_all_blocks=1 00:11:05.298 --rc geninfo_unexecuted_blocks=1 00:11:05.298 00:11:05.298 ' 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:11:05.298 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:05.298 --rc genhtml_branch_coverage=1 00:11:05.298 --rc genhtml_function_coverage=1 00:11:05.298 --rc genhtml_legend=1 00:11:05.298 --rc geninfo_all_blocks=1 00:11:05.298 --rc geninfo_unexecuted_blocks=1 00:11:05.298 00:11:05.298 ' 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@7 -- # uname -s 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@15 -- # shopt -s extglob 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- paths/export.sh@5 -- # export PATH 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@51 -- # : 0 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:11:05.298 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@55 -- # have_pci_nics=0 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@12 -- # nvmftestinit 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@476 -- # prepare_net_devs 00:11:05.298 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@438 -- # local -g is_hw=no 00:11:05.299 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@440 -- # remove_spdk_ns 00:11:05.299 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:05.299 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:11:05.299 17:26:46 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:05.299 17:26:47 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:11:05.299 17:26:47 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:11:05.299 17:26:47 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@309 -- # xtrace_disable 00:11:05.299 17:26:47 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@315 -- # pci_devs=() 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@315 -- # local -a pci_devs 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@316 -- # pci_net_devs=() 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@317 -- # pci_drivers=() 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@317 -- # local -A pci_drivers 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@319 -- # net_devs=() 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@319 -- # local -ga net_devs 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@320 -- # e810=() 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@320 -- # local -ga e810 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@321 -- # x722=() 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@321 -- # local -ga x722 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@322 -- # mlx=() 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@322 -- # local -ga mlx 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:11:07.836 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:11:07.836 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@418 -- # [[ up == up ]] 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:07.836 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:11:07.836 Found net devices under 0000:0a:00.0: cvl_0_0 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@418 -- # [[ up == up ]] 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:11:07.837 Found net devices under 0000:0a:00.1: cvl_0_1 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@442 -- # is_hw=yes 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:11:07.837 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:11:07.837 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.252 ms 00:11:07.837 00:11:07.837 --- 10.0.0.2 ping statistics --- 00:11:07.837 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:07.837 rtt min/avg/max/mdev = 0.252/0.252/0.252/0.000 ms 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:11:07.837 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:11:07.837 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.164 ms 00:11:07.837 00:11:07.837 --- 10.0.0.1 ping statistics --- 00:11:07.837 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:07.837 rtt min/avg/max/mdev = 0.164/0.164/0.164/0.000 ms 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@450 -- # return 0 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@13 -- # nvmfappstart -m 0x3 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@726 -- # xtrace_disable 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@509 -- # nvmfpid=133830 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@510 -- # waitforlisten 133830 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x3 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@835 -- # '[' -z 133830 ']' 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:07.837 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:11:07.837 [2024-12-06 17:26:49.422365] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:11:07.837 [2024-12-06 17:26:49.422466] [ DPDK EAL parameters: nvmf -c 0x3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:07.837 [2024-12-06 17:26:49.492654] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:11:07.837 [2024-12-06 17:26:49.534352] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:11:07.837 [2024-12-06 17:26:49.534411] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:11:07.837 [2024-12-06 17:26:49.534435] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:11:07.837 [2024-12-06 17:26:49.534446] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:11:07.837 [2024-12-06 17:26:49.534454] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:11:07.837 [2024-12-06 17:26:49.535753] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:11:07.837 [2024-12-06 17:26:49.535759] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@868 -- # return 0 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@732 -- # xtrace_disable 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@15 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:11:07.837 [2024-12-06 17:26:49.667341] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@16 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:07.837 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:11:08.096 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.096 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@17 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:11:08.096 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.096 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:11:08.096 [2024-12-06 17:26:49.683521] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:11:08.096 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.096 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@18 -- # rpc_cmd bdev_null_create NULL1 1000 512 00:11:08.096 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.096 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:11:08.096 NULL1 00:11:08.096 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.096 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@23 -- # rpc_cmd bdev_delay_create -b NULL1 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:11:08.096 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.096 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:11:08.096 Delay0 00:11:08.096 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.096 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:11:08.096 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:08.096 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:11:08.096 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:08.096 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@28 -- # perf_pid=133862 00:11:08.096 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0xC -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 5 -q 128 -w randrw -M 70 -o 512 -P 4 00:11:08.096 17:26:49 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@30 -- # sleep 2 00:11:08.096 [2024-12-06 17:26:49.768404] subsystem.c:1641:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:11:10.006 17:26:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@32 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:11:10.006 17:26:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:10.006 17:26:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 starting I/O failed: -6 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 starting I/O failed: -6 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 starting I/O failed: -6 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 starting I/O failed: -6 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 starting I/O failed: -6 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 starting I/O failed: -6 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 starting I/O failed: -6 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 starting I/O failed: -6 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 starting I/O failed: -6 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 starting I/O failed: -6 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 starting I/O failed: -6 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 [2024-12-06 17:26:51.891966] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7f6a78000c40 is same with the state(6) to be set 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 starting I/O failed: -6 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 starting I/O failed: -6 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 starting I/O failed: -6 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 starting I/O failed: -6 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 starting I/O failed: -6 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 starting I/O failed: -6 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 starting I/O failed: -6 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 starting I/O failed: -6 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 Write completed with error (sct=0, sc=8) 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.266 starting I/O failed: -6 00:11:10.266 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Write completed with error (sct=0, sc=8) 00:11:10.267 Write completed with error (sct=0, sc=8) 00:11:10.267 starting I/O failed: -6 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 starting I/O failed: -6 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Write completed with error (sct=0, sc=8) 00:11:10.267 Write completed with error (sct=0, sc=8) 00:11:10.267 Write completed with error (sct=0, sc=8) 00:11:10.267 starting I/O failed: -6 00:11:10.267 [2024-12-06 17:26:51.892754] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdca510 is same with the state(6) to be set 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Write completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Write completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Write completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Write completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Write completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Write completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Write completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Write completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Write completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Write completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Write completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Write completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Write completed with error (sct=0, sc=8) 00:11:10.267 Read completed with error (sct=0, sc=8) 00:11:10.267 Write completed with error (sct=0, sc=8) 00:11:11.203 [2024-12-06 17:26:52.863168] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdc8190 is same with the state(6) to be set 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Write completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Write completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Write completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Write completed with error (sct=0, sc=8) 00:11:11.203 Write completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Write completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Write completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Write completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Write completed with error (sct=0, sc=8) 00:11:11.203 Write completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 [2024-12-06 17:26:52.894114] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdca330 is same with the state(6) to be set 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Write completed with error (sct=0, sc=8) 00:11:11.203 Write completed with error (sct=0, sc=8) 00:11:11.203 Write completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Write completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Write completed with error (sct=0, sc=8) 00:11:11.203 Write completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Read completed with error (sct=0, sc=8) 00:11:11.203 Write completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 [2024-12-06 17:26:52.894368] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7f6a7800d020 is same with the state(6) to be set 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Write completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Write completed with error (sct=0, sc=8) 00:11:11.204 Write completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Write completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Write completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Write completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Write completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Write completed with error (sct=0, sc=8) 00:11:11.204 [2024-12-06 17:26:52.894551] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7f6a7800d7e0 is same with the state(6) to be set 00:11:11.204 Write completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Write completed with error (sct=0, sc=8) 00:11:11.204 Write completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Write completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Write completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Read completed with error (sct=0, sc=8) 00:11:11.204 Write completed with error (sct=0, sc=8) 00:11:11.204 [2024-12-06 17:26:52.894742] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdc9f70 is same with the state(6) to be set 00:11:11.204 Initializing NVMe Controllers 00:11:11.204 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:11:11.204 Controller IO queue size 128, less than required. 00:11:11.204 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:11:11.204 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 2 00:11:11.204 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 3 00:11:11.204 Initialization complete. Launching workers. 00:11:11.204 ======================================================== 00:11:11.204 Latency(us) 00:11:11.204 Device Information : IOPS MiB/s Average min max 00:11:11.204 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 2: 172.57 0.08 891216.86 401.31 1014532.99 00:11:11.204 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 3: 171.08 0.08 893045.96 655.67 1014155.18 00:11:11.204 ======================================================== 00:11:11.204 Total : 343.66 0.17 892127.45 401.31 1014532.99 00:11:11.204 00:11:11.204 [2024-12-06 17:26:52.895895] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xdc8190 (9): Bad file descriptor 00:11:11.204 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf: errors occurred 00:11:11.204 17:26:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:11.204 17:26:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@34 -- # delay=0 00:11:11.204 17:26:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@35 -- # kill -0 133862 00:11:11.204 17:26:52 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@36 -- # sleep 0.5 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@38 -- # (( delay++ > 30 )) 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@35 -- # kill -0 133862 00:11:11.771 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh: line 35: kill: (133862) - No such process 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@45 -- # NOT wait 133862 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@652 -- # local es=0 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@654 -- # valid_exec_arg wait 133862 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@640 -- # local arg=wait 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@644 -- # type -t wait 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@655 -- # wait 133862 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@655 -- # es=1 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@48 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@49 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:11:11.771 [2024-12-06 17:26:53.418048] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@50 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@54 -- # perf_pid=134267 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@56 -- # delay=0 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0xC -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 3 -q 128 -w randrw -M 70 -o 512 -P 4 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 134267 00:11:11.771 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:11:11.771 [2024-12-06 17:26:53.480783] subsystem.c:1641:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:11:12.338 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:11:12.338 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 134267 00:11:12.338 17:26:53 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:11:12.905 17:26:54 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:11:12.905 17:26:54 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 134267 00:11:12.905 17:26:54 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:11:13.163 17:26:54 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:11:13.163 17:26:54 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 134267 00:11:13.163 17:26:54 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:11:13.734 17:26:55 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:11:13.734 17:26:55 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 134267 00:11:13.734 17:26:55 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:11:14.300 17:26:55 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:11:14.300 17:26:55 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 134267 00:11:14.300 17:26:55 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:11:14.867 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:11:14.867 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 134267 00:11:14.867 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:11:15.126 Initializing NVMe Controllers 00:11:15.126 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:11:15.126 Controller IO queue size 128, less than required. 00:11:15.126 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:11:15.126 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 2 00:11:15.126 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 3 00:11:15.126 Initialization complete. Launching workers. 00:11:15.126 ======================================================== 00:11:15.126 Latency(us) 00:11:15.126 Device Information : IOPS MiB/s Average min max 00:11:15.126 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 2: 128.00 0.06 1004222.46 1000216.04 1041498.34 00:11:15.126 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 3: 128.00 0.06 1005052.98 1000224.87 1013412.49 00:11:15.126 ======================================================== 00:11:15.126 Total : 256.00 0.12 1004637.72 1000216.04 1041498.34 00:11:15.126 00:11:15.126 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:11:15.126 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 134267 00:11:15.126 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh: line 57: kill: (134267) - No such process 00:11:15.126 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@67 -- # wait 134267 00:11:15.126 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:11:15.126 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@71 -- # nvmftestfini 00:11:15.126 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@516 -- # nvmfcleanup 00:11:15.126 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@121 -- # sync 00:11:15.126 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:11:15.126 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@124 -- # set +e 00:11:15.126 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@125 -- # for i in {1..20} 00:11:15.126 17:26:56 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:11:15.126 rmmod nvme_tcp 00:11:15.385 rmmod nvme_fabrics 00:11:15.385 rmmod nvme_keyring 00:11:15.385 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:11:15.385 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@128 -- # set -e 00:11:15.385 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@129 -- # return 0 00:11:15.385 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@517 -- # '[' -n 133830 ']' 00:11:15.385 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@518 -- # killprocess 133830 00:11:15.385 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@954 -- # '[' -z 133830 ']' 00:11:15.385 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@958 -- # kill -0 133830 00:11:15.385 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@959 -- # uname 00:11:15.385 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:15.385 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 133830 00:11:15.385 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:15.385 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:15.385 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@972 -- # echo 'killing process with pid 133830' 00:11:15.385 killing process with pid 133830 00:11:15.385 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@973 -- # kill 133830 00:11:15.385 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@978 -- # wait 133830 00:11:15.645 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:11:15.645 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:11:15.645 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:11:15.645 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@297 -- # iptr 00:11:15.645 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@791 -- # iptables-save 00:11:15.645 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:11:15.645 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@791 -- # iptables-restore 00:11:15.645 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:11:15.645 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@302 -- # remove_spdk_ns 00:11:15.645 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:15.645 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:11:15.645 17:26:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:17.557 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:11:17.557 00:11:17.557 real 0m12.464s 00:11:17.557 user 0m27.898s 00:11:17.557 sys 0m3.081s 00:11:17.557 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:17.557 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:11:17.557 ************************************ 00:11:17.557 END TEST nvmf_delete_subsystem 00:11:17.557 ************************************ 00:11:17.557 17:26:59 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@26 -- # run_test nvmf_host_management /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh --transport=tcp 00:11:17.557 17:26:59 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:11:17.557 17:26:59 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:17.557 17:26:59 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:11:17.557 ************************************ 00:11:17.557 START TEST nvmf_host_management 00:11:17.557 ************************************ 00:11:17.557 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh --transport=tcp 00:11:17.557 * Looking for test storage... 00:11:17.818 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1711 -- # lcov --version 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@333 -- # local ver1 ver1_l 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@334 -- # local ver2 ver2_l 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@336 -- # IFS=.-: 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@336 -- # read -ra ver1 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@337 -- # IFS=.-: 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@337 -- # read -ra ver2 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@338 -- # local 'op=<' 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@340 -- # ver1_l=2 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@341 -- # ver2_l=1 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@344 -- # case "$op" in 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@345 -- # : 1 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@364 -- # (( v = 0 )) 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@365 -- # decimal 1 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@353 -- # local d=1 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@355 -- # echo 1 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@365 -- # ver1[v]=1 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@366 -- # decimal 2 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@353 -- # local d=2 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@355 -- # echo 2 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@366 -- # ver2[v]=2 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@368 -- # return 0 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:11:17.818 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:17.818 --rc genhtml_branch_coverage=1 00:11:17.818 --rc genhtml_function_coverage=1 00:11:17.818 --rc genhtml_legend=1 00:11:17.818 --rc geninfo_all_blocks=1 00:11:17.818 --rc geninfo_unexecuted_blocks=1 00:11:17.818 00:11:17.818 ' 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:11:17.818 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:17.818 --rc genhtml_branch_coverage=1 00:11:17.818 --rc genhtml_function_coverage=1 00:11:17.818 --rc genhtml_legend=1 00:11:17.818 --rc geninfo_all_blocks=1 00:11:17.818 --rc geninfo_unexecuted_blocks=1 00:11:17.818 00:11:17.818 ' 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:11:17.818 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:17.818 --rc genhtml_branch_coverage=1 00:11:17.818 --rc genhtml_function_coverage=1 00:11:17.818 --rc genhtml_legend=1 00:11:17.818 --rc geninfo_all_blocks=1 00:11:17.818 --rc geninfo_unexecuted_blocks=1 00:11:17.818 00:11:17.818 ' 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:11:17.818 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:17.818 --rc genhtml_branch_coverage=1 00:11:17.818 --rc genhtml_function_coverage=1 00:11:17.818 --rc genhtml_legend=1 00:11:17.818 --rc geninfo_all_blocks=1 00:11:17.818 --rc geninfo_unexecuted_blocks=1 00:11:17.818 00:11:17.818 ' 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@7 -- # uname -s 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@15 -- # shopt -s extglob 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:11:17.818 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- paths/export.sh@5 -- # export PATH 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@51 -- # : 0 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:11:17.819 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@55 -- # have_pci_nics=0 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@11 -- # MALLOC_BDEV_SIZE=64 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@105 -- # nvmftestinit 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@476 -- # prepare_net_devs 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@438 -- # local -g is_hw=no 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@440 -- # remove_spdk_ns 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@309 -- # xtrace_disable 00:11:17.819 17:26:59 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@315 -- # pci_devs=() 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@315 -- # local -a pci_devs 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@316 -- # pci_net_devs=() 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@317 -- # pci_drivers=() 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@317 -- # local -A pci_drivers 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@319 -- # net_devs=() 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@319 -- # local -ga net_devs 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@320 -- # e810=() 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@320 -- # local -ga e810 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@321 -- # x722=() 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@321 -- # local -ga x722 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@322 -- # mlx=() 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@322 -- # local -ga mlx 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:11:19.758 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:19.758 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:11:19.759 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@418 -- # [[ up == up ]] 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:11:19.759 Found net devices under 0000:0a:00.0: cvl_0_0 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@418 -- # [[ up == up ]] 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:11:19.759 Found net devices under 0000:0a:00.1: cvl_0_1 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@442 -- # is_hw=yes 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:11:19.759 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:11:20.016 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:11:20.016 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:11:20.016 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:11:20.016 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:11:20.016 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:11:20.016 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:11:20.017 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:11:20.017 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.238 ms 00:11:20.017 00:11:20.017 --- 10.0.0.2 ping statistics --- 00:11:20.017 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:20.017 rtt min/avg/max/mdev = 0.238/0.238/0.238/0.000 ms 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:11:20.017 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:11:20.017 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.115 ms 00:11:20.017 00:11:20.017 --- 10.0.0.1 ping statistics --- 00:11:20.017 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:20.017 rtt min/avg/max/mdev = 0.115/0.115/0.115/0.000 ms 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@450 -- # return 0 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@107 -- # nvmf_host_management 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@69 -- # starttarget 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@16 -- # nvmfappstart -m 0x1E 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@726 -- # xtrace_disable 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@509 -- # nvmfpid=136843 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1E 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@510 -- # waitforlisten 136843 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@835 -- # '[' -z 136843 ']' 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:20.017 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:20.017 17:27:01 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:11:20.017 [2024-12-06 17:27:01.829693] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:11:20.017 [2024-12-06 17:27:01.829798] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:20.275 [2024-12-06 17:27:01.904242] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:11:20.275 [2024-12-06 17:27:01.950668] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:11:20.275 [2024-12-06 17:27:01.950723] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:11:20.275 [2024-12-06 17:27:01.950746] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:11:20.275 [2024-12-06 17:27:01.950758] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:11:20.275 [2024-12-06 17:27:01.950767] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:11:20.275 [2024-12-06 17:27:01.952202] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:11:20.275 [2024-12-06 17:27:01.952339] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:11:20.275 [2024-12-06 17:27:01.952397] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:11:20.275 [2024-12-06 17:27:01.952400] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:11:20.275 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:20.275 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@868 -- # return 0 00:11:20.275 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:11:20.275 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@732 -- # xtrace_disable 00:11:20.275 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:11:20.275 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:11:20.275 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:11:20.275 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.275 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:11:20.275 [2024-12-06 17:27:02.097317] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:11:20.275 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.275 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@20 -- # timing_enter create_subsystem 00:11:20.275 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@726 -- # xtrace_disable 00:11:20.275 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:11:20.275 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@22 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:11:20.275 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@23 -- # cat 00:11:20.532 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@30 -- # rpc_cmd 00:11:20.532 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:20.532 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:11:20.532 Malloc0 00:11:20.532 [2024-12-06 17:27:02.177883] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:11:20.532 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:20.532 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@31 -- # timing_exit create_subsystems 00:11:20.532 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@732 -- # xtrace_disable 00:11:20.532 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:11:20.532 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@73 -- # perfpid=136900 00:11:20.532 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@74 -- # waitforlisten 136900 /var/tmp/bdevperf.sock 00:11:20.532 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@835 -- # '[' -z 136900 ']' 00:11:20.532 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:11:20.532 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@72 -- # gen_nvmf_target_json 0 00:11:20.532 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock --json /dev/fd/63 -q 64 -o 65536 -w verify -t 10 00:11:20.532 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:20.532 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:11:20.532 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:11:20.532 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@560 -- # config=() 00:11:20.532 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:20.532 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@560 -- # local subsystem config 00:11:20.532 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:11:20.532 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:11:20.532 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:11:20.532 { 00:11:20.532 "params": { 00:11:20.532 "name": "Nvme$subsystem", 00:11:20.532 "trtype": "$TEST_TRANSPORT", 00:11:20.532 "traddr": "$NVMF_FIRST_TARGET_IP", 00:11:20.532 "adrfam": "ipv4", 00:11:20.532 "trsvcid": "$NVMF_PORT", 00:11:20.532 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:11:20.532 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:11:20.532 "hdgst": ${hdgst:-false}, 00:11:20.532 "ddgst": ${ddgst:-false} 00:11:20.532 }, 00:11:20.533 "method": "bdev_nvme_attach_controller" 00:11:20.533 } 00:11:20.533 EOF 00:11:20.533 )") 00:11:20.533 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@582 -- # cat 00:11:20.533 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@584 -- # jq . 00:11:20.533 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@585 -- # IFS=, 00:11:20.533 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:11:20.533 "params": { 00:11:20.533 "name": "Nvme0", 00:11:20.533 "trtype": "tcp", 00:11:20.533 "traddr": "10.0.0.2", 00:11:20.533 "adrfam": "ipv4", 00:11:20.533 "trsvcid": "4420", 00:11:20.533 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:11:20.533 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:11:20.533 "hdgst": false, 00:11:20.533 "ddgst": false 00:11:20.533 }, 00:11:20.533 "method": "bdev_nvme_attach_controller" 00:11:20.533 }' 00:11:20.533 [2024-12-06 17:27:02.263506] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:11:20.533 [2024-12-06 17:27:02.263595] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid136900 ] 00:11:20.533 [2024-12-06 17:27:02.336410] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:20.790 [2024-12-06 17:27:02.384418] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:20.790 Running I/O for 10 seconds... 00:11:21.049 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:21.049 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@868 -- # return 0 00:11:21.049 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@75 -- # rpc_cmd -s /var/tmp/bdevperf.sock framework_wait_init 00:11:21.049 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.049 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:11:21.049 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.049 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@78 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; kill -9 $perfpid || true; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:11:21.049 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@80 -- # waitforio /var/tmp/bdevperf.sock Nvme0n1 00:11:21.049 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@45 -- # '[' -z /var/tmp/bdevperf.sock ']' 00:11:21.049 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@49 -- # '[' -z Nvme0n1 ']' 00:11:21.049 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@52 -- # local ret=1 00:11:21.049 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@53 -- # local i 00:11:21.049 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@54 -- # (( i = 10 )) 00:11:21.049 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@54 -- # (( i != 0 )) 00:11:21.049 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@55 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme0n1 00:11:21.049 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@55 -- # jq -r '.bdevs[0].num_read_ops' 00:11:21.049 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.049 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:11:21.049 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.049 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@55 -- # read_io_count=67 00:11:21.049 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@58 -- # '[' 67 -ge 100 ']' 00:11:21.049 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@62 -- # sleep 0.25 00:11:21.310 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@54 -- # (( i-- )) 00:11:21.310 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@54 -- # (( i != 0 )) 00:11:21.310 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@55 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme0n1 00:11:21.310 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@55 -- # jq -r '.bdevs[0].num_read_ops' 00:11:21.310 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.310 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:11:21.310 17:27:02 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.310 17:27:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@55 -- # read_io_count=515 00:11:21.310 17:27:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@58 -- # '[' 515 -ge 100 ']' 00:11:21.310 17:27:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@59 -- # ret=0 00:11:21.310 17:27:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@60 -- # break 00:11:21.310 17:27:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@64 -- # return 0 00:11:21.310 17:27:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@84 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2016-06.io.spdk:cnode0 nqn.2016-06.io.spdk:host0 00:11:21.310 17:27:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.310 17:27:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:11:21.310 [2024-12-06 17:27:03.010822] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:81792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.310 [2024-12-06 17:27:03.010880] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.310 [2024-12-06 17:27:03.010918] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:73728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.310 [2024-12-06 17:27:03.010935] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.310 [2024-12-06 17:27:03.010951] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:73856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.310 [2024-12-06 17:27:03.010977] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.310 [2024-12-06 17:27:03.010993] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:73984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.310 [2024-12-06 17:27:03.011007] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.310 [2024-12-06 17:27:03.011023] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:74112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.310 [2024-12-06 17:27:03.011046] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.310 [2024-12-06 17:27:03.011062] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:74240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.310 [2024-12-06 17:27:03.011076] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.011091] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:74368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.011116] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.011132] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:74496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.011147] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.011163] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:74624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.011177] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.011193] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:74752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.011207] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.011222] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:74880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.011243] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.011260] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:75008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.011274] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.011304] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:75136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.011319] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.011334] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:75264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.011348] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.011363] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:75392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.011376] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.011391] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:75520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.011405] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.011420] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:75648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.011433] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.011448] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:75776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.011463] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.011478] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:75904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.011492] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.011512] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:76032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.011527] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.011543] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:76160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.011556] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.011571] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:76288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.011585] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.011599] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:76416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.011613] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.011628] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:76544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.011641] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.011681] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:76672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.011696] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.011717] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:76800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.011731] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.011746] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:76928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.011765] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.011781] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:77056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.011795] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.011810] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:77184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.011823] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.011839] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:77312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.011853] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.011868] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:77440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.011882] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.011897] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:77568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.011914] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.011930] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:77696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.011944] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.011969] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:77824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.011998] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.012014] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:77952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.012032] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.012047] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:78080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.012062] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.012076] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:78208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.012090] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.012105] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:78336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.012118] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.012133] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:78464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.012147] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.012162] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:78592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.012175] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.012191] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:78720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.012204] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.012220] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:78848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.012233] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.012248] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:78976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.012266] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.012282] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:79104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.012297] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.012315] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:79232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.311 [2024-12-06 17:27:03.012329] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.311 [2024-12-06 17:27:03.012344] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:79360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.312 [2024-12-06 17:27:03.012374] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.312 [2024-12-06 17:27:03.012390] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:79488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.312 [2024-12-06 17:27:03.012404] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.312 [2024-12-06 17:27:03.012420] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:79616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.312 [2024-12-06 17:27:03.012434] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.312 [2024-12-06 17:27:03.012449] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:79744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.312 [2024-12-06 17:27:03.012462] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.312 [2024-12-06 17:27:03.012478] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:79872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.312 [2024-12-06 17:27:03.012492] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.312 [2024-12-06 17:27:03.012507] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:80000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.312 [2024-12-06 17:27:03.012520] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.312 [2024-12-06 17:27:03.012536] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:80128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.312 [2024-12-06 17:27:03.012550] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.312 [2024-12-06 17:27:03.012565] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:80256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.312 [2024-12-06 17:27:03.012580] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.312 [2024-12-06 17:27:03.012594] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:80384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.312 [2024-12-06 17:27:03.012609] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.312 [2024-12-06 17:27:03.012624] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:80512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.312 [2024-12-06 17:27:03.012638] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.312 [2024-12-06 17:27:03.012654] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:80640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.312 [2024-12-06 17:27:03.012676] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.312 [2024-12-06 17:27:03.012695] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:80768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.312 [2024-12-06 17:27:03.012720] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.312 [2024-12-06 17:27:03.012737] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:80896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.312 [2024-12-06 17:27:03.012751] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.312 [2024-12-06 17:27:03.012766] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:81024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.312 [2024-12-06 17:27:03.012784] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.312 [2024-12-06 17:27:03.012800] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:81152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.312 [2024-12-06 17:27:03.012814] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.312 [2024-12-06 17:27:03.012829] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:81280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.312 [2024-12-06 17:27:03.012843] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.312 [2024-12-06 17:27:03.012858] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:81408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.312 [2024-12-06 17:27:03.012872] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.312 [2024-12-06 17:27:03.012888] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:81536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.312 [2024-12-06 17:27:03.012902] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.312 [2024-12-06 17:27:03.012917] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:81664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:11:21.312 [2024-12-06 17:27:03.012931] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.312 [2024-12-06 17:27:03.012945] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x252a920 is same with the state(6) to be set 00:11:21.312 17:27:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.312 [2024-12-06 17:27:03.013098] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:11:21.312 [2024-12-06 17:27:03.013121] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.312 [2024-12-06 17:27:03.013136] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:11:21.312 [2024-12-06 17:27:03.013150] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.312 [2024-12-06 17:27:03.013164] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 ns 17:27:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@85 -- # rpc_cmd nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode0 nqn.2016-06.io.spdk:host0 00:11:21.312 id:0 cdw10:00000000 cdw11:00000000 00:11:21.312 [2024-12-06 17:27:03.013180] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.312 [2024-12-06 17:27:03.013194] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:11:21.312 [2024-12-06 17:27:03.013207] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:21.312 [2024-12-06 17:27:03.013224] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2311980 is same with the state(6) to be set 00:11:21.312 17:27:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@563 -- # xtrace_disable 00:11:21.312 17:27:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:11:21.312 [2024-12-06 17:27:03.014384] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:11:21.312 task offset: 81792 on job bdev=Nvme0n1 fails 00:11:21.312 00:11:21.312 Latency(us) 00:11:21.312 [2024-12-06T16:27:03.151Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:21.312 Job: Nvme0n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:11:21.312 Job: Nvme0n1 ended in about 0.41 seconds with error 00:11:21.312 Verification LBA range: start 0x0 length 0x400 00:11:21.312 Nvme0n1 : 0.41 1409.11 88.07 156.57 0.00 39722.59 7233.23 39807.05 00:11:21.312 [2024-12-06T16:27:03.151Z] =================================================================================================================== 00:11:21.312 [2024-12-06T16:27:03.151Z] Total : 1409.11 88.07 156.57 0.00 39722.59 7233.23 39807.05 00:11:21.312 [2024-12-06 17:27:03.016349] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:11:21.312 [2024-12-06 17:27:03.016378] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x2311980 (9): Bad file descriptor 00:11:21.312 17:27:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:11:21.312 17:27:03 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@87 -- # sleep 1 00:11:21.572 [2024-12-06 17:27:03.148814] bdev_nvme.c:2286:bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 2] Resetting controller successful. 00:11:22.508 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@91 -- # kill -9 136900 00:11:22.508 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh: line 91: kill: (136900) - No such process 00:11:22.508 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@91 -- # true 00:11:22.508 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@97 -- # rm -f /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 /var/tmp/spdk_cpu_lock_003 /var/tmp/spdk_cpu_lock_004 00:11:22.508 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -q 64 -o 65536 -w verify -t 1 00:11:22.508 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@100 -- # gen_nvmf_target_json 0 00:11:22.508 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@560 -- # config=() 00:11:22.508 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@560 -- # local subsystem config 00:11:22.508 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:11:22.508 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:11:22.508 { 00:11:22.508 "params": { 00:11:22.508 "name": "Nvme$subsystem", 00:11:22.508 "trtype": "$TEST_TRANSPORT", 00:11:22.508 "traddr": "$NVMF_FIRST_TARGET_IP", 00:11:22.508 "adrfam": "ipv4", 00:11:22.508 "trsvcid": "$NVMF_PORT", 00:11:22.508 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:11:22.508 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:11:22.508 "hdgst": ${hdgst:-false}, 00:11:22.508 "ddgst": ${ddgst:-false} 00:11:22.508 }, 00:11:22.508 "method": "bdev_nvme_attach_controller" 00:11:22.508 } 00:11:22.508 EOF 00:11:22.508 )") 00:11:22.508 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@582 -- # cat 00:11:22.508 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@584 -- # jq . 00:11:22.508 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@585 -- # IFS=, 00:11:22.508 17:27:04 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:11:22.508 "params": { 00:11:22.508 "name": "Nvme0", 00:11:22.508 "trtype": "tcp", 00:11:22.508 "traddr": "10.0.0.2", 00:11:22.508 "adrfam": "ipv4", 00:11:22.508 "trsvcid": "4420", 00:11:22.508 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:11:22.508 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:11:22.508 "hdgst": false, 00:11:22.508 "ddgst": false 00:11:22.508 }, 00:11:22.508 "method": "bdev_nvme_attach_controller" 00:11:22.508 }' 00:11:22.508 [2024-12-06 17:27:04.072218] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:11:22.508 [2024-12-06 17:27:04.072294] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid137181 ] 00:11:22.508 [2024-12-06 17:27:04.143387] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:22.508 [2024-12-06 17:27:04.192037] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:22.767 Running I/O for 1 seconds... 00:11:23.704 1625.00 IOPS, 101.56 MiB/s 00:11:23.704 Latency(us) 00:11:23.704 [2024-12-06T16:27:05.543Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:23.704 Job: Nvme0n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:11:23.704 Verification LBA range: start 0x0 length 0x400 00:11:23.704 Nvme0n1 : 1.01 1664.81 104.05 0.00 0.00 37652.28 2051.03 33593.27 00:11:23.704 [2024-12-06T16:27:05.543Z] =================================================================================================================== 00:11:23.704 [2024-12-06T16:27:05.543Z] Total : 1664.81 104.05 0.00 0.00 37652.28 2051.03 33593.27 00:11:23.962 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@102 -- # stoptarget 00:11:23.962 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@36 -- # rm -f ./local-job0-0-verify.state 00:11:23.962 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@37 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:11:23.962 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@38 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:11:23.962 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@40 -- # nvmftestfini 00:11:23.962 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@516 -- # nvmfcleanup 00:11:23.962 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@121 -- # sync 00:11:23.962 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:11:23.962 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@124 -- # set +e 00:11:23.962 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@125 -- # for i in {1..20} 00:11:23.962 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:11:23.962 rmmod nvme_tcp 00:11:23.962 rmmod nvme_fabrics 00:11:23.962 rmmod nvme_keyring 00:11:23.962 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:11:23.962 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@128 -- # set -e 00:11:23.962 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@129 -- # return 0 00:11:23.962 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@517 -- # '[' -n 136843 ']' 00:11:23.963 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@518 -- # killprocess 136843 00:11:23.963 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@954 -- # '[' -z 136843 ']' 00:11:23.963 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@958 -- # kill -0 136843 00:11:23.963 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@959 -- # uname 00:11:23.963 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:23.963 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 136843 00:11:23.963 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:11:23.963 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:11:23.963 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@972 -- # echo 'killing process with pid 136843' 00:11:23.963 killing process with pid 136843 00:11:23.963 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@973 -- # kill 136843 00:11:23.963 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@978 -- # wait 136843 00:11:24.223 [2024-12-06 17:27:05.941839] app.c: 721:unclaim_cpu_cores: *ERROR*: Failed to unlink lock fd for core 1, errno: 2 00:11:24.223 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:11:24.223 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:11:24.223 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:11:24.223 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@297 -- # iptr 00:11:24.223 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@791 -- # iptables-save 00:11:24.223 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:11:24.223 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@791 -- # iptables-restore 00:11:24.223 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:11:24.223 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@302 -- # remove_spdk_ns 00:11:24.223 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:24.223 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:11:24.223 17:27:05 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:26.766 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:11:26.766 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@109 -- # trap - SIGINT SIGTERM EXIT 00:11:26.766 00:11:26.766 real 0m8.673s 00:11:26.766 user 0m19.176s 00:11:26.766 sys 0m2.711s 00:11:26.766 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:26.766 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:11:26.766 ************************************ 00:11:26.766 END TEST nvmf_host_management 00:11:26.766 ************************************ 00:11:26.766 17:27:08 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@27 -- # run_test nvmf_lvol /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvol.sh --transport=tcp 00:11:26.766 17:27:08 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:11:26.766 17:27:08 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:26.766 17:27:08 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:11:26.767 ************************************ 00:11:26.767 START TEST nvmf_lvol 00:11:26.767 ************************************ 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvol.sh --transport=tcp 00:11:26.767 * Looking for test storage... 00:11:26.767 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1711 -- # lcov --version 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@333 -- # local ver1 ver1_l 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@334 -- # local ver2 ver2_l 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@336 -- # IFS=.-: 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@336 -- # read -ra ver1 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@337 -- # IFS=.-: 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@337 -- # read -ra ver2 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@338 -- # local 'op=<' 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@340 -- # ver1_l=2 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@341 -- # ver2_l=1 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@344 -- # case "$op" in 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@345 -- # : 1 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@364 -- # (( v = 0 )) 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@365 -- # decimal 1 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@353 -- # local d=1 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@355 -- # echo 1 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@365 -- # ver1[v]=1 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@366 -- # decimal 2 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@353 -- # local d=2 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@355 -- # echo 2 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@366 -- # ver2[v]=2 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@368 -- # return 0 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:11:26.767 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:26.767 --rc genhtml_branch_coverage=1 00:11:26.767 --rc genhtml_function_coverage=1 00:11:26.767 --rc genhtml_legend=1 00:11:26.767 --rc geninfo_all_blocks=1 00:11:26.767 --rc geninfo_unexecuted_blocks=1 00:11:26.767 00:11:26.767 ' 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:11:26.767 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:26.767 --rc genhtml_branch_coverage=1 00:11:26.767 --rc genhtml_function_coverage=1 00:11:26.767 --rc genhtml_legend=1 00:11:26.767 --rc geninfo_all_blocks=1 00:11:26.767 --rc geninfo_unexecuted_blocks=1 00:11:26.767 00:11:26.767 ' 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:11:26.767 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:26.767 --rc genhtml_branch_coverage=1 00:11:26.767 --rc genhtml_function_coverage=1 00:11:26.767 --rc genhtml_legend=1 00:11:26.767 --rc geninfo_all_blocks=1 00:11:26.767 --rc geninfo_unexecuted_blocks=1 00:11:26.767 00:11:26.767 ' 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:11:26.767 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:26.767 --rc genhtml_branch_coverage=1 00:11:26.767 --rc genhtml_function_coverage=1 00:11:26.767 --rc genhtml_legend=1 00:11:26.767 --rc geninfo_all_blocks=1 00:11:26.767 --rc geninfo_unexecuted_blocks=1 00:11:26.767 00:11:26.767 ' 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@7 -- # uname -s 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@15 -- # shopt -s extglob 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- paths/export.sh@5 -- # export PATH 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@51 -- # : 0 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:11:26.767 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:11:26.767 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@55 -- # have_pci_nics=0 00:11:26.768 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@11 -- # MALLOC_BDEV_SIZE=64 00:11:26.768 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:11:26.768 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@13 -- # LVOL_BDEV_INIT_SIZE=20 00:11:26.768 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@14 -- # LVOL_BDEV_FINAL_SIZE=30 00:11:26.768 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@16 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:11:26.768 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@18 -- # nvmftestinit 00:11:26.768 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:11:26.768 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:11:26.768 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@476 -- # prepare_net_devs 00:11:26.768 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@438 -- # local -g is_hw=no 00:11:26.768 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@440 -- # remove_spdk_ns 00:11:26.768 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:26.768 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:11:26.768 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:26.768 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:11:26.768 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:11:26.768 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@309 -- # xtrace_disable 00:11:26.768 17:27:08 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@315 -- # pci_devs=() 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@315 -- # local -a pci_devs 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@316 -- # pci_net_devs=() 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@317 -- # pci_drivers=() 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@317 -- # local -A pci_drivers 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@319 -- # net_devs=() 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@319 -- # local -ga net_devs 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@320 -- # e810=() 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@320 -- # local -ga e810 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@321 -- # x722=() 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@321 -- # local -ga x722 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@322 -- # mlx=() 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@322 -- # local -ga mlx 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:11:28.675 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:11:28.675 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@418 -- # [[ up == up ]] 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:11:28.675 Found net devices under 0000:0a:00.0: cvl_0_0 00:11:28.675 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@418 -- # [[ up == up ]] 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:11:28.676 Found net devices under 0000:0a:00.1: cvl_0_1 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@442 -- # is_hw=yes 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:11:28.676 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:11:28.676 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.155 ms 00:11:28.676 00:11:28.676 --- 10.0.0.2 ping statistics --- 00:11:28.676 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:28.676 rtt min/avg/max/mdev = 0.155/0.155/0.155/0.000 ms 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:11:28.676 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:11:28.676 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.044 ms 00:11:28.676 00:11:28.676 --- 10.0.0.1 ping statistics --- 00:11:28.676 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:28.676 rtt min/avg/max/mdev = 0.044/0.044/0.044/0.000 ms 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@450 -- # return 0 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:11:28.676 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:11:28.936 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@19 -- # nvmfappstart -m 0x7 00:11:28.936 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:11:28.936 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@726 -- # xtrace_disable 00:11:28.936 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:11:28.936 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@509 -- # nvmfpid=139782 00:11:28.936 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x7 00:11:28.936 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@510 -- # waitforlisten 139782 00:11:28.936 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@835 -- # '[' -z 139782 ']' 00:11:28.936 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:28.936 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:28.936 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:28.936 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:28.936 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:28.936 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:11:28.936 [2024-12-06 17:27:10.585959] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:11:28.936 [2024-12-06 17:27:10.586050] [ DPDK EAL parameters: nvmf -c 0x7 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:28.936 [2024-12-06 17:27:10.660242] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:11:28.936 [2024-12-06 17:27:10.709965] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:11:28.936 [2024-12-06 17:27:10.710021] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:11:28.937 [2024-12-06 17:27:10.710035] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:11:28.937 [2024-12-06 17:27:10.710047] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:11:28.937 [2024-12-06 17:27:10.710057] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:11:28.937 [2024-12-06 17:27:10.711493] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:11:28.937 [2024-12-06 17:27:10.711566] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:11:28.937 [2024-12-06 17:27:10.711569] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:29.196 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:29.196 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@868 -- # return 0 00:11:29.196 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:11:29.196 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@732 -- # xtrace_disable 00:11:29.196 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:11:29.196 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:11:29.196 17:27:10 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:11:29.475 [2024-12-06 17:27:11.109198] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:11:29.475 17:27:11 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:11:29.734 17:27:11 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@24 -- # base_bdevs='Malloc0 ' 00:11:29.734 17:27:11 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:11:29.993 17:27:11 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@25 -- # base_bdevs+=Malloc1 00:11:29.993 17:27:11 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n raid0 -z 64 -r 0 -b 'Malloc0 Malloc1' 00:11:30.251 17:27:11 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore raid0 lvs 00:11:30.510 17:27:12 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@29 -- # lvs=9700fe89-f4a5-4fff-a7f4-39e4cdf882b6 00:11:30.510 17:27:12 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u 9700fe89-f4a5-4fff-a7f4-39e4cdf882b6 lvol 20 00:11:30.769 17:27:12 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@32 -- # lvol=7fd3416e-a387-4757-ae0e-edd35602416b 00:11:30.769 17:27:12 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:11:31.028 17:27:12 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 7fd3416e-a387-4757-ae0e-edd35602416b 00:11:31.286 17:27:13 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:11:31.545 [2024-12-06 17:27:13.353544] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:11:31.545 17:27:13 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:11:31.803 17:27:13 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@42 -- # perf_pid=140208 00:11:31.803 17:27:13 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@44 -- # sleep 1 00:11:31.803 17:27:13 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -o 4096 -q 128 -s 512 -w randwrite -t 10 -c 0x18 00:11:33.179 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_snapshot 7fd3416e-a387-4757-ae0e-edd35602416b MY_SNAPSHOT 00:11:33.179 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@47 -- # snapshot=7d55c899-2fe2-427f-ad52-9c5f32e6b35b 00:11:33.179 17:27:14 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_resize 7fd3416e-a387-4757-ae0e-edd35602416b 30 00:11:33.440 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@49 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_clone 7d55c899-2fe2-427f-ad52-9c5f32e6b35b MY_CLONE 00:11:34.006 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@49 -- # clone=dc2f939e-07a4-4f67-9949-c5ef53438949 00:11:34.006 17:27:15 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_inflate dc2f939e-07a4-4f67-9949-c5ef53438949 00:11:34.573 17:27:16 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@53 -- # wait 140208 00:11:42.684 Initializing NVMe Controllers 00:11:42.684 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode0 00:11:42.684 Controller IO queue size 128, less than required. 00:11:42.684 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:11:42.684 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 3 00:11:42.684 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 4 00:11:42.684 Initialization complete. Launching workers. 00:11:42.684 ======================================================== 00:11:42.684 Latency(us) 00:11:42.684 Device Information : IOPS MiB/s Average min max 00:11:42.684 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 from core 3: 10167.60 39.72 12594.07 1892.78 80585.82 00:11:42.684 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 from core 4: 10557.30 41.24 12131.63 2424.61 77798.52 00:11:42.684 ======================================================== 00:11:42.684 Total : 20724.90 80.96 12358.50 1892.78 80585.82 00:11:42.684 00:11:42.684 17:27:24 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:11:42.684 17:27:24 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete 7fd3416e-a387-4757-ae0e-edd35602416b 00:11:42.941 17:27:24 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u 9700fe89-f4a5-4fff-a7f4-39e4cdf882b6 00:11:43.200 17:27:24 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@60 -- # rm -f 00:11:43.200 17:27:24 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@62 -- # trap - SIGINT SIGTERM EXIT 00:11:43.200 17:27:24 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@64 -- # nvmftestfini 00:11:43.200 17:27:24 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@516 -- # nvmfcleanup 00:11:43.200 17:27:24 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@121 -- # sync 00:11:43.200 17:27:24 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:11:43.200 17:27:24 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@124 -- # set +e 00:11:43.200 17:27:24 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@125 -- # for i in {1..20} 00:11:43.200 17:27:24 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:11:43.200 rmmod nvme_tcp 00:11:43.200 rmmod nvme_fabrics 00:11:43.200 rmmod nvme_keyring 00:11:43.200 17:27:25 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:11:43.200 17:27:25 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@128 -- # set -e 00:11:43.200 17:27:25 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@129 -- # return 0 00:11:43.200 17:27:25 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@517 -- # '[' -n 139782 ']' 00:11:43.200 17:27:25 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@518 -- # killprocess 139782 00:11:43.458 17:27:25 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@954 -- # '[' -z 139782 ']' 00:11:43.458 17:27:25 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@958 -- # kill -0 139782 00:11:43.458 17:27:25 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@959 -- # uname 00:11:43.458 17:27:25 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:11:43.458 17:27:25 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 139782 00:11:43.458 17:27:25 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:11:43.458 17:27:25 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:11:43.458 17:27:25 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@972 -- # echo 'killing process with pid 139782' 00:11:43.458 killing process with pid 139782 00:11:43.458 17:27:25 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@973 -- # kill 139782 00:11:43.458 17:27:25 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@978 -- # wait 139782 00:11:43.718 17:27:25 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:11:43.718 17:27:25 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:11:43.718 17:27:25 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:11:43.718 17:27:25 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@297 -- # iptr 00:11:43.718 17:27:25 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@791 -- # iptables-save 00:11:43.718 17:27:25 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@791 -- # iptables-restore 00:11:43.718 17:27:25 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:11:43.718 17:27:25 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:11:43.718 17:27:25 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@302 -- # remove_spdk_ns 00:11:43.718 17:27:25 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:43.718 17:27:25 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:11:43.718 17:27:25 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:45.630 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:11:45.630 00:11:45.630 real 0m19.314s 00:11:45.630 user 1m5.106s 00:11:45.630 sys 0m5.898s 00:11:45.630 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1130 -- # xtrace_disable 00:11:45.630 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:11:45.630 ************************************ 00:11:45.630 END TEST nvmf_lvol 00:11:45.630 ************************************ 00:11:45.630 17:27:27 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@28 -- # run_test nvmf_lvs_grow /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh --transport=tcp 00:11:45.630 17:27:27 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:11:45.630 17:27:27 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:45.630 17:27:27 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:11:45.630 ************************************ 00:11:45.630 START TEST nvmf_lvs_grow 00:11:45.630 ************************************ 00:11:45.630 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh --transport=tcp 00:11:45.891 * Looking for test storage... 00:11:45.891 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1711 -- # lcov --version 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@333 -- # local ver1 ver1_l 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@334 -- # local ver2 ver2_l 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@336 -- # IFS=.-: 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@336 -- # read -ra ver1 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@337 -- # IFS=.-: 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@337 -- # read -ra ver2 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@338 -- # local 'op=<' 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@340 -- # ver1_l=2 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@341 -- # ver2_l=1 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@344 -- # case "$op" in 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@345 -- # : 1 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@364 -- # (( v = 0 )) 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@365 -- # decimal 1 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@353 -- # local d=1 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@355 -- # echo 1 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@365 -- # ver1[v]=1 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@366 -- # decimal 2 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@353 -- # local d=2 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@355 -- # echo 2 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@366 -- # ver2[v]=2 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@368 -- # return 0 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:11:45.891 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:45.891 --rc genhtml_branch_coverage=1 00:11:45.891 --rc genhtml_function_coverage=1 00:11:45.891 --rc genhtml_legend=1 00:11:45.891 --rc geninfo_all_blocks=1 00:11:45.891 --rc geninfo_unexecuted_blocks=1 00:11:45.891 00:11:45.891 ' 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:11:45.891 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:45.891 --rc genhtml_branch_coverage=1 00:11:45.891 --rc genhtml_function_coverage=1 00:11:45.891 --rc genhtml_legend=1 00:11:45.891 --rc geninfo_all_blocks=1 00:11:45.891 --rc geninfo_unexecuted_blocks=1 00:11:45.891 00:11:45.891 ' 00:11:45.891 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:11:45.891 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:45.891 --rc genhtml_branch_coverage=1 00:11:45.891 --rc genhtml_function_coverage=1 00:11:45.891 --rc genhtml_legend=1 00:11:45.891 --rc geninfo_all_blocks=1 00:11:45.892 --rc geninfo_unexecuted_blocks=1 00:11:45.892 00:11:45.892 ' 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:11:45.892 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:45.892 --rc genhtml_branch_coverage=1 00:11:45.892 --rc genhtml_function_coverage=1 00:11:45.892 --rc genhtml_legend=1 00:11:45.892 --rc geninfo_all_blocks=1 00:11:45.892 --rc geninfo_unexecuted_blocks=1 00:11:45.892 00:11:45.892 ' 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@7 -- # uname -s 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@15 -- # shopt -s extglob 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- paths/export.sh@5 -- # export PATH 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@51 -- # : 0 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:11:45.892 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@55 -- # have_pci_nics=0 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@11 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@12 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@98 -- # nvmftestinit 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@476 -- # prepare_net_devs 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@438 -- # local -g is_hw=no 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@440 -- # remove_spdk_ns 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@309 -- # xtrace_disable 00:11:45.892 17:27:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:11:48.430 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:11:48.430 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@315 -- # pci_devs=() 00:11:48.430 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@315 -- # local -a pci_devs 00:11:48.430 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@316 -- # pci_net_devs=() 00:11:48.430 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:11:48.430 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@317 -- # pci_drivers=() 00:11:48.430 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@317 -- # local -A pci_drivers 00:11:48.430 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@319 -- # net_devs=() 00:11:48.430 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@319 -- # local -ga net_devs 00:11:48.430 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@320 -- # e810=() 00:11:48.430 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@320 -- # local -ga e810 00:11:48.430 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@321 -- # x722=() 00:11:48.430 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@321 -- # local -ga x722 00:11:48.430 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@322 -- # mlx=() 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@322 -- # local -ga mlx 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:11:48.431 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:11:48.431 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@418 -- # [[ up == up ]] 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:11:48.431 Found net devices under 0000:0a:00.0: cvl_0_0 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@418 -- # [[ up == up ]] 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:11:48.431 Found net devices under 0000:0a:00.1: cvl_0_1 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@442 -- # is_hw=yes 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:11:48.431 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:11:48.432 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:11:48.432 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.243 ms 00:11:48.432 00:11:48.432 --- 10.0.0.2 ping statistics --- 00:11:48.432 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:48.432 rtt min/avg/max/mdev = 0.243/0.243/0.243/0.000 ms 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:11:48.432 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:11:48.432 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.136 ms 00:11:48.432 00:11:48.432 --- 10.0.0.1 ping statistics --- 00:11:48.432 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:48.432 rtt min/avg/max/mdev = 0.136/0.136/0.136/0.000 ms 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@450 -- # return 0 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@99 -- # nvmfappstart -m 0x1 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@726 -- # xtrace_disable 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@509 -- # nvmfpid=143501 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@510 -- # waitforlisten 143501 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@835 -- # '[' -z 143501 ']' 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:48.432 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:48.432 17:27:29 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:11:48.432 [2024-12-06 17:27:29.951983] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:11:48.432 [2024-12-06 17:27:29.952084] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:48.432 [2024-12-06 17:27:30.032158] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:48.432 [2024-12-06 17:27:30.079458] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:11:48.432 [2024-12-06 17:27:30.079529] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:11:48.432 [2024-12-06 17:27:30.079552] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:11:48.432 [2024-12-06 17:27:30.079562] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:11:48.432 [2024-12-06 17:27:30.079571] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:11:48.432 [2024-12-06 17:27:30.080210] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:48.432 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:48.432 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@868 -- # return 0 00:11:48.432 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:11:48.432 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@732 -- # xtrace_disable 00:11:48.432 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:11:48.432 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:11:48.432 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:11:48.692 [2024-12-06 17:27:30.477061] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:11:48.692 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@102 -- # run_test lvs_grow_clean lvs_grow 00:11:48.692 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:11:48.692 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1111 -- # xtrace_disable 00:11:48.692 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:11:48.692 ************************************ 00:11:48.692 START TEST lvs_grow_clean 00:11:48.692 ************************************ 00:11:48.692 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@1129 -- # lvs_grow 00:11:48.692 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@15 -- # local aio_bdev lvs lvol 00:11:48.692 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@16 -- # local data_clusters free_clusters 00:11:48.692 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@17 -- # local bdevperf_pid run_test_pid 00:11:48.692 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@18 -- # local aio_init_size_mb=200 00:11:48.950 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@19 -- # local aio_final_size_mb=400 00:11:48.950 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@20 -- # local lvol_bdev_size_mb=150 00:11:48.950 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@23 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:11:48.950 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@24 -- # truncate -s 200M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:11:48.950 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:11:49.209 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@25 -- # aio_bdev=aio_bdev 00:11:49.209 17:27:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --cluster-sz 4194304 --md-pages-per-cluster-ratio 300 aio_bdev lvs 00:11:49.469 17:27:31 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@28 -- # lvs=abd4b4c0-140f-48dd-95cf-65f1c56a12a2 00:11:49.469 17:27:31 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u abd4b4c0-140f-48dd-95cf-65f1c56a12a2 00:11:49.469 17:27:31 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@29 -- # jq -r '.[0].total_data_clusters' 00:11:49.732 17:27:31 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@29 -- # data_clusters=49 00:11:49.732 17:27:31 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@30 -- # (( data_clusters == 49 )) 00:11:49.732 17:27:31 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u abd4b4c0-140f-48dd-95cf-65f1c56a12a2 lvol 150 00:11:49.991 17:27:31 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@33 -- # lvol=f6a3201b-9dd4-4596-ad81-fa500ad4e7c7 00:11:49.991 17:27:31 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@36 -- # truncate -s 400M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:11:49.991 17:27:31 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_rescan aio_bdev 00:11:50.249 [2024-12-06 17:27:31.904101] bdev_aio.c:1053:bdev_aio_rescan: *NOTICE*: AIO device is resized: bdev name /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev, old block count 51200, new block count 102400 00:11:50.249 [2024-12-06 17:27:31.904185] vbdev_lvol.c: 165:vbdev_lvs_base_bdev_event_cb: *NOTICE*: Unsupported bdev event: type 1 00:11:50.249 true 00:11:50.249 17:27:31 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u abd4b4c0-140f-48dd-95cf-65f1c56a12a2 00:11:50.249 17:27:31 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@38 -- # jq -r '.[0].total_data_clusters' 00:11:50.507 17:27:32 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@38 -- # (( data_clusters == 49 )) 00:11:50.507 17:27:32 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:11:50.786 17:27:32 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 f6a3201b-9dd4-4596-ad81-fa500ad4e7c7 00:11:51.044 17:27:32 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:11:51.303 [2024-12-06 17:27:32.979272] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:11:51.303 17:27:32 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:11:51.563 17:27:33 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@48 -- # bdevperf_pid=143933 00:11:51.563 17:27:33 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock -m 0x2 -o 4096 -q 128 -w randwrite -t 10 -S 1 -z 00:11:51.563 17:27:33 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@49 -- # trap 'killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:11:51.563 17:27:33 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@50 -- # waitforlisten 143933 /var/tmp/bdevperf.sock 00:11:51.563 17:27:33 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@835 -- # '[' -z 143933 ']' 00:11:51.563 17:27:33 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:11:51.563 17:27:33 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@840 -- # local max_retries=100 00:11:51.563 17:27:33 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:11:51.563 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:11:51.563 17:27:33 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@844 -- # xtrace_disable 00:11:51.563 17:27:33 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@10 -- # set +x 00:11:51.563 [2024-12-06 17:27:33.319778] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:11:51.563 [2024-12-06 17:27:33.319863] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid143933 ] 00:11:51.563 [2024-12-06 17:27:33.391096] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:51.822 [2024-12-06 17:27:33.437606] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:11:51.822 17:27:33 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:11:51.822 17:27:33 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@868 -- # return 0 00:11:51.822 17:27:33 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 00:11:52.081 Nvme0n1 00:11:52.081 17:27:33 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_get_bdevs -b Nvme0n1 -t 3000 00:11:52.340 [ 00:11:52.340 { 00:11:52.340 "name": "Nvme0n1", 00:11:52.340 "aliases": [ 00:11:52.340 "f6a3201b-9dd4-4596-ad81-fa500ad4e7c7" 00:11:52.340 ], 00:11:52.340 "product_name": "NVMe disk", 00:11:52.340 "block_size": 4096, 00:11:52.340 "num_blocks": 38912, 00:11:52.340 "uuid": "f6a3201b-9dd4-4596-ad81-fa500ad4e7c7", 00:11:52.340 "numa_id": 0, 00:11:52.340 "assigned_rate_limits": { 00:11:52.340 "rw_ios_per_sec": 0, 00:11:52.340 "rw_mbytes_per_sec": 0, 00:11:52.340 "r_mbytes_per_sec": 0, 00:11:52.340 "w_mbytes_per_sec": 0 00:11:52.340 }, 00:11:52.340 "claimed": false, 00:11:52.340 "zoned": false, 00:11:52.340 "supported_io_types": { 00:11:52.340 "read": true, 00:11:52.340 "write": true, 00:11:52.340 "unmap": true, 00:11:52.340 "flush": true, 00:11:52.340 "reset": true, 00:11:52.340 "nvme_admin": true, 00:11:52.340 "nvme_io": true, 00:11:52.340 "nvme_io_md": false, 00:11:52.340 "write_zeroes": true, 00:11:52.340 "zcopy": false, 00:11:52.340 "get_zone_info": false, 00:11:52.340 "zone_management": false, 00:11:52.340 "zone_append": false, 00:11:52.340 "compare": true, 00:11:52.340 "compare_and_write": true, 00:11:52.340 "abort": true, 00:11:52.340 "seek_hole": false, 00:11:52.340 "seek_data": false, 00:11:52.340 "copy": true, 00:11:52.340 "nvme_iov_md": false 00:11:52.340 }, 00:11:52.340 "memory_domains": [ 00:11:52.340 { 00:11:52.340 "dma_device_id": "system", 00:11:52.340 "dma_device_type": 1 00:11:52.340 } 00:11:52.340 ], 00:11:52.340 "driver_specific": { 00:11:52.341 "nvme": [ 00:11:52.341 { 00:11:52.341 "trid": { 00:11:52.341 "trtype": "TCP", 00:11:52.341 "adrfam": "IPv4", 00:11:52.341 "traddr": "10.0.0.2", 00:11:52.341 "trsvcid": "4420", 00:11:52.341 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:11:52.341 }, 00:11:52.341 "ctrlr_data": { 00:11:52.341 "cntlid": 1, 00:11:52.341 "vendor_id": "0x8086", 00:11:52.341 "model_number": "SPDK bdev Controller", 00:11:52.341 "serial_number": "SPDK0", 00:11:52.341 "firmware_revision": "25.01", 00:11:52.341 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:11:52.341 "oacs": { 00:11:52.341 "security": 0, 00:11:52.341 "format": 0, 00:11:52.341 "firmware": 0, 00:11:52.341 "ns_manage": 0 00:11:52.341 }, 00:11:52.341 "multi_ctrlr": true, 00:11:52.341 "ana_reporting": false 00:11:52.341 }, 00:11:52.341 "vs": { 00:11:52.341 "nvme_version": "1.3" 00:11:52.341 }, 00:11:52.341 "ns_data": { 00:11:52.341 "id": 1, 00:11:52.341 "can_share": true 00:11:52.341 } 00:11:52.341 } 00:11:52.341 ], 00:11:52.341 "mp_policy": "active_passive" 00:11:52.341 } 00:11:52.341 } 00:11:52.341 ] 00:11:52.600 17:27:34 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@56 -- # run_test_pid=144069 00:11:52.600 17:27:34 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:11:52.600 17:27:34 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@57 -- # sleep 2 00:11:52.600 Running I/O for 10 seconds... 00:11:53.543 Latency(us) 00:11:53.543 [2024-12-06T16:27:35.382Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:53.543 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:53.543 Nvme0n1 : 1.00 15241.00 59.54 0.00 0.00 0.00 0.00 0.00 00:11:53.543 [2024-12-06T16:27:35.382Z] =================================================================================================================== 00:11:53.543 [2024-12-06T16:27:35.382Z] Total : 15241.00 59.54 0.00 0.00 0.00 0.00 0.00 00:11:53.543 00:11:54.478 17:27:36 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_grow_lvstore -u abd4b4c0-140f-48dd-95cf-65f1c56a12a2 00:11:54.478 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:54.478 Nvme0n1 : 2.00 15400.00 60.16 0.00 0.00 0.00 0.00 0.00 00:11:54.478 [2024-12-06T16:27:36.317Z] =================================================================================================================== 00:11:54.478 [2024-12-06T16:27:36.317Z] Total : 15400.00 60.16 0.00 0.00 0.00 0.00 0.00 00:11:54.478 00:11:54.737 true 00:11:54.737 17:27:36 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u abd4b4c0-140f-48dd-95cf-65f1c56a12a2 00:11:54.737 17:27:36 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@61 -- # jq -r '.[0].total_data_clusters' 00:11:54.996 17:27:36 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@61 -- # data_clusters=99 00:11:54.996 17:27:36 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@62 -- # (( data_clusters == 99 )) 00:11:54.996 17:27:36 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@65 -- # wait 144069 00:11:55.563 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:55.563 Nvme0n1 : 3.00 15516.00 60.61 0.00 0.00 0.00 0.00 0.00 00:11:55.563 [2024-12-06T16:27:37.402Z] =================================================================================================================== 00:11:55.563 [2024-12-06T16:27:37.402Z] Total : 15516.00 60.61 0.00 0.00 0.00 0.00 0.00 00:11:55.563 00:11:56.502 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:56.502 Nvme0n1 : 4.00 15574.00 60.84 0.00 0.00 0.00 0.00 0.00 00:11:56.502 [2024-12-06T16:27:38.341Z] =================================================================================================================== 00:11:56.502 [2024-12-06T16:27:38.341Z] Total : 15574.00 60.84 0.00 0.00 0.00 0.00 0.00 00:11:56.502 00:11:57.879 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:57.879 Nvme0n1 : 5.00 15659.60 61.17 0.00 0.00 0.00 0.00 0.00 00:11:57.879 [2024-12-06T16:27:39.718Z] =================================================================================================================== 00:11:57.879 [2024-12-06T16:27:39.718Z] Total : 15659.60 61.17 0.00 0.00 0.00 0.00 0.00 00:11:57.879 00:11:58.814 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:58.814 Nvme0n1 : 6.00 15695.50 61.31 0.00 0.00 0.00 0.00 0.00 00:11:58.814 [2024-12-06T16:27:40.653Z] =================================================================================================================== 00:11:58.814 [2024-12-06T16:27:40.653Z] Total : 15695.50 61.31 0.00 0.00 0.00 0.00 0.00 00:11:58.814 00:11:59.749 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:59.749 Nvme0n1 : 7.00 15739.29 61.48 0.00 0.00 0.00 0.00 0.00 00:11:59.749 [2024-12-06T16:27:41.588Z] =================================================================================================================== 00:11:59.749 [2024-12-06T16:27:41.588Z] Total : 15739.29 61.48 0.00 0.00 0.00 0.00 0.00 00:11:59.749 00:12:00.687 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:12:00.687 Nvme0n1 : 8.00 15790.12 61.68 0.00 0.00 0.00 0.00 0.00 00:12:00.687 [2024-12-06T16:27:42.526Z] =================================================================================================================== 00:12:00.687 [2024-12-06T16:27:42.526Z] Total : 15790.12 61.68 0.00 0.00 0.00 0.00 0.00 00:12:00.687 00:12:01.625 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:12:01.625 Nvme0n1 : 9.00 15835.00 61.86 0.00 0.00 0.00 0.00 0.00 00:12:01.625 [2024-12-06T16:27:43.464Z] =================================================================================================================== 00:12:01.625 [2024-12-06T16:27:43.464Z] Total : 15835.00 61.86 0.00 0.00 0.00 0.00 0.00 00:12:01.625 00:12:02.561 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:12:02.562 Nvme0n1 : 10.00 15864.40 61.97 0.00 0.00 0.00 0.00 0.00 00:12:02.562 [2024-12-06T16:27:44.401Z] =================================================================================================================== 00:12:02.562 [2024-12-06T16:27:44.401Z] Total : 15864.40 61.97 0.00 0.00 0.00 0.00 0.00 00:12:02.562 00:12:02.562 00:12:02.562 Latency(us) 00:12:02.562 [2024-12-06T16:27:44.401Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:02.562 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:12:02.562 Nvme0n1 : 10.01 15867.54 61.98 0.00 0.00 8061.82 2585.03 15728.64 00:12:02.562 [2024-12-06T16:27:44.401Z] =================================================================================================================== 00:12:02.562 [2024-12-06T16:27:44.401Z] Total : 15867.54 61.98 0.00 0.00 8061.82 2585.03 15728.64 00:12:02.562 { 00:12:02.562 "results": [ 00:12:02.562 { 00:12:02.562 "job": "Nvme0n1", 00:12:02.562 "core_mask": "0x2", 00:12:02.562 "workload": "randwrite", 00:12:02.562 "status": "finished", 00:12:02.562 "queue_depth": 128, 00:12:02.562 "io_size": 4096, 00:12:02.562 "runtime": 10.006087, 00:12:02.562 "iops": 15867.54142753306, 00:12:02.562 "mibps": 61.98258370130102, 00:12:02.562 "io_failed": 0, 00:12:02.562 "io_timeout": 0, 00:12:02.562 "avg_latency_us": 8061.816146778377, 00:12:02.562 "min_latency_us": 2585.031111111111, 00:12:02.562 "max_latency_us": 15728.64 00:12:02.562 } 00:12:02.562 ], 00:12:02.562 "core_count": 1 00:12:02.562 } 00:12:02.562 17:27:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@66 -- # killprocess 143933 00:12:02.562 17:27:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@954 -- # '[' -z 143933 ']' 00:12:02.562 17:27:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@958 -- # kill -0 143933 00:12:02.562 17:27:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@959 -- # uname 00:12:02.562 17:27:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:02.562 17:27:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 143933 00:12:02.562 17:27:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:12:02.562 17:27:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:12:02.562 17:27:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@972 -- # echo 'killing process with pid 143933' 00:12:02.562 killing process with pid 143933 00:12:02.562 17:27:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@973 -- # kill 143933 00:12:02.562 Received shutdown signal, test time was about 10.000000 seconds 00:12:02.562 00:12:02.562 Latency(us) 00:12:02.562 [2024-12-06T16:27:44.401Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:02.562 [2024-12-06T16:27:44.401Z] =================================================================================================================== 00:12:02.562 [2024-12-06T16:27:44.401Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:12:02.562 17:27:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@978 -- # wait 143933 00:12:02.821 17:27:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@68 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:12:03.079 17:27:44 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@69 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:12:03.338 17:27:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@70 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u abd4b4c0-140f-48dd-95cf-65f1c56a12a2 00:12:03.338 17:27:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@70 -- # jq -r '.[0].free_clusters' 00:12:03.595 17:27:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@70 -- # free_clusters=61 00:12:03.595 17:27:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@72 -- # [[ '' == \d\i\r\t\y ]] 00:12:03.595 17:27:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:12:03.853 [2024-12-06 17:27:45.611515] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev aio_bdev being removed: closing lvstore lvs 00:12:03.853 17:27:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@85 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u abd4b4c0-140f-48dd-95cf-65f1c56a12a2 00:12:03.853 17:27:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@652 -- # local es=0 00:12:03.853 17:27:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@654 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u abd4b4c0-140f-48dd-95cf-65f1c56a12a2 00:12:03.853 17:27:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@640 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:12:03.853 17:27:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:03.853 17:27:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@644 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:12:03.853 17:27:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:03.853 17:27:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@646 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:12:03.853 17:27:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:03.853 17:27:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@646 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:12:03.853 17:27:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@646 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:12:03.853 17:27:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@655 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u abd4b4c0-140f-48dd-95cf-65f1c56a12a2 00:12:04.111 request: 00:12:04.111 { 00:12:04.111 "uuid": "abd4b4c0-140f-48dd-95cf-65f1c56a12a2", 00:12:04.111 "method": "bdev_lvol_get_lvstores", 00:12:04.111 "req_id": 1 00:12:04.111 } 00:12:04.111 Got JSON-RPC error response 00:12:04.111 response: 00:12:04.111 { 00:12:04.111 "code": -19, 00:12:04.111 "message": "No such device" 00:12:04.111 } 00:12:04.111 17:27:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@655 -- # es=1 00:12:04.111 17:27:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:12:04.111 17:27:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:12:04.111 17:27:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:12:04.111 17:27:45 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:12:04.368 aio_bdev 00:12:04.368 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@87 -- # waitforbdev f6a3201b-9dd4-4596-ad81-fa500ad4e7c7 00:12:04.368 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@903 -- # local bdev_name=f6a3201b-9dd4-4596-ad81-fa500ad4e7c7 00:12:04.368 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:04.368 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@905 -- # local i 00:12:04.368 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:04.368 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:04.368 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@908 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:12:04.626 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@910 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b f6a3201b-9dd4-4596-ad81-fa500ad4e7c7 -t 2000 00:12:04.884 [ 00:12:04.884 { 00:12:04.884 "name": "f6a3201b-9dd4-4596-ad81-fa500ad4e7c7", 00:12:04.884 "aliases": [ 00:12:04.884 "lvs/lvol" 00:12:04.884 ], 00:12:04.884 "product_name": "Logical Volume", 00:12:04.884 "block_size": 4096, 00:12:04.884 "num_blocks": 38912, 00:12:04.884 "uuid": "f6a3201b-9dd4-4596-ad81-fa500ad4e7c7", 00:12:04.884 "assigned_rate_limits": { 00:12:04.884 "rw_ios_per_sec": 0, 00:12:04.884 "rw_mbytes_per_sec": 0, 00:12:04.884 "r_mbytes_per_sec": 0, 00:12:04.884 "w_mbytes_per_sec": 0 00:12:04.884 }, 00:12:04.884 "claimed": false, 00:12:04.884 "zoned": false, 00:12:04.884 "supported_io_types": { 00:12:04.884 "read": true, 00:12:04.884 "write": true, 00:12:04.884 "unmap": true, 00:12:04.884 "flush": false, 00:12:04.884 "reset": true, 00:12:04.884 "nvme_admin": false, 00:12:04.884 "nvme_io": false, 00:12:04.884 "nvme_io_md": false, 00:12:04.884 "write_zeroes": true, 00:12:04.884 "zcopy": false, 00:12:04.884 "get_zone_info": false, 00:12:04.884 "zone_management": false, 00:12:04.884 "zone_append": false, 00:12:04.884 "compare": false, 00:12:04.884 "compare_and_write": false, 00:12:04.884 "abort": false, 00:12:04.884 "seek_hole": true, 00:12:04.884 "seek_data": true, 00:12:04.884 "copy": false, 00:12:04.884 "nvme_iov_md": false 00:12:04.884 }, 00:12:04.884 "driver_specific": { 00:12:04.884 "lvol": { 00:12:04.884 "lvol_store_uuid": "abd4b4c0-140f-48dd-95cf-65f1c56a12a2", 00:12:04.884 "base_bdev": "aio_bdev", 00:12:04.884 "thin_provision": false, 00:12:04.884 "num_allocated_clusters": 38, 00:12:04.884 "snapshot": false, 00:12:04.884 "clone": false, 00:12:04.884 "esnap_clone": false 00:12:04.884 } 00:12:04.884 } 00:12:04.884 } 00:12:04.884 ] 00:12:04.884 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@911 -- # return 0 00:12:05.142 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u abd4b4c0-140f-48dd-95cf-65f1c56a12a2 00:12:05.142 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@88 -- # jq -r '.[0].free_clusters' 00:12:05.401 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@88 -- # (( free_clusters == 61 )) 00:12:05.401 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u abd4b4c0-140f-48dd-95cf-65f1c56a12a2 00:12:05.401 17:27:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@89 -- # jq -r '.[0].total_data_clusters' 00:12:05.660 17:27:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@89 -- # (( data_clusters == 99 )) 00:12:05.660 17:27:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@92 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete f6a3201b-9dd4-4596-ad81-fa500ad4e7c7 00:12:05.919 17:27:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@93 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u abd4b4c0-140f-48dd-95cf-65f1c56a12a2 00:12:06.178 17:27:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@94 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:12:06.438 17:27:48 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@95 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:12:06.438 00:12:06.438 real 0m17.604s 00:12:06.438 user 0m17.088s 00:12:06.438 sys 0m1.904s 00:12:06.438 17:27:48 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:06.438 17:27:48 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@10 -- # set +x 00:12:06.438 ************************************ 00:12:06.438 END TEST lvs_grow_clean 00:12:06.438 ************************************ 00:12:06.438 17:27:48 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@103 -- # run_test lvs_grow_dirty lvs_grow dirty 00:12:06.438 17:27:48 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:12:06.438 17:27:48 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:06.438 17:27:48 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:12:06.438 ************************************ 00:12:06.438 START TEST lvs_grow_dirty 00:12:06.438 ************************************ 00:12:06.438 17:27:48 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@1129 -- # lvs_grow dirty 00:12:06.438 17:27:48 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@15 -- # local aio_bdev lvs lvol 00:12:06.438 17:27:48 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@16 -- # local data_clusters free_clusters 00:12:06.438 17:27:48 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@17 -- # local bdevperf_pid run_test_pid 00:12:06.438 17:27:48 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@18 -- # local aio_init_size_mb=200 00:12:06.438 17:27:48 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@19 -- # local aio_final_size_mb=400 00:12:06.438 17:27:48 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@20 -- # local lvol_bdev_size_mb=150 00:12:06.438 17:27:48 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@23 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:12:06.438 17:27:48 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@24 -- # truncate -s 200M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:12:06.438 17:27:48 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:12:06.696 17:27:48 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@25 -- # aio_bdev=aio_bdev 00:12:06.697 17:27:48 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --cluster-sz 4194304 --md-pages-per-cluster-ratio 300 aio_bdev lvs 00:12:06.955 17:27:48 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@28 -- # lvs=4edc6b8b-03f2-4934-b920-115813401976 00:12:06.955 17:27:48 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 4edc6b8b-03f2-4934-b920-115813401976 00:12:06.955 17:27:48 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@29 -- # jq -r '.[0].total_data_clusters' 00:12:07.214 17:27:48 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@29 -- # data_clusters=49 00:12:07.214 17:27:48 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@30 -- # (( data_clusters == 49 )) 00:12:07.214 17:27:48 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u 4edc6b8b-03f2-4934-b920-115813401976 lvol 150 00:12:07.473 17:27:49 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@33 -- # lvol=bcc5cd7d-1608-4625-8611-8cb86edc7b75 00:12:07.473 17:27:49 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@36 -- # truncate -s 400M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:12:07.473 17:27:49 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_rescan aio_bdev 00:12:07.732 [2024-12-06 17:27:49.526024] bdev_aio.c:1053:bdev_aio_rescan: *NOTICE*: AIO device is resized: bdev name /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev, old block count 51200, new block count 102400 00:12:07.732 [2024-12-06 17:27:49.526114] vbdev_lvol.c: 165:vbdev_lvs_base_bdev_event_cb: *NOTICE*: Unsupported bdev event: type 1 00:12:07.732 true 00:12:07.732 17:27:49 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 4edc6b8b-03f2-4934-b920-115813401976 00:12:07.732 17:27:49 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@38 -- # jq -r '.[0].total_data_clusters' 00:12:07.990 17:27:49 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@38 -- # (( data_clusters == 49 )) 00:12:07.990 17:27:49 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:12:08.558 17:27:50 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bcc5cd7d-1608-4625-8611-8cb86edc7b75 00:12:08.558 17:27:50 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:12:08.816 [2024-12-06 17:27:50.609256] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:12:08.816 17:27:50 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:12:09.074 17:27:50 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@48 -- # bdevperf_pid=146120 00:12:09.074 17:27:50 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock -m 0x2 -o 4096 -q 128 -w randwrite -t 10 -S 1 -z 00:12:09.075 17:27:50 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@49 -- # trap 'killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:12:09.075 17:27:50 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@50 -- # waitforlisten 146120 /var/tmp/bdevperf.sock 00:12:09.075 17:27:50 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@835 -- # '[' -z 146120 ']' 00:12:09.075 17:27:50 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:12:09.075 17:27:50 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:09.075 17:27:50 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:12:09.075 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:12:09.075 17:27:50 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:09.075 17:27:50 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:12:09.333 [2024-12-06 17:27:50.940074] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:12:09.333 [2024-12-06 17:27:50.940144] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid146120 ] 00:12:09.333 [2024-12-06 17:27:51.005978] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:09.333 [2024-12-06 17:27:51.050375] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:12:09.333 17:27:51 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:09.333 17:27:51 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@868 -- # return 0 00:12:09.334 17:27:51 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 00:12:09.901 Nvme0n1 00:12:09.901 17:27:51 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_get_bdevs -b Nvme0n1 -t 3000 00:12:10.159 [ 00:12:10.159 { 00:12:10.159 "name": "Nvme0n1", 00:12:10.159 "aliases": [ 00:12:10.159 "bcc5cd7d-1608-4625-8611-8cb86edc7b75" 00:12:10.159 ], 00:12:10.159 "product_name": "NVMe disk", 00:12:10.159 "block_size": 4096, 00:12:10.159 "num_blocks": 38912, 00:12:10.159 "uuid": "bcc5cd7d-1608-4625-8611-8cb86edc7b75", 00:12:10.159 "numa_id": 0, 00:12:10.159 "assigned_rate_limits": { 00:12:10.159 "rw_ios_per_sec": 0, 00:12:10.159 "rw_mbytes_per_sec": 0, 00:12:10.159 "r_mbytes_per_sec": 0, 00:12:10.159 "w_mbytes_per_sec": 0 00:12:10.159 }, 00:12:10.159 "claimed": false, 00:12:10.159 "zoned": false, 00:12:10.159 "supported_io_types": { 00:12:10.159 "read": true, 00:12:10.159 "write": true, 00:12:10.159 "unmap": true, 00:12:10.159 "flush": true, 00:12:10.159 "reset": true, 00:12:10.159 "nvme_admin": true, 00:12:10.159 "nvme_io": true, 00:12:10.159 "nvme_io_md": false, 00:12:10.159 "write_zeroes": true, 00:12:10.159 "zcopy": false, 00:12:10.159 "get_zone_info": false, 00:12:10.159 "zone_management": false, 00:12:10.159 "zone_append": false, 00:12:10.159 "compare": true, 00:12:10.159 "compare_and_write": true, 00:12:10.159 "abort": true, 00:12:10.159 "seek_hole": false, 00:12:10.159 "seek_data": false, 00:12:10.159 "copy": true, 00:12:10.159 "nvme_iov_md": false 00:12:10.159 }, 00:12:10.159 "memory_domains": [ 00:12:10.159 { 00:12:10.159 "dma_device_id": "system", 00:12:10.159 "dma_device_type": 1 00:12:10.159 } 00:12:10.159 ], 00:12:10.159 "driver_specific": { 00:12:10.159 "nvme": [ 00:12:10.159 { 00:12:10.159 "trid": { 00:12:10.159 "trtype": "TCP", 00:12:10.159 "adrfam": "IPv4", 00:12:10.159 "traddr": "10.0.0.2", 00:12:10.159 "trsvcid": "4420", 00:12:10.159 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:12:10.159 }, 00:12:10.159 "ctrlr_data": { 00:12:10.159 "cntlid": 1, 00:12:10.159 "vendor_id": "0x8086", 00:12:10.159 "model_number": "SPDK bdev Controller", 00:12:10.159 "serial_number": "SPDK0", 00:12:10.159 "firmware_revision": "25.01", 00:12:10.159 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:12:10.159 "oacs": { 00:12:10.159 "security": 0, 00:12:10.159 "format": 0, 00:12:10.159 "firmware": 0, 00:12:10.159 "ns_manage": 0 00:12:10.159 }, 00:12:10.159 "multi_ctrlr": true, 00:12:10.159 "ana_reporting": false 00:12:10.159 }, 00:12:10.159 "vs": { 00:12:10.159 "nvme_version": "1.3" 00:12:10.159 }, 00:12:10.159 "ns_data": { 00:12:10.159 "id": 1, 00:12:10.159 "can_share": true 00:12:10.159 } 00:12:10.159 } 00:12:10.159 ], 00:12:10.159 "mp_policy": "active_passive" 00:12:10.159 } 00:12:10.159 } 00:12:10.159 ] 00:12:10.159 17:27:51 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@56 -- # run_test_pid=146256 00:12:10.159 17:27:51 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@57 -- # sleep 2 00:12:10.159 17:27:51 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:12:10.419 Running I/O for 10 seconds... 00:12:11.355 Latency(us) 00:12:11.355 [2024-12-06T16:27:53.194Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:11.355 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:12:11.355 Nvme0n1 : 1.00 15114.00 59.04 0.00 0.00 0.00 0.00 0.00 00:12:11.355 [2024-12-06T16:27:53.194Z] =================================================================================================================== 00:12:11.355 [2024-12-06T16:27:53.194Z] Total : 15114.00 59.04 0.00 0.00 0.00 0.00 0.00 00:12:11.355 00:12:12.290 17:27:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_grow_lvstore -u 4edc6b8b-03f2-4934-b920-115813401976 00:12:12.290 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:12:12.290 Nvme0n1 : 2.00 15336.00 59.91 0.00 0.00 0.00 0.00 0.00 00:12:12.290 [2024-12-06T16:27:54.129Z] =================================================================================================================== 00:12:12.290 [2024-12-06T16:27:54.129Z] Total : 15336.00 59.91 0.00 0.00 0.00 0.00 0.00 00:12:12.290 00:12:12.550 true 00:12:12.550 17:27:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 4edc6b8b-03f2-4934-b920-115813401976 00:12:12.550 17:27:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@61 -- # jq -r '.[0].total_data_clusters' 00:12:12.808 17:27:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@61 -- # data_clusters=99 00:12:12.808 17:27:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@62 -- # (( data_clusters == 99 )) 00:12:12.808 17:27:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@65 -- # wait 146256 00:12:13.383 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:12:13.383 Nvme0n1 : 3.00 15452.00 60.36 0.00 0.00 0.00 0.00 0.00 00:12:13.383 [2024-12-06T16:27:55.222Z] =================================================================================================================== 00:12:13.383 [2024-12-06T16:27:55.222Z] Total : 15452.00 60.36 0.00 0.00 0.00 0.00 0.00 00:12:13.383 00:12:14.377 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:12:14.377 Nvme0n1 : 4.00 15557.75 60.77 0.00 0.00 0.00 0.00 0.00 00:12:14.377 [2024-12-06T16:27:56.216Z] =================================================================================================================== 00:12:14.377 [2024-12-06T16:27:56.216Z] Total : 15557.75 60.77 0.00 0.00 0.00 0.00 0.00 00:12:14.377 00:12:15.312 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:12:15.312 Nvme0n1 : 5.00 15646.60 61.12 0.00 0.00 0.00 0.00 0.00 00:12:15.312 [2024-12-06T16:27:57.151Z] =================================================================================================================== 00:12:15.312 [2024-12-06T16:27:57.151Z] Total : 15646.60 61.12 0.00 0.00 0.00 0.00 0.00 00:12:15.312 00:12:16.247 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:12:16.247 Nvme0n1 : 6.00 15705.83 61.35 0.00 0.00 0.00 0.00 0.00 00:12:16.247 [2024-12-06T16:27:58.086Z] =================================================================================================================== 00:12:16.247 [2024-12-06T16:27:58.086Z] Total : 15705.83 61.35 0.00 0.00 0.00 0.00 0.00 00:12:16.247 00:12:17.628 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:12:17.628 Nvme0n1 : 7.00 15748.14 61.52 0.00 0.00 0.00 0.00 0.00 00:12:17.628 [2024-12-06T16:27:59.467Z] =================================================================================================================== 00:12:17.628 [2024-12-06T16:27:59.467Z] Total : 15748.14 61.52 0.00 0.00 0.00 0.00 0.00 00:12:17.628 00:12:18.565 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:12:18.565 Nvme0n1 : 8.00 15795.75 61.70 0.00 0.00 0.00 0.00 0.00 00:12:18.565 [2024-12-06T16:28:00.404Z] =================================================================================================================== 00:12:18.565 [2024-12-06T16:28:00.404Z] Total : 15795.75 61.70 0.00 0.00 0.00 0.00 0.00 00:12:18.565 00:12:19.501 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:12:19.501 Nvme0n1 : 9.00 15818.67 61.79 0.00 0.00 0.00 0.00 0.00 00:12:19.501 [2024-12-06T16:28:01.340Z] =================================================================================================================== 00:12:19.501 [2024-12-06T16:28:01.341Z] Total : 15818.67 61.79 0.00 0.00 0.00 0.00 0.00 00:12:19.502 00:12:20.438 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:12:20.438 Nvme0n1 : 10.00 15843.50 61.89 0.00 0.00 0.00 0.00 0.00 00:12:20.438 [2024-12-06T16:28:02.277Z] =================================================================================================================== 00:12:20.438 [2024-12-06T16:28:02.277Z] Total : 15843.50 61.89 0.00 0.00 0.00 0.00 0.00 00:12:20.438 00:12:20.438 00:12:20.438 Latency(us) 00:12:20.438 [2024-12-06T16:28:02.277Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:20.438 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:12:20.438 Nvme0n1 : 10.00 15848.97 61.91 0.00 0.00 8071.79 2330.17 16019.91 00:12:20.438 [2024-12-06T16:28:02.277Z] =================================================================================================================== 00:12:20.438 [2024-12-06T16:28:02.277Z] Total : 15848.97 61.91 0.00 0.00 8071.79 2330.17 16019.91 00:12:20.438 { 00:12:20.438 "results": [ 00:12:20.438 { 00:12:20.438 "job": "Nvme0n1", 00:12:20.438 "core_mask": "0x2", 00:12:20.438 "workload": "randwrite", 00:12:20.438 "status": "finished", 00:12:20.438 "queue_depth": 128, 00:12:20.438 "io_size": 4096, 00:12:20.438 "runtime": 10.004622, 00:12:20.438 "iops": 15848.97460393806, 00:12:20.438 "mibps": 61.91005704663305, 00:12:20.438 "io_failed": 0, 00:12:20.438 "io_timeout": 0, 00:12:20.438 "avg_latency_us": 8071.790139411814, 00:12:20.438 "min_latency_us": 2330.168888888889, 00:12:20.438 "max_latency_us": 16019.91111111111 00:12:20.438 } 00:12:20.438 ], 00:12:20.438 "core_count": 1 00:12:20.438 } 00:12:20.438 17:28:02 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@66 -- # killprocess 146120 00:12:20.438 17:28:02 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@954 -- # '[' -z 146120 ']' 00:12:20.438 17:28:02 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@958 -- # kill -0 146120 00:12:20.438 17:28:02 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@959 -- # uname 00:12:20.438 17:28:02 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:20.438 17:28:02 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 146120 00:12:20.438 17:28:02 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:12:20.438 17:28:02 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:12:20.438 17:28:02 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@972 -- # echo 'killing process with pid 146120' 00:12:20.438 killing process with pid 146120 00:12:20.438 17:28:02 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@973 -- # kill 146120 00:12:20.438 Received shutdown signal, test time was about 10.000000 seconds 00:12:20.438 00:12:20.438 Latency(us) 00:12:20.438 [2024-12-06T16:28:02.277Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:20.438 [2024-12-06T16:28:02.277Z] =================================================================================================================== 00:12:20.438 [2024-12-06T16:28:02.277Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:12:20.438 17:28:02 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@978 -- # wait 146120 00:12:20.697 17:28:02 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@68 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:12:20.955 17:28:02 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@69 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:12:21.226 17:28:02 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@70 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 4edc6b8b-03f2-4934-b920-115813401976 00:12:21.226 17:28:02 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@70 -- # jq -r '.[0].free_clusters' 00:12:21.489 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@70 -- # free_clusters=61 00:12:21.489 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@72 -- # [[ dirty == \d\i\r\t\y ]] 00:12:21.489 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@74 -- # kill -9 143501 00:12:21.489 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@75 -- # wait 143501 00:12:21.489 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh: line 75: 143501 Killed "${NVMF_APP[@]}" "$@" 00:12:21.489 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@75 -- # true 00:12:21.489 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@76 -- # nvmfappstart -m 0x1 00:12:21.489 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:12:21.489 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@726 -- # xtrace_disable 00:12:21.489 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:12:21.489 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@509 -- # nvmfpid=147592 00:12:21.489 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1 00:12:21.489 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@510 -- # waitforlisten 147592 00:12:21.489 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@835 -- # '[' -z 147592 ']' 00:12:21.489 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:21.489 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:21.489 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:21.489 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:21.489 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:21.489 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:12:21.489 [2024-12-06 17:28:03.227181] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:12:21.489 [2024-12-06 17:28:03.227268] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:12:21.489 [2024-12-06 17:28:03.299729] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:21.747 [2024-12-06 17:28:03.348225] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:12:21.747 [2024-12-06 17:28:03.348285] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:12:21.747 [2024-12-06 17:28:03.348305] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:12:21.747 [2024-12-06 17:28:03.348316] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:12:21.747 [2024-12-06 17:28:03.348326] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:12:21.747 [2024-12-06 17:28:03.348915] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:21.747 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:21.747 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@868 -- # return 0 00:12:21.747 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:12:21.747 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@732 -- # xtrace_disable 00:12:21.747 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:12:21.747 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:12:21.747 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@77 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:12:22.006 [2024-12-06 17:28:03.742814] blobstore.c:4899:bs_recover: *NOTICE*: Performing recovery on blobstore 00:12:22.006 [2024-12-06 17:28:03.742963] blobstore.c:4846:bs_load_replay_md_cpl: *NOTICE*: Recover: blob 0x0 00:12:22.006 [2024-12-06 17:28:03.743029] blobstore.c:4846:bs_load_replay_md_cpl: *NOTICE*: Recover: blob 0x1 00:12:22.006 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@77 -- # aio_bdev=aio_bdev 00:12:22.006 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@78 -- # waitforbdev bcc5cd7d-1608-4625-8611-8cb86edc7b75 00:12:22.006 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@903 -- # local bdev_name=bcc5cd7d-1608-4625-8611-8cb86edc7b75 00:12:22.006 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:22.006 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@905 -- # local i 00:12:22.006 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:22.006 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:22.006 17:28:03 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@908 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:12:22.265 17:28:04 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@910 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b bcc5cd7d-1608-4625-8611-8cb86edc7b75 -t 2000 00:12:22.525 [ 00:12:22.525 { 00:12:22.525 "name": "bcc5cd7d-1608-4625-8611-8cb86edc7b75", 00:12:22.525 "aliases": [ 00:12:22.525 "lvs/lvol" 00:12:22.525 ], 00:12:22.525 "product_name": "Logical Volume", 00:12:22.525 "block_size": 4096, 00:12:22.525 "num_blocks": 38912, 00:12:22.525 "uuid": "bcc5cd7d-1608-4625-8611-8cb86edc7b75", 00:12:22.525 "assigned_rate_limits": { 00:12:22.525 "rw_ios_per_sec": 0, 00:12:22.525 "rw_mbytes_per_sec": 0, 00:12:22.525 "r_mbytes_per_sec": 0, 00:12:22.525 "w_mbytes_per_sec": 0 00:12:22.525 }, 00:12:22.525 "claimed": false, 00:12:22.525 "zoned": false, 00:12:22.525 "supported_io_types": { 00:12:22.525 "read": true, 00:12:22.525 "write": true, 00:12:22.525 "unmap": true, 00:12:22.525 "flush": false, 00:12:22.525 "reset": true, 00:12:22.525 "nvme_admin": false, 00:12:22.525 "nvme_io": false, 00:12:22.525 "nvme_io_md": false, 00:12:22.525 "write_zeroes": true, 00:12:22.525 "zcopy": false, 00:12:22.525 "get_zone_info": false, 00:12:22.525 "zone_management": false, 00:12:22.525 "zone_append": false, 00:12:22.525 "compare": false, 00:12:22.525 "compare_and_write": false, 00:12:22.525 "abort": false, 00:12:22.525 "seek_hole": true, 00:12:22.525 "seek_data": true, 00:12:22.525 "copy": false, 00:12:22.525 "nvme_iov_md": false 00:12:22.525 }, 00:12:22.525 "driver_specific": { 00:12:22.525 "lvol": { 00:12:22.525 "lvol_store_uuid": "4edc6b8b-03f2-4934-b920-115813401976", 00:12:22.525 "base_bdev": "aio_bdev", 00:12:22.525 "thin_provision": false, 00:12:22.525 "num_allocated_clusters": 38, 00:12:22.525 "snapshot": false, 00:12:22.525 "clone": false, 00:12:22.525 "esnap_clone": false 00:12:22.525 } 00:12:22.525 } 00:12:22.525 } 00:12:22.525 ] 00:12:22.525 17:28:04 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@911 -- # return 0 00:12:22.525 17:28:04 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 4edc6b8b-03f2-4934-b920-115813401976 00:12:22.525 17:28:04 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@79 -- # jq -r '.[0].free_clusters' 00:12:22.784 17:28:04 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@79 -- # (( free_clusters == 61 )) 00:12:22.785 17:28:04 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@80 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 4edc6b8b-03f2-4934-b920-115813401976 00:12:22.785 17:28:04 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@80 -- # jq -r '.[0].total_data_clusters' 00:12:23.043 17:28:04 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@80 -- # (( data_clusters == 99 )) 00:12:23.043 17:28:04 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:12:23.302 [2024-12-06 17:28:05.120439] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev aio_bdev being removed: closing lvstore lvs 00:12:23.562 17:28:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@85 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 4edc6b8b-03f2-4934-b920-115813401976 00:12:23.562 17:28:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@652 -- # local es=0 00:12:23.562 17:28:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@654 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 4edc6b8b-03f2-4934-b920-115813401976 00:12:23.562 17:28:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@640 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:12:23.562 17:28:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:23.562 17:28:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@644 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:12:23.562 17:28:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:23.562 17:28:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@646 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:12:23.562 17:28:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:12:23.562 17:28:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@646 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:12:23.562 17:28:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@646 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:12:23.562 17:28:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@655 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 4edc6b8b-03f2-4934-b920-115813401976 00:12:23.820 request: 00:12:23.820 { 00:12:23.820 "uuid": "4edc6b8b-03f2-4934-b920-115813401976", 00:12:23.820 "method": "bdev_lvol_get_lvstores", 00:12:23.820 "req_id": 1 00:12:23.820 } 00:12:23.821 Got JSON-RPC error response 00:12:23.821 response: 00:12:23.821 { 00:12:23.821 "code": -19, 00:12:23.821 "message": "No such device" 00:12:23.821 } 00:12:23.821 17:28:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@655 -- # es=1 00:12:23.821 17:28:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:12:23.821 17:28:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:12:23.821 17:28:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:12:23.821 17:28:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:12:24.078 aio_bdev 00:12:24.078 17:28:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@87 -- # waitforbdev bcc5cd7d-1608-4625-8611-8cb86edc7b75 00:12:24.078 17:28:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@903 -- # local bdev_name=bcc5cd7d-1608-4625-8611-8cb86edc7b75 00:12:24.078 17:28:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:12:24.078 17:28:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@905 -- # local i 00:12:24.078 17:28:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:12:24.078 17:28:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:12:24.078 17:28:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@908 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:12:24.336 17:28:05 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@910 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b bcc5cd7d-1608-4625-8611-8cb86edc7b75 -t 2000 00:12:24.595 [ 00:12:24.595 { 00:12:24.595 "name": "bcc5cd7d-1608-4625-8611-8cb86edc7b75", 00:12:24.595 "aliases": [ 00:12:24.595 "lvs/lvol" 00:12:24.595 ], 00:12:24.595 "product_name": "Logical Volume", 00:12:24.595 "block_size": 4096, 00:12:24.595 "num_blocks": 38912, 00:12:24.595 "uuid": "bcc5cd7d-1608-4625-8611-8cb86edc7b75", 00:12:24.595 "assigned_rate_limits": { 00:12:24.595 "rw_ios_per_sec": 0, 00:12:24.595 "rw_mbytes_per_sec": 0, 00:12:24.595 "r_mbytes_per_sec": 0, 00:12:24.595 "w_mbytes_per_sec": 0 00:12:24.595 }, 00:12:24.595 "claimed": false, 00:12:24.595 "zoned": false, 00:12:24.595 "supported_io_types": { 00:12:24.595 "read": true, 00:12:24.595 "write": true, 00:12:24.595 "unmap": true, 00:12:24.595 "flush": false, 00:12:24.595 "reset": true, 00:12:24.595 "nvme_admin": false, 00:12:24.595 "nvme_io": false, 00:12:24.595 "nvme_io_md": false, 00:12:24.595 "write_zeroes": true, 00:12:24.595 "zcopy": false, 00:12:24.595 "get_zone_info": false, 00:12:24.595 "zone_management": false, 00:12:24.595 "zone_append": false, 00:12:24.595 "compare": false, 00:12:24.595 "compare_and_write": false, 00:12:24.595 "abort": false, 00:12:24.595 "seek_hole": true, 00:12:24.595 "seek_data": true, 00:12:24.595 "copy": false, 00:12:24.595 "nvme_iov_md": false 00:12:24.595 }, 00:12:24.595 "driver_specific": { 00:12:24.595 "lvol": { 00:12:24.595 "lvol_store_uuid": "4edc6b8b-03f2-4934-b920-115813401976", 00:12:24.595 "base_bdev": "aio_bdev", 00:12:24.595 "thin_provision": false, 00:12:24.595 "num_allocated_clusters": 38, 00:12:24.595 "snapshot": false, 00:12:24.595 "clone": false, 00:12:24.595 "esnap_clone": false 00:12:24.595 } 00:12:24.595 } 00:12:24.595 } 00:12:24.595 ] 00:12:24.595 17:28:06 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@911 -- # return 0 00:12:24.595 17:28:06 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 4edc6b8b-03f2-4934-b920-115813401976 00:12:24.595 17:28:06 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@88 -- # jq -r '.[0].free_clusters' 00:12:24.854 17:28:06 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@88 -- # (( free_clusters == 61 )) 00:12:24.854 17:28:06 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 4edc6b8b-03f2-4934-b920-115813401976 00:12:24.854 17:28:06 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@89 -- # jq -r '.[0].total_data_clusters' 00:12:25.113 17:28:06 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@89 -- # (( data_clusters == 99 )) 00:12:25.113 17:28:06 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@92 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete bcc5cd7d-1608-4625-8611-8cb86edc7b75 00:12:25.372 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@93 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u 4edc6b8b-03f2-4934-b920-115813401976 00:12:25.630 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@94 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:12:25.889 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@95 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:12:25.889 00:12:25.889 real 0m19.490s 00:12:25.889 user 0m49.170s 00:12:25.889 sys 0m4.571s 00:12:25.889 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:25.889 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:12:25.889 ************************************ 00:12:25.889 END TEST lvs_grow_dirty 00:12:25.889 ************************************ 00:12:25.889 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@1 -- # process_shm --id 0 00:12:25.889 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@812 -- # type=--id 00:12:25.889 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@813 -- # id=0 00:12:25.889 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@814 -- # '[' --id = --pid ']' 00:12:25.889 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@818 -- # find /dev/shm -name '*.0' -printf '%f\n' 00:12:25.889 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@818 -- # shm_files=nvmf_trace.0 00:12:25.889 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@820 -- # [[ -z nvmf_trace.0 ]] 00:12:25.889 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@824 -- # for n in $shm_files 00:12:25.889 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@825 -- # tar -C /dev/shm/ -cvzf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf_trace.0_shm.tar.gz nvmf_trace.0 00:12:25.889 nvmf_trace.0 00:12:26.147 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@827 -- # return 0 00:12:26.147 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@1 -- # nvmftestfini 00:12:26.147 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@516 -- # nvmfcleanup 00:12:26.147 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@121 -- # sync 00:12:26.147 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:12:26.147 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@124 -- # set +e 00:12:26.147 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@125 -- # for i in {1..20} 00:12:26.147 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:12:26.147 rmmod nvme_tcp 00:12:26.147 rmmod nvme_fabrics 00:12:26.147 rmmod nvme_keyring 00:12:26.147 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:12:26.147 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@128 -- # set -e 00:12:26.147 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@129 -- # return 0 00:12:26.147 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@517 -- # '[' -n 147592 ']' 00:12:26.147 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@518 -- # killprocess 147592 00:12:26.147 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@954 -- # '[' -z 147592 ']' 00:12:26.147 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@958 -- # kill -0 147592 00:12:26.147 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@959 -- # uname 00:12:26.147 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:26.147 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 147592 00:12:26.147 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:26.147 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:26.147 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@972 -- # echo 'killing process with pid 147592' 00:12:26.147 killing process with pid 147592 00:12:26.147 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@973 -- # kill 147592 00:12:26.147 17:28:07 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@978 -- # wait 147592 00:12:26.407 17:28:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:12:26.407 17:28:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:12:26.407 17:28:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:12:26.408 17:28:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@297 -- # iptr 00:12:26.408 17:28:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@791 -- # iptables-save 00:12:26.408 17:28:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:12:26.408 17:28:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@791 -- # iptables-restore 00:12:26.408 17:28:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:12:26.408 17:28:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@302 -- # remove_spdk_ns 00:12:26.408 17:28:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:26.408 17:28:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:26.408 17:28:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:28.314 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:12:28.314 00:12:28.314 real 0m42.654s 00:12:28.314 user 1m12.321s 00:12:28.314 sys 0m8.526s 00:12:28.314 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:28.314 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:12:28.314 ************************************ 00:12:28.314 END TEST nvmf_lvs_grow 00:12:28.314 ************************************ 00:12:28.314 17:28:10 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@29 -- # run_test nvmf_bdev_io_wait /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdev_io_wait.sh --transport=tcp 00:12:28.314 17:28:10 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:12:28.314 17:28:10 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:28.314 17:28:10 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:12:28.314 ************************************ 00:12:28.314 START TEST nvmf_bdev_io_wait 00:12:28.314 ************************************ 00:12:28.314 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdev_io_wait.sh --transport=tcp 00:12:28.573 * Looking for test storage... 00:12:28.573 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1711 -- # lcov --version 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@333 -- # local ver1 ver1_l 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@334 -- # local ver2 ver2_l 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@336 -- # IFS=.-: 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@336 -- # read -ra ver1 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@337 -- # IFS=.-: 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@337 -- # read -ra ver2 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@338 -- # local 'op=<' 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@340 -- # ver1_l=2 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@341 -- # ver2_l=1 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@344 -- # case "$op" in 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@345 -- # : 1 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@364 -- # (( v = 0 )) 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@365 -- # decimal 1 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@353 -- # local d=1 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@355 -- # echo 1 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@365 -- # ver1[v]=1 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@366 -- # decimal 2 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@353 -- # local d=2 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@355 -- # echo 2 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@366 -- # ver2[v]=2 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@368 -- # return 0 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:12:28.573 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:28.573 --rc genhtml_branch_coverage=1 00:12:28.573 --rc genhtml_function_coverage=1 00:12:28.573 --rc genhtml_legend=1 00:12:28.573 --rc geninfo_all_blocks=1 00:12:28.573 --rc geninfo_unexecuted_blocks=1 00:12:28.573 00:12:28.573 ' 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:12:28.573 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:28.573 --rc genhtml_branch_coverage=1 00:12:28.573 --rc genhtml_function_coverage=1 00:12:28.573 --rc genhtml_legend=1 00:12:28.573 --rc geninfo_all_blocks=1 00:12:28.573 --rc geninfo_unexecuted_blocks=1 00:12:28.573 00:12:28.573 ' 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:12:28.573 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:28.573 --rc genhtml_branch_coverage=1 00:12:28.573 --rc genhtml_function_coverage=1 00:12:28.573 --rc genhtml_legend=1 00:12:28.573 --rc geninfo_all_blocks=1 00:12:28.573 --rc geninfo_unexecuted_blocks=1 00:12:28.573 00:12:28.573 ' 00:12:28.573 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:12:28.574 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:28.574 --rc genhtml_branch_coverage=1 00:12:28.574 --rc genhtml_function_coverage=1 00:12:28.574 --rc genhtml_legend=1 00:12:28.574 --rc geninfo_all_blocks=1 00:12:28.574 --rc geninfo_unexecuted_blocks=1 00:12:28.574 00:12:28.574 ' 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@7 -- # uname -s 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@15 -- # shopt -s extglob 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- paths/export.sh@5 -- # export PATH 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@51 -- # : 0 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:12:28.574 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@55 -- # have_pci_nics=0 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@11 -- # MALLOC_BDEV_SIZE=64 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@14 -- # nvmftestinit 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@476 -- # prepare_net_devs 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@438 -- # local -g is_hw=no 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@440 -- # remove_spdk_ns 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@309 -- # xtrace_disable 00:12:28.574 17:28:10 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:12:31.114 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:12:31.114 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@315 -- # pci_devs=() 00:12:31.114 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@315 -- # local -a pci_devs 00:12:31.114 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@316 -- # pci_net_devs=() 00:12:31.114 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:12:31.114 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@317 -- # pci_drivers=() 00:12:31.114 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@317 -- # local -A pci_drivers 00:12:31.114 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@319 -- # net_devs=() 00:12:31.114 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@319 -- # local -ga net_devs 00:12:31.114 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@320 -- # e810=() 00:12:31.114 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@320 -- # local -ga e810 00:12:31.114 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@321 -- # x722=() 00:12:31.114 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@321 -- # local -ga x722 00:12:31.114 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@322 -- # mlx=() 00:12:31.114 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@322 -- # local -ga mlx 00:12:31.114 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:12:31.114 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:12:31.114 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:12:31.115 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:12:31.115 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@418 -- # [[ up == up ]] 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:12:31.115 Found net devices under 0000:0a:00.0: cvl_0_0 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@418 -- # [[ up == up ]] 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:12:31.115 Found net devices under 0000:0a:00.1: cvl_0_1 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@442 -- # is_hw=yes 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:12:31.115 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:12:31.115 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.348 ms 00:12:31.115 00:12:31.115 --- 10.0.0.2 ping statistics --- 00:12:31.115 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:31.115 rtt min/avg/max/mdev = 0.348/0.348/0.348/0.000 ms 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:12:31.115 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:12:31.115 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.147 ms 00:12:31.115 00:12:31.115 --- 10.0.0.1 ping statistics --- 00:12:31.115 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:31.115 rtt min/avg/max/mdev = 0.147/0.147/0.147/0.000 ms 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@450 -- # return 0 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@15 -- # nvmfappstart -m 0xF --wait-for-rpc 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@726 -- # xtrace_disable 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@509 -- # nvmfpid=150130 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@510 -- # waitforlisten 150130 00:12:31.115 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@835 -- # '[' -z 150130 ']' 00:12:31.116 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF --wait-for-rpc 00:12:31.116 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:31.116 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:31.116 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:31.116 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:31.116 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:31.116 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:12:31.116 [2024-12-06 17:28:12.608712] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:12:31.116 [2024-12-06 17:28:12.608798] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:12:31.116 [2024-12-06 17:28:12.680389] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:12:31.116 [2024-12-06 17:28:12.731350] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:12:31.116 [2024-12-06 17:28:12.731404] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:12:31.116 [2024-12-06 17:28:12.731417] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:12:31.116 [2024-12-06 17:28:12.731428] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:12:31.116 [2024-12-06 17:28:12.731438] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:12:31.116 [2024-12-06 17:28:12.732959] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:12:31.116 [2024-12-06 17:28:12.733022] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:12:31.116 [2024-12-06 17:28:12.733086] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:12:31.116 [2024-12-06 17:28:12.733089] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:31.116 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:31.116 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@868 -- # return 0 00:12:31.116 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:12:31.116 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@732 -- # xtrace_disable 00:12:31.116 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:12:31.116 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:12:31.116 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@18 -- # rpc_cmd bdev_set_options -p 5 -c 1 00:12:31.116 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.116 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:12:31.116 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.116 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@19 -- # rpc_cmd framework_start_init 00:12:31.116 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.116 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:12:31.116 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.116 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@20 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:12:31.116 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.116 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:12:31.116 [2024-12-06 17:28:12.951033] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:12:31.375 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.375 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@22 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:12:31.375 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.375 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:12:31.375 Malloc0 00:12:31.375 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.375 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@23 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:12:31.375 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.375 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:12:31.375 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.375 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:12:31.375 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.375 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:12:31.375 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.375 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:12:31.375 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:31.375 17:28:12 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:12:31.375 [2024-12-06 17:28:13.002227] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:12:31.375 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:31.375 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@28 -- # WRITE_PID=150196 00:12:31.375 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@30 -- # READ_PID=150199 00:12:31.375 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@27 -- # gen_nvmf_target_json 00:12:31.375 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x10 -i 1 --json /dev/fd/63 -q 128 -o 4096 -w write -t 1 -s 256 00:12:31.375 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # config=() 00:12:31.375 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # local subsystem config 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:12:31.376 { 00:12:31.376 "params": { 00:12:31.376 "name": "Nvme$subsystem", 00:12:31.376 "trtype": "$TEST_TRANSPORT", 00:12:31.376 "traddr": "$NVMF_FIRST_TARGET_IP", 00:12:31.376 "adrfam": "ipv4", 00:12:31.376 "trsvcid": "$NVMF_PORT", 00:12:31.376 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:12:31.376 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:12:31.376 "hdgst": ${hdgst:-false}, 00:12:31.376 "ddgst": ${ddgst:-false} 00:12:31.376 }, 00:12:31.376 "method": "bdev_nvme_attach_controller" 00:12:31.376 } 00:12:31.376 EOF 00:12:31.376 )") 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@32 -- # FLUSH_PID=150202 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@29 -- # gen_nvmf_target_json 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x20 -i 2 --json /dev/fd/63 -q 128 -o 4096 -w read -t 1 -s 256 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # config=() 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # local subsystem config 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x40 -i 3 --json /dev/fd/63 -q 128 -o 4096 -w flush -t 1 -s 256 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@31 -- # gen_nvmf_target_json 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:12:31.376 { 00:12:31.376 "params": { 00:12:31.376 "name": "Nvme$subsystem", 00:12:31.376 "trtype": "$TEST_TRANSPORT", 00:12:31.376 "traddr": "$NVMF_FIRST_TARGET_IP", 00:12:31.376 "adrfam": "ipv4", 00:12:31.376 "trsvcid": "$NVMF_PORT", 00:12:31.376 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:12:31.376 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:12:31.376 "hdgst": ${hdgst:-false}, 00:12:31.376 "ddgst": ${ddgst:-false} 00:12:31.376 }, 00:12:31.376 "method": "bdev_nvme_attach_controller" 00:12:31.376 } 00:12:31.376 EOF 00:12:31.376 )") 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@34 -- # UNMAP_PID=150206 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@35 -- # sync 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # config=() 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # cat 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # local subsystem config 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@33 -- # gen_nvmf_target_json 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x80 -i 4 --json /dev/fd/63 -q 128 -o 4096 -w unmap -t 1 -s 256 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:12:31.376 { 00:12:31.376 "params": { 00:12:31.376 "name": "Nvme$subsystem", 00:12:31.376 "trtype": "$TEST_TRANSPORT", 00:12:31.376 "traddr": "$NVMF_FIRST_TARGET_IP", 00:12:31.376 "adrfam": "ipv4", 00:12:31.376 "trsvcid": "$NVMF_PORT", 00:12:31.376 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:12:31.376 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:12:31.376 "hdgst": ${hdgst:-false}, 00:12:31.376 "ddgst": ${ddgst:-false} 00:12:31.376 }, 00:12:31.376 "method": "bdev_nvme_attach_controller" 00:12:31.376 } 00:12:31.376 EOF 00:12:31.376 )") 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # config=() 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # local subsystem config 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:12:31.376 { 00:12:31.376 "params": { 00:12:31.376 "name": "Nvme$subsystem", 00:12:31.376 "trtype": "$TEST_TRANSPORT", 00:12:31.376 "traddr": "$NVMF_FIRST_TARGET_IP", 00:12:31.376 "adrfam": "ipv4", 00:12:31.376 "trsvcid": "$NVMF_PORT", 00:12:31.376 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:12:31.376 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:12:31.376 "hdgst": ${hdgst:-false}, 00:12:31.376 "ddgst": ${ddgst:-false} 00:12:31.376 }, 00:12:31.376 "method": "bdev_nvme_attach_controller" 00:12:31.376 } 00:12:31.376 EOF 00:12:31.376 )") 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # cat 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # cat 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@37 -- # wait 150196 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # cat 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # jq . 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # jq . 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # jq . 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@585 -- # IFS=, 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:12:31.376 "params": { 00:12:31.376 "name": "Nvme1", 00:12:31.376 "trtype": "tcp", 00:12:31.376 "traddr": "10.0.0.2", 00:12:31.376 "adrfam": "ipv4", 00:12:31.376 "trsvcid": "4420", 00:12:31.376 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:12:31.376 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:12:31.376 "hdgst": false, 00:12:31.376 "ddgst": false 00:12:31.376 }, 00:12:31.376 "method": "bdev_nvme_attach_controller" 00:12:31.376 }' 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # jq . 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@585 -- # IFS=, 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:12:31.376 "params": { 00:12:31.376 "name": "Nvme1", 00:12:31.376 "trtype": "tcp", 00:12:31.376 "traddr": "10.0.0.2", 00:12:31.376 "adrfam": "ipv4", 00:12:31.376 "trsvcid": "4420", 00:12:31.376 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:12:31.376 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:12:31.376 "hdgst": false, 00:12:31.376 "ddgst": false 00:12:31.376 }, 00:12:31.376 "method": "bdev_nvme_attach_controller" 00:12:31.376 }' 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@585 -- # IFS=, 00:12:31.376 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:12:31.376 "params": { 00:12:31.376 "name": "Nvme1", 00:12:31.376 "trtype": "tcp", 00:12:31.376 "traddr": "10.0.0.2", 00:12:31.376 "adrfam": "ipv4", 00:12:31.376 "trsvcid": "4420", 00:12:31.376 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:12:31.376 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:12:31.376 "hdgst": false, 00:12:31.376 "ddgst": false 00:12:31.377 }, 00:12:31.377 "method": "bdev_nvme_attach_controller" 00:12:31.377 }' 00:12:31.377 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@585 -- # IFS=, 00:12:31.377 17:28:13 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:12:31.377 "params": { 00:12:31.377 "name": "Nvme1", 00:12:31.377 "trtype": "tcp", 00:12:31.377 "traddr": "10.0.0.2", 00:12:31.377 "adrfam": "ipv4", 00:12:31.377 "trsvcid": "4420", 00:12:31.377 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:12:31.377 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:12:31.377 "hdgst": false, 00:12:31.377 "ddgst": false 00:12:31.377 }, 00:12:31.377 "method": "bdev_nvme_attach_controller" 00:12:31.377 }' 00:12:31.377 [2024-12-06 17:28:13.053305] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:12:31.377 [2024-12-06 17:28:13.053301] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:12:31.377 [2024-12-06 17:28:13.053310] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:12:31.377 [2024-12-06 17:28:13.053363] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:12:31.377 [2024-12-06 17:28:13.053403] [ DPDK EAL parameters: bdevperf -c 0x10 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib[2024-12-06 17:28:13.053404] [ DPDK EAL parameters: bdevperf -c 0x40 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib[2024-12-06 17:28:13.053403] [ DPDK EAL parameters: bdevperf -c 0x20 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk1 .cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk3 .cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk2 --proc-type=auto ] 00:12:31.377 --proc-type=auto ] 00:12:31.377 --proc-type=auto ] 00:12:31.377 [2024-12-06 17:28:13.053442] [ DPDK EAL parameters: bdevperf -c 0x80 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk4 --proc-type=auto ] 00:12:31.635 [2024-12-06 17:28:13.240267] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:31.635 [2024-12-06 17:28:13.281987] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:12:31.635 [2024-12-06 17:28:13.339196] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:31.635 [2024-12-06 17:28:13.381355] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 6 00:12:31.635 [2024-12-06 17:28:13.437167] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:31.893 [2024-12-06 17:28:13.481564] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 5 00:12:31.893 [2024-12-06 17:28:13.512500] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:31.893 [2024-12-06 17:28:13.551503] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 7 00:12:31.893 Running I/O for 1 seconds... 00:12:31.893 Running I/O for 1 seconds... 00:12:32.151 Running I/O for 1 seconds... 00:12:32.151 Running I/O for 1 seconds... 00:12:33.093 192248.00 IOPS, 750.97 MiB/s [2024-12-06T16:28:14.932Z] 6772.00 IOPS, 26.45 MiB/s 00:12:33.093 Latency(us) 00:12:33.093 [2024-12-06T16:28:14.932Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:33.093 Job: Nvme1n1 (Core Mask 0x40, workload: flush, depth: 128, IO size: 4096) 00:12:33.093 Nvme1n1 : 1.00 191890.07 749.57 0.00 0.00 663.41 282.17 1844.72 00:12:33.093 [2024-12-06T16:28:14.932Z] =================================================================================================================== 00:12:33.093 [2024-12-06T16:28:14.932Z] Total : 191890.07 749.57 0.00 0.00 663.41 282.17 1844.72 00:12:33.093 00:12:33.093 Latency(us) 00:12:33.093 [2024-12-06T16:28:14.932Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:33.093 Job: Nvme1n1 (Core Mask 0x10, workload: write, depth: 128, IO size: 4096) 00:12:33.093 Nvme1n1 : 1.02 6803.56 26.58 0.00 0.00 18679.92 6990.51 28932.93 00:12:33.093 [2024-12-06T16:28:14.932Z] =================================================================================================================== 00:12:33.093 [2024-12-06T16:28:14.932Z] Total : 6803.56 26.58 0.00 0.00 18679.92 6990.51 28932.93 00:12:33.093 8781.00 IOPS, 34.30 MiB/s [2024-12-06T16:28:14.932Z] 6585.00 IOPS, 25.72 MiB/s 00:12:33.093 Latency(us) 00:12:33.093 [2024-12-06T16:28:14.932Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:33.093 Job: Nvme1n1 (Core Mask 0x20, workload: read, depth: 128, IO size: 4096) 00:12:33.093 Nvme1n1 : 1.01 8821.94 34.46 0.00 0.00 14432.50 8689.59 24175.50 00:12:33.093 [2024-12-06T16:28:14.932Z] =================================================================================================================== 00:12:33.093 [2024-12-06T16:28:14.932Z] Total : 8821.94 34.46 0.00 0.00 14432.50 8689.59 24175.50 00:12:33.093 00:12:33.093 Latency(us) 00:12:33.093 [2024-12-06T16:28:14.932Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:33.093 Job: Nvme1n1 (Core Mask 0x80, workload: unmap, depth: 128, IO size: 4096) 00:12:33.093 Nvme1n1 : 1.01 6701.14 26.18 0.00 0.00 19046.45 4199.16 44079.03 00:12:33.093 [2024-12-06T16:28:14.932Z] =================================================================================================================== 00:12:33.093 [2024-12-06T16:28:14.932Z] Total : 6701.14 26.18 0.00 0.00 19046.45 4199.16 44079.03 00:12:33.093 17:28:14 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@38 -- # wait 150199 00:12:33.355 17:28:14 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@39 -- # wait 150202 00:12:33.355 17:28:14 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@40 -- # wait 150206 00:12:33.355 17:28:14 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@42 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:12:33.355 17:28:14 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:33.355 17:28:14 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:12:33.355 17:28:14 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:33.355 17:28:14 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@44 -- # trap - SIGINT SIGTERM EXIT 00:12:33.355 17:28:14 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@46 -- # nvmftestfini 00:12:33.355 17:28:14 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@516 -- # nvmfcleanup 00:12:33.355 17:28:14 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@121 -- # sync 00:12:33.355 17:28:14 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:12:33.355 17:28:14 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@124 -- # set +e 00:12:33.355 17:28:14 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@125 -- # for i in {1..20} 00:12:33.355 17:28:14 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:12:33.355 rmmod nvme_tcp 00:12:33.355 rmmod nvme_fabrics 00:12:33.355 rmmod nvme_keyring 00:12:33.355 17:28:14 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:12:33.355 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@128 -- # set -e 00:12:33.355 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@129 -- # return 0 00:12:33.355 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@517 -- # '[' -n 150130 ']' 00:12:33.355 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@518 -- # killprocess 150130 00:12:33.355 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@954 -- # '[' -z 150130 ']' 00:12:33.355 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@958 -- # kill -0 150130 00:12:33.355 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@959 -- # uname 00:12:33.355 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:33.355 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 150130 00:12:33.355 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:33.355 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:33.355 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@972 -- # echo 'killing process with pid 150130' 00:12:33.355 killing process with pid 150130 00:12:33.355 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@973 -- # kill 150130 00:12:33.355 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@978 -- # wait 150130 00:12:33.621 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:12:33.621 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:12:33.621 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:12:33.621 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@297 -- # iptr 00:12:33.621 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@791 -- # iptables-save 00:12:33.621 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:12:33.621 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@791 -- # iptables-restore 00:12:33.621 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:12:33.621 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@302 -- # remove_spdk_ns 00:12:33.621 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:33.621 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:33.621 17:28:15 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:35.612 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:12:35.612 00:12:35.612 real 0m7.135s 00:12:35.612 user 0m15.653s 00:12:35.612 sys 0m3.439s 00:12:35.612 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:35.612 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:12:35.612 ************************************ 00:12:35.612 END TEST nvmf_bdev_io_wait 00:12:35.612 ************************************ 00:12:35.612 17:28:17 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@30 -- # run_test nvmf_queue_depth /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/queue_depth.sh --transport=tcp 00:12:35.612 17:28:17 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:12:35.612 17:28:17 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:35.612 17:28:17 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:12:35.612 ************************************ 00:12:35.612 START TEST nvmf_queue_depth 00:12:35.612 ************************************ 00:12:35.612 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/queue_depth.sh --transport=tcp 00:12:35.612 * Looking for test storage... 00:12:35.612 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:12:35.612 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:12:35.612 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1711 -- # lcov --version 00:12:35.612 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:12:35.880 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@333 -- # local ver1 ver1_l 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@334 -- # local ver2 ver2_l 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@336 -- # IFS=.-: 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@336 -- # read -ra ver1 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@337 -- # IFS=.-: 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@337 -- # read -ra ver2 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@338 -- # local 'op=<' 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@340 -- # ver1_l=2 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@341 -- # ver2_l=1 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@344 -- # case "$op" in 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@345 -- # : 1 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@364 -- # (( v = 0 )) 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@365 -- # decimal 1 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@353 -- # local d=1 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@355 -- # echo 1 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@365 -- # ver1[v]=1 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@366 -- # decimal 2 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@353 -- # local d=2 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@355 -- # echo 2 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@366 -- # ver2[v]=2 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@368 -- # return 0 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:12:35.881 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:35.881 --rc genhtml_branch_coverage=1 00:12:35.881 --rc genhtml_function_coverage=1 00:12:35.881 --rc genhtml_legend=1 00:12:35.881 --rc geninfo_all_blocks=1 00:12:35.881 --rc geninfo_unexecuted_blocks=1 00:12:35.881 00:12:35.881 ' 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:12:35.881 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:35.881 --rc genhtml_branch_coverage=1 00:12:35.881 --rc genhtml_function_coverage=1 00:12:35.881 --rc genhtml_legend=1 00:12:35.881 --rc geninfo_all_blocks=1 00:12:35.881 --rc geninfo_unexecuted_blocks=1 00:12:35.881 00:12:35.881 ' 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:12:35.881 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:35.881 --rc genhtml_branch_coverage=1 00:12:35.881 --rc genhtml_function_coverage=1 00:12:35.881 --rc genhtml_legend=1 00:12:35.881 --rc geninfo_all_blocks=1 00:12:35.881 --rc geninfo_unexecuted_blocks=1 00:12:35.881 00:12:35.881 ' 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:12:35.881 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:35.881 --rc genhtml_branch_coverage=1 00:12:35.881 --rc genhtml_function_coverage=1 00:12:35.881 --rc genhtml_legend=1 00:12:35.881 --rc geninfo_all_blocks=1 00:12:35.881 --rc geninfo_unexecuted_blocks=1 00:12:35.881 00:12:35.881 ' 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@12 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@7 -- # uname -s 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@15 -- # shopt -s extglob 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- paths/export.sh@5 -- # export PATH 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@51 -- # : 0 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:12:35.881 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@55 -- # have_pci_nics=0 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@14 -- # MALLOC_BDEV_SIZE=64 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@15 -- # MALLOC_BLOCK_SIZE=512 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@17 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:12:35.881 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@19 -- # nvmftestinit 00:12:35.882 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:12:35.882 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:12:35.882 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@476 -- # prepare_net_devs 00:12:35.882 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@438 -- # local -g is_hw=no 00:12:35.882 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@440 -- # remove_spdk_ns 00:12:35.882 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:35.882 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:35.882 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:35.882 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:12:35.882 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:12:35.882 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@309 -- # xtrace_disable 00:12:35.882 17:28:17 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@315 -- # pci_devs=() 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@315 -- # local -a pci_devs 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@316 -- # pci_net_devs=() 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@317 -- # pci_drivers=() 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@317 -- # local -A pci_drivers 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@319 -- # net_devs=() 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@319 -- # local -ga net_devs 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@320 -- # e810=() 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@320 -- # local -ga e810 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@321 -- # x722=() 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@321 -- # local -ga x722 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@322 -- # mlx=() 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@322 -- # local -ga mlx 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:12:37.883 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:12:37.883 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:37.883 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@418 -- # [[ up == up ]] 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:12:37.884 Found net devices under 0000:0a:00.0: cvl_0_0 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@418 -- # [[ up == up ]] 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:12:37.884 Found net devices under 0000:0a:00.1: cvl_0_1 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@442 -- # is_hw=yes 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:12:37.884 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:12:38.184 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:12:38.184 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:12:38.184 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.137 ms 00:12:38.184 00:12:38.184 --- 10.0.0.2 ping statistics --- 00:12:38.185 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:38.185 rtt min/avg/max/mdev = 0.137/0.137/0.137/0.000 ms 00:12:38.185 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:12:38.185 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:12:38.185 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.109 ms 00:12:38.185 00:12:38.185 --- 10.0.0.1 ping statistics --- 00:12:38.185 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:38.185 rtt min/avg/max/mdev = 0.109/0.109/0.109/0.000 ms 00:12:38.185 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:12:38.185 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@450 -- # return 0 00:12:38.185 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:12:38.185 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:12:38.185 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:12:38.185 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:12:38.185 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:12:38.185 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:12:38.185 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:12:38.185 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@21 -- # nvmfappstart -m 0x2 00:12:38.185 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:12:38.185 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@726 -- # xtrace_disable 00:12:38.185 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:12:38.185 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@509 -- # nvmfpid=152415 00:12:38.185 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:12:38.185 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@510 -- # waitforlisten 152415 00:12:38.185 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@835 -- # '[' -z 152415 ']' 00:12:38.185 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:38.185 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:38.185 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:38.185 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:38.185 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:38.185 17:28:19 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:12:38.185 [2024-12-06 17:28:19.801409] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:12:38.185 [2024-12-06 17:28:19.801499] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:12:38.185 [2024-12-06 17:28:19.878443] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:38.185 [2024-12-06 17:28:19.920257] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:12:38.185 [2024-12-06 17:28:19.920317] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:12:38.185 [2024-12-06 17:28:19.920346] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:12:38.185 [2024-12-06 17:28:19.920357] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:12:38.185 [2024-12-06 17:28:19.920367] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:12:38.185 [2024-12-06 17:28:19.920944] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@868 -- # return 0 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@732 -- # xtrace_disable 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@23 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:12:38.475 [2024-12-06 17:28:20.055898] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@24 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:12:38.475 Malloc0 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@25 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@26 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:12:38.475 [2024-12-06 17:28:20.104541] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@30 -- # bdevperf_pid=152552 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -z -r /var/tmp/bdevperf.sock -q 1024 -o 4096 -w verify -t 10 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@32 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@33 -- # waitforlisten 152552 /var/tmp/bdevperf.sock 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@835 -- # '[' -z 152552 ']' 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:12:38.475 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:38.475 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:12:38.475 [2024-12-06 17:28:20.152816] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:12:38.475 [2024-12-06 17:28:20.152882] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid152552 ] 00:12:38.475 [2024-12-06 17:28:20.223030] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:38.475 [2024-12-06 17:28:20.269139] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:38.750 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:38.750 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@868 -- # return 0 00:12:38.750 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@34 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:12:38.750 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:38.750 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:12:39.031 NVMe0n1 00:12:39.031 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:39.031 17:28:20 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:12:39.031 Running I/O for 10 seconds... 00:12:41.089 8192.00 IOPS, 32.00 MiB/s [2024-12-06T16:28:23.925Z] 8482.00 IOPS, 33.13 MiB/s [2024-12-06T16:28:24.923Z] 8531.33 IOPS, 33.33 MiB/s [2024-12-06T16:28:25.908Z] 8573.50 IOPS, 33.49 MiB/s [2024-12-06T16:28:26.865Z] 8600.80 IOPS, 33.60 MiB/s [2024-12-06T16:28:27.802Z] 8663.17 IOPS, 33.84 MiB/s [2024-12-06T16:28:29.181Z] 8635.14 IOPS, 33.73 MiB/s [2024-12-06T16:28:30.116Z] 8697.50 IOPS, 33.97 MiB/s [2024-12-06T16:28:31.052Z] 8705.00 IOPS, 34.00 MiB/s [2024-12-06T16:28:31.052Z] 8699.30 IOPS, 33.98 MiB/s 00:12:49.213 Latency(us) 00:12:49.213 [2024-12-06T16:28:31.052Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:49.213 Job: NVMe0n1 (Core Mask 0x1, workload: verify, depth: 1024, IO size: 4096) 00:12:49.213 Verification LBA range: start 0x0 length 0x4000 00:12:49.213 NVMe0n1 : 10.08 8728.56 34.10 0.00 0.00 116880.14 21165.70 71846.87 00:12:49.213 [2024-12-06T16:28:31.052Z] =================================================================================================================== 00:12:49.213 [2024-12-06T16:28:31.052Z] Total : 8728.56 34.10 0.00 0.00 116880.14 21165.70 71846.87 00:12:49.213 { 00:12:49.213 "results": [ 00:12:49.213 { 00:12:49.213 "job": "NVMe0n1", 00:12:49.213 "core_mask": "0x1", 00:12:49.213 "workload": "verify", 00:12:49.213 "status": "finished", 00:12:49.213 "verify_range": { 00:12:49.213 "start": 0, 00:12:49.213 "length": 16384 00:12:49.213 }, 00:12:49.213 "queue_depth": 1024, 00:12:49.213 "io_size": 4096, 00:12:49.213 "runtime": 10.082078, 00:12:49.213 "iops": 8728.557743750842, 00:12:49.213 "mibps": 34.095928686526726, 00:12:49.213 "io_failed": 0, 00:12:49.213 "io_timeout": 0, 00:12:49.213 "avg_latency_us": 116880.13544688799, 00:12:49.213 "min_latency_us": 21165.70074074074, 00:12:49.213 "max_latency_us": 71846.87407407408 00:12:49.213 } 00:12:49.213 ], 00:12:49.213 "core_count": 1 00:12:49.213 } 00:12:49.213 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@39 -- # killprocess 152552 00:12:49.213 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@954 -- # '[' -z 152552 ']' 00:12:49.213 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@958 -- # kill -0 152552 00:12:49.213 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@959 -- # uname 00:12:49.213 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:49.213 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 152552 00:12:49.213 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:12:49.213 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:12:49.213 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@972 -- # echo 'killing process with pid 152552' 00:12:49.213 killing process with pid 152552 00:12:49.213 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@973 -- # kill 152552 00:12:49.213 Received shutdown signal, test time was about 10.000000 seconds 00:12:49.213 00:12:49.213 Latency(us) 00:12:49.213 [2024-12-06T16:28:31.052Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:49.213 [2024-12-06T16:28:31.052Z] =================================================================================================================== 00:12:49.213 [2024-12-06T16:28:31.052Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:12:49.213 17:28:30 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@978 -- # wait 152552 00:12:49.472 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@41 -- # trap - SIGINT SIGTERM EXIT 00:12:49.472 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@43 -- # nvmftestfini 00:12:49.472 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@516 -- # nvmfcleanup 00:12:49.472 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@121 -- # sync 00:12:49.472 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:12:49.472 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@124 -- # set +e 00:12:49.472 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@125 -- # for i in {1..20} 00:12:49.472 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:12:49.472 rmmod nvme_tcp 00:12:49.472 rmmod nvme_fabrics 00:12:49.472 rmmod nvme_keyring 00:12:49.472 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:12:49.472 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@128 -- # set -e 00:12:49.472 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@129 -- # return 0 00:12:49.472 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@517 -- # '[' -n 152415 ']' 00:12:49.472 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@518 -- # killprocess 152415 00:12:49.472 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@954 -- # '[' -z 152415 ']' 00:12:49.472 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@958 -- # kill -0 152415 00:12:49.472 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@959 -- # uname 00:12:49.472 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:12:49.472 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 152415 00:12:49.472 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:12:49.472 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:12:49.472 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@972 -- # echo 'killing process with pid 152415' 00:12:49.472 killing process with pid 152415 00:12:49.472 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@973 -- # kill 152415 00:12:49.472 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@978 -- # wait 152415 00:12:49.733 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:12:49.733 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:12:49.733 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:12:49.733 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@297 -- # iptr 00:12:49.733 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@791 -- # iptables-save 00:12:49.733 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:12:49.733 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@791 -- # iptables-restore 00:12:49.733 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:12:49.733 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@302 -- # remove_spdk_ns 00:12:49.733 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:49.733 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:49.733 17:28:31 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:51.641 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:12:51.641 00:12:51.641 real 0m16.109s 00:12:51.641 user 0m22.633s 00:12:51.641 sys 0m3.102s 00:12:51.641 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:51.641 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:12:51.641 ************************************ 00:12:51.641 END TEST nvmf_queue_depth 00:12:51.641 ************************************ 00:12:51.641 17:28:33 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@31 -- # run_test nvmf_target_multipath /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multipath.sh --transport=tcp 00:12:51.641 17:28:33 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:12:51.641 17:28:33 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:51.641 17:28:33 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:12:51.901 ************************************ 00:12:51.901 START TEST nvmf_target_multipath 00:12:51.901 ************************************ 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multipath.sh --transport=tcp 00:12:51.901 * Looking for test storage... 00:12:51.901 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1711 -- # lcov --version 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@333 -- # local ver1 ver1_l 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@334 -- # local ver2 ver2_l 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@336 -- # IFS=.-: 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@336 -- # read -ra ver1 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@337 -- # IFS=.-: 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@337 -- # read -ra ver2 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@338 -- # local 'op=<' 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@340 -- # ver1_l=2 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@341 -- # ver2_l=1 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@344 -- # case "$op" in 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@345 -- # : 1 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@364 -- # (( v = 0 )) 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@365 -- # decimal 1 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@353 -- # local d=1 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@355 -- # echo 1 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@365 -- # ver1[v]=1 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@366 -- # decimal 2 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@353 -- # local d=2 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@355 -- # echo 2 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@366 -- # ver2[v]=2 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@368 -- # return 0 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:12:51.901 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:51.901 --rc genhtml_branch_coverage=1 00:12:51.901 --rc genhtml_function_coverage=1 00:12:51.901 --rc genhtml_legend=1 00:12:51.901 --rc geninfo_all_blocks=1 00:12:51.901 --rc geninfo_unexecuted_blocks=1 00:12:51.901 00:12:51.901 ' 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:12:51.901 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:51.901 --rc genhtml_branch_coverage=1 00:12:51.901 --rc genhtml_function_coverage=1 00:12:51.901 --rc genhtml_legend=1 00:12:51.901 --rc geninfo_all_blocks=1 00:12:51.901 --rc geninfo_unexecuted_blocks=1 00:12:51.901 00:12:51.901 ' 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:12:51.901 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:51.901 --rc genhtml_branch_coverage=1 00:12:51.901 --rc genhtml_function_coverage=1 00:12:51.901 --rc genhtml_legend=1 00:12:51.901 --rc geninfo_all_blocks=1 00:12:51.901 --rc geninfo_unexecuted_blocks=1 00:12:51.901 00:12:51.901 ' 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:12:51.901 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:51.901 --rc genhtml_branch_coverage=1 00:12:51.901 --rc genhtml_function_coverage=1 00:12:51.901 --rc genhtml_legend=1 00:12:51.901 --rc geninfo_all_blocks=1 00:12:51.901 --rc geninfo_unexecuted_blocks=1 00:12:51.901 00:12:51.901 ' 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@7 -- # uname -s 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@15 -- # shopt -s extglob 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:12:51.901 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- paths/export.sh@5 -- # export PATH 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@51 -- # : 0 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:12:51.902 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@55 -- # have_pci_nics=0 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@11 -- # MALLOC_BDEV_SIZE=64 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@13 -- # nqn=nqn.2016-06.io.spdk:cnode1 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@15 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@43 -- # nvmftestinit 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@476 -- # prepare_net_devs 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@438 -- # local -g is_hw=no 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@440 -- # remove_spdk_ns 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@309 -- # xtrace_disable 00:12:51.902 17:28:33 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@10 -- # set +x 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@315 -- # pci_devs=() 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@315 -- # local -a pci_devs 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@316 -- # pci_net_devs=() 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@317 -- # pci_drivers=() 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@317 -- # local -A pci_drivers 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@319 -- # net_devs=() 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@319 -- # local -ga net_devs 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@320 -- # e810=() 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@320 -- # local -ga e810 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@321 -- # x722=() 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@321 -- # local -ga x722 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@322 -- # mlx=() 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@322 -- # local -ga mlx 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:12:54.434 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:12:54.434 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@418 -- # [[ up == up ]] 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:12:54.434 Found net devices under 0000:0a:00.0: cvl_0_0 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@418 -- # [[ up == up ]] 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:12:54.434 Found net devices under 0000:0a:00.1: cvl_0_1 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@442 -- # is_hw=yes 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:12:54.434 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:12:54.434 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.295 ms 00:12:54.434 00:12:54.434 --- 10.0.0.2 ping statistics --- 00:12:54.434 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:54.434 rtt min/avg/max/mdev = 0.295/0.295/0.295/0.000 ms 00:12:54.434 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:12:54.434 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:12:54.434 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.130 ms 00:12:54.434 00:12:54.434 --- 10.0.0.1 ping statistics --- 00:12:54.434 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:54.435 rtt min/avg/max/mdev = 0.130/0.130/0.130/0.000 ms 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@450 -- # return 0 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@45 -- # '[' -z ']' 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@46 -- # echo 'only one NIC for nvmf test' 00:12:54.435 only one NIC for nvmf test 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@47 -- # nvmftestfini 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@516 -- # nvmfcleanup 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@121 -- # sync 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@124 -- # set +e 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@125 -- # for i in {1..20} 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:12:54.435 rmmod nvme_tcp 00:12:54.435 rmmod nvme_fabrics 00:12:54.435 rmmod nvme_keyring 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@128 -- # set -e 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@129 -- # return 0 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@517 -- # '[' -n '' ']' 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@297 -- # iptr 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@791 -- # iptables-save 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@791 -- # iptables-restore 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@302 -- # remove_spdk_ns 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:54.435 17:28:35 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:56.345 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:12:56.345 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@48 -- # exit 0 00:12:56.345 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@1 -- # nvmftestfini 00:12:56.345 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@516 -- # nvmfcleanup 00:12:56.345 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@121 -- # sync 00:12:56.345 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:12:56.345 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@124 -- # set +e 00:12:56.345 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@125 -- # for i in {1..20} 00:12:56.345 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:12:56.345 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:12:56.345 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@128 -- # set -e 00:12:56.345 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@129 -- # return 0 00:12:56.345 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@517 -- # '[' -n '' ']' 00:12:56.345 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:12:56.345 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:12:56.345 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:12:56.345 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@297 -- # iptr 00:12:56.345 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@791 -- # iptables-save 00:12:56.345 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:12:56.345 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@791 -- # iptables-restore 00:12:56.345 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:12:56.345 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@302 -- # remove_spdk_ns 00:12:56.345 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:56.345 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:56.345 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:56.345 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:12:56.345 00:12:56.345 real 0m4.555s 00:12:56.345 user 0m0.937s 00:12:56.345 sys 0m1.626s 00:12:56.345 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1130 -- # xtrace_disable 00:12:56.345 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@10 -- # set +x 00:12:56.345 ************************************ 00:12:56.345 END TEST nvmf_target_multipath 00:12:56.345 ************************************ 00:12:56.346 17:28:38 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@32 -- # run_test nvmf_zcopy /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh --transport=tcp 00:12:56.346 17:28:38 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:12:56.346 17:28:38 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1111 -- # xtrace_disable 00:12:56.346 17:28:38 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:12:56.346 ************************************ 00:12:56.346 START TEST nvmf_zcopy 00:12:56.346 ************************************ 00:12:56.346 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh --transport=tcp 00:12:56.346 * Looking for test storage... 00:12:56.346 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:12:56.346 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:12:56.346 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1711 -- # lcov --version 00:12:56.346 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@333 -- # local ver1 ver1_l 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@334 -- # local ver2 ver2_l 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@336 -- # IFS=.-: 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@336 -- # read -ra ver1 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@337 -- # IFS=.-: 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@337 -- # read -ra ver2 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@338 -- # local 'op=<' 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@340 -- # ver1_l=2 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@341 -- # ver2_l=1 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@344 -- # case "$op" in 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@345 -- # : 1 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@364 -- # (( v = 0 )) 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@365 -- # decimal 1 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@353 -- # local d=1 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@355 -- # echo 1 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@365 -- # ver1[v]=1 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@366 -- # decimal 2 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@353 -- # local d=2 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@355 -- # echo 2 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@366 -- # ver2[v]=2 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@368 -- # return 0 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:12:56.605 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:56.605 --rc genhtml_branch_coverage=1 00:12:56.605 --rc genhtml_function_coverage=1 00:12:56.605 --rc genhtml_legend=1 00:12:56.605 --rc geninfo_all_blocks=1 00:12:56.605 --rc geninfo_unexecuted_blocks=1 00:12:56.605 00:12:56.605 ' 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:12:56.605 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:56.605 --rc genhtml_branch_coverage=1 00:12:56.605 --rc genhtml_function_coverage=1 00:12:56.605 --rc genhtml_legend=1 00:12:56.605 --rc geninfo_all_blocks=1 00:12:56.605 --rc geninfo_unexecuted_blocks=1 00:12:56.605 00:12:56.605 ' 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:12:56.605 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:56.605 --rc genhtml_branch_coverage=1 00:12:56.605 --rc genhtml_function_coverage=1 00:12:56.605 --rc genhtml_legend=1 00:12:56.605 --rc geninfo_all_blocks=1 00:12:56.605 --rc geninfo_unexecuted_blocks=1 00:12:56.605 00:12:56.605 ' 00:12:56.605 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:12:56.606 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:56.606 --rc genhtml_branch_coverage=1 00:12:56.606 --rc genhtml_function_coverage=1 00:12:56.606 --rc genhtml_legend=1 00:12:56.606 --rc geninfo_all_blocks=1 00:12:56.606 --rc geninfo_unexecuted_blocks=1 00:12:56.606 00:12:56.606 ' 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@7 -- # uname -s 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@15 -- # shopt -s extglob 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- paths/export.sh@5 -- # export PATH 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@51 -- # : 0 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:12:56.606 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@55 -- # have_pci_nics=0 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@12 -- # nvmftestinit 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@476 -- # prepare_net_devs 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@438 -- # local -g is_hw=no 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@440 -- # remove_spdk_ns 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@309 -- # xtrace_disable 00:12:56.606 17:28:38 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:12:59.143 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:12:59.143 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@315 -- # pci_devs=() 00:12:59.143 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@315 -- # local -a pci_devs 00:12:59.143 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@316 -- # pci_net_devs=() 00:12:59.143 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:12:59.143 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@317 -- # pci_drivers=() 00:12:59.143 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@317 -- # local -A pci_drivers 00:12:59.143 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@319 -- # net_devs=() 00:12:59.143 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@319 -- # local -ga net_devs 00:12:59.143 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@320 -- # e810=() 00:12:59.143 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@320 -- # local -ga e810 00:12:59.143 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@321 -- # x722=() 00:12:59.143 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@321 -- # local -ga x722 00:12:59.143 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@322 -- # mlx=() 00:12:59.143 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@322 -- # local -ga mlx 00:12:59.143 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:12:59.143 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:12:59.143 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:12:59.143 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:12:59.143 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:12:59.143 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:12:59.143 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:12:59.143 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:12:59.143 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:12:59.143 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:12:59.143 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:12:59.144 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:12:59.144 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@418 -- # [[ up == up ]] 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:12:59.144 Found net devices under 0000:0a:00.0: cvl_0_0 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@418 -- # [[ up == up ]] 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:12:59.144 Found net devices under 0000:0a:00.1: cvl_0_1 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@442 -- # is_hw=yes 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:12:59.144 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:12:59.144 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.182 ms 00:12:59.144 00:12:59.144 --- 10.0.0.2 ping statistics --- 00:12:59.144 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:59.144 rtt min/avg/max/mdev = 0.182/0.182/0.182/0.000 ms 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:12:59.144 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:12:59.144 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.087 ms 00:12:59.144 00:12:59.144 --- 10.0.0.1 ping statistics --- 00:12:59.144 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:59.144 rtt min/avg/max/mdev = 0.087/0.087/0.087/0.000 ms 00:12:59.144 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@450 -- # return 0 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@13 -- # nvmfappstart -m 0x2 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@726 -- # xtrace_disable 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@509 -- # nvmfpid=157798 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@510 -- # waitforlisten 157798 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@835 -- # '[' -z 157798 ']' 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@840 -- # local max_retries=100 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:59.145 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@844 -- # xtrace_disable 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:12:59.145 [2024-12-06 17:28:40.674312] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:12:59.145 [2024-12-06 17:28:40.674397] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:12:59.145 [2024-12-06 17:28:40.748640] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:59.145 [2024-12-06 17:28:40.791170] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:12:59.145 [2024-12-06 17:28:40.791232] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:12:59.145 [2024-12-06 17:28:40.791260] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:12:59.145 [2024-12-06 17:28:40.791271] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:12:59.145 [2024-12-06 17:28:40.791281] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:12:59.145 [2024-12-06 17:28:40.791885] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@868 -- # return 0 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@732 -- # xtrace_disable 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@15 -- # '[' tcp '!=' tcp ']' 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@22 -- # rpc_cmd nvmf_create_transport -t tcp -o -c 0 --zcopy 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:12:59.145 [2024-12-06 17:28:40.932676] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:12:59.145 [2024-12-06 17:28:40.948879] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@29 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc0 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:12:59.145 malloc0 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@30 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:12:59.145 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:12:59.404 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:12:59.404 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -t 10 -q 128 -w verify -o 8192 00:12:59.404 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@33 -- # gen_nvmf_target_json 00:12:59.404 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@560 -- # config=() 00:12:59.404 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@560 -- # local subsystem config 00:12:59.404 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:12:59.404 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:12:59.404 { 00:12:59.404 "params": { 00:12:59.404 "name": "Nvme$subsystem", 00:12:59.404 "trtype": "$TEST_TRANSPORT", 00:12:59.404 "traddr": "$NVMF_FIRST_TARGET_IP", 00:12:59.404 "adrfam": "ipv4", 00:12:59.404 "trsvcid": "$NVMF_PORT", 00:12:59.404 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:12:59.404 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:12:59.404 "hdgst": ${hdgst:-false}, 00:12:59.404 "ddgst": ${ddgst:-false} 00:12:59.404 }, 00:12:59.404 "method": "bdev_nvme_attach_controller" 00:12:59.404 } 00:12:59.404 EOF 00:12:59.404 )") 00:12:59.404 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@582 -- # cat 00:12:59.404 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@584 -- # jq . 00:12:59.404 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@585 -- # IFS=, 00:12:59.404 17:28:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:12:59.404 "params": { 00:12:59.404 "name": "Nvme1", 00:12:59.404 "trtype": "tcp", 00:12:59.404 "traddr": "10.0.0.2", 00:12:59.404 "adrfam": "ipv4", 00:12:59.404 "trsvcid": "4420", 00:12:59.404 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:12:59.404 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:12:59.404 "hdgst": false, 00:12:59.404 "ddgst": false 00:12:59.404 }, 00:12:59.404 "method": "bdev_nvme_attach_controller" 00:12:59.404 }' 00:12:59.404 [2024-12-06 17:28:41.032076] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:12:59.404 [2024-12-06 17:28:41.032142] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid157823 ] 00:12:59.404 [2024-12-06 17:28:41.099309] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:59.404 [2024-12-06 17:28:41.147049] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:59.663 Running I/O for 10 seconds... 00:13:01.533 5867.00 IOPS, 45.84 MiB/s [2024-12-06T16:28:44.749Z] 5938.50 IOPS, 46.39 MiB/s [2024-12-06T16:28:45.684Z] 5973.33 IOPS, 46.67 MiB/s [2024-12-06T16:28:46.620Z] 5978.50 IOPS, 46.71 MiB/s [2024-12-06T16:28:47.555Z] 5980.60 IOPS, 46.72 MiB/s [2024-12-06T16:28:48.497Z] 5992.50 IOPS, 46.82 MiB/s [2024-12-06T16:28:49.431Z] 6002.43 IOPS, 46.89 MiB/s [2024-12-06T16:28:50.363Z] 6008.88 IOPS, 46.94 MiB/s [2024-12-06T16:28:51.755Z] 6007.11 IOPS, 46.93 MiB/s [2024-12-06T16:28:51.755Z] 6012.20 IOPS, 46.97 MiB/s 00:13:09.916 Latency(us) 00:13:09.916 [2024-12-06T16:28:51.755Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:09.916 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 8192) 00:13:09.916 Verification LBA range: start 0x0 length 0x1000 00:13:09.916 Nvme1n1 : 10.01 6016.50 47.00 0.00 0.00 21216.31 3543.80 29127.11 00:13:09.916 [2024-12-06T16:28:51.755Z] =================================================================================================================== 00:13:09.916 [2024-12-06T16:28:51.755Z] Total : 6016.50 47.00 0.00 0.00 21216.31 3543.80 29127.11 00:13:09.916 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@39 -- # perfpid=159024 00:13:09.916 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@41 -- # xtrace_disable 00:13:09.916 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:13:09.916 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/63 -t 5 -q 128 -w randrw -M 50 -o 8192 00:13:09.916 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@37 -- # gen_nvmf_target_json 00:13:09.916 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@560 -- # config=() 00:13:09.916 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@560 -- # local subsystem config 00:13:09.916 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:13:09.916 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:13:09.916 { 00:13:09.916 "params": { 00:13:09.916 "name": "Nvme$subsystem", 00:13:09.916 "trtype": "$TEST_TRANSPORT", 00:13:09.916 "traddr": "$NVMF_FIRST_TARGET_IP", 00:13:09.916 "adrfam": "ipv4", 00:13:09.916 "trsvcid": "$NVMF_PORT", 00:13:09.916 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:13:09.916 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:13:09.916 "hdgst": ${hdgst:-false}, 00:13:09.916 "ddgst": ${ddgst:-false} 00:13:09.916 }, 00:13:09.916 "method": "bdev_nvme_attach_controller" 00:13:09.916 } 00:13:09.916 EOF 00:13:09.916 )") 00:13:09.916 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@582 -- # cat 00:13:09.916 [2024-12-06 17:28:51.579126] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:09.916 [2024-12-06 17:28:51.579173] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:09.916 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@584 -- # jq . 00:13:09.916 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@585 -- # IFS=, 00:13:09.916 17:28:51 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:13:09.916 "params": { 00:13:09.916 "name": "Nvme1", 00:13:09.916 "trtype": "tcp", 00:13:09.916 "traddr": "10.0.0.2", 00:13:09.916 "adrfam": "ipv4", 00:13:09.916 "trsvcid": "4420", 00:13:09.916 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:13:09.916 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:13:09.916 "hdgst": false, 00:13:09.916 "ddgst": false 00:13:09.916 }, 00:13:09.916 "method": "bdev_nvme_attach_controller" 00:13:09.916 }' 00:13:09.916 [2024-12-06 17:28:51.587094] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:09.916 [2024-12-06 17:28:51.587115] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:09.916 [2024-12-06 17:28:51.595102] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:09.916 [2024-12-06 17:28:51.595122] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:09.916 [2024-12-06 17:28:51.603124] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:09.916 [2024-12-06 17:28:51.603143] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:09.916 [2024-12-06 17:28:51.611149] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:09.916 [2024-12-06 17:28:51.611170] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:09.916 [2024-12-06 17:28:51.615464] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:13:09.916 [2024-12-06 17:28:51.615533] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid159024 ] 00:13:09.916 [2024-12-06 17:28:51.619174] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:09.916 [2024-12-06 17:28:51.619193] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:09.916 [2024-12-06 17:28:51.627197] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:09.916 [2024-12-06 17:28:51.627218] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:09.916 [2024-12-06 17:28:51.635220] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:09.916 [2024-12-06 17:28:51.635240] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:09.916 [2024-12-06 17:28:51.643239] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:09.916 [2024-12-06 17:28:51.643269] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:09.916 [2024-12-06 17:28:51.651261] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:09.916 [2024-12-06 17:28:51.651281] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:09.916 [2024-12-06 17:28:51.659284] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:09.916 [2024-12-06 17:28:51.659305] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:09.916 [2024-12-06 17:28:51.667302] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:09.916 [2024-12-06 17:28:51.667322] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:09.916 [2024-12-06 17:28:51.675324] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:09.916 [2024-12-06 17:28:51.675353] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:09.916 [2024-12-06 17:28:51.683345] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:09.916 [2024-12-06 17:28:51.683365] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:09.916 [2024-12-06 17:28:51.685571] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:09.916 [2024-12-06 17:28:51.691399] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:09.916 [2024-12-06 17:28:51.691425] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:09.916 [2024-12-06 17:28:51.699439] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:09.916 [2024-12-06 17:28:51.699479] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:09.916 [2024-12-06 17:28:51.707416] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:09.916 [2024-12-06 17:28:51.707437] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:09.916 [2024-12-06 17:28:51.715435] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:09.916 [2024-12-06 17:28:51.715457] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:09.916 [2024-12-06 17:28:51.723457] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:09.916 [2024-12-06 17:28:51.723477] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:09.916 [2024-12-06 17:28:51.731474] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:09.916 [2024-12-06 17:28:51.731494] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:09.916 [2024-12-06 17:28:51.733443] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:09.916 [2024-12-06 17:28:51.739496] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:09.916 [2024-12-06 17:28:51.739526] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:09.916 [2024-12-06 17:28:51.747530] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:09.916 [2024-12-06 17:28:51.747555] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.175 [2024-12-06 17:28:51.755581] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.175 [2024-12-06 17:28:51.755625] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.175 [2024-12-06 17:28:51.763601] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.175 [2024-12-06 17:28:51.763659] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.175 [2024-12-06 17:28:51.771621] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.175 [2024-12-06 17:28:51.771691] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.175 [2024-12-06 17:28:51.779658] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.175 [2024-12-06 17:28:51.779704] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.175 [2024-12-06 17:28:51.787687] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.175 [2024-12-06 17:28:51.787739] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.175 [2024-12-06 17:28:51.795710] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.175 [2024-12-06 17:28:51.795757] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.175 [2024-12-06 17:28:51.803699] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.175 [2024-12-06 17:28:51.803732] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.175 [2024-12-06 17:28:51.811776] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.175 [2024-12-06 17:28:51.811817] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.175 [2024-12-06 17:28:51.819792] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.175 [2024-12-06 17:28:51.819833] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.175 [2024-12-06 17:28:51.827793] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.175 [2024-12-06 17:28:51.827835] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.175 [2024-12-06 17:28:51.835782] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.175 [2024-12-06 17:28:51.835804] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.175 [2024-12-06 17:28:51.843800] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.175 [2024-12-06 17:28:51.843821] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.175 [2024-12-06 17:28:51.852222] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.175 [2024-12-06 17:28:51.852247] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.175 [2024-12-06 17:28:51.860251] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.175 [2024-12-06 17:28:51.860273] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.175 [2024-12-06 17:28:51.868273] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.175 [2024-12-06 17:28:51.868295] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.175 [2024-12-06 17:28:51.876295] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.175 [2024-12-06 17:28:51.876317] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.175 [2024-12-06 17:28:51.884314] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.175 [2024-12-06 17:28:51.884335] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.175 [2024-12-06 17:28:51.892335] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.175 [2024-12-06 17:28:51.892355] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.175 [2024-12-06 17:28:51.900356] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.175 [2024-12-06 17:28:51.900375] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.175 [2024-12-06 17:28:51.908377] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.175 [2024-12-06 17:28:51.908396] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.175 [2024-12-06 17:28:51.916456] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.175 [2024-12-06 17:28:51.916478] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.175 [2024-12-06 17:28:51.924422] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.175 [2024-12-06 17:28:51.924443] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.175 [2024-12-06 17:28:51.932509] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.175 [2024-12-06 17:28:51.932533] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.175 [2024-12-06 17:28:51.940464] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.175 [2024-12-06 17:28:51.940485] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.175 [2024-12-06 17:28:51.948578] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.175 [2024-12-06 17:28:51.948605] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.175 [2024-12-06 17:28:51.989873] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.175 [2024-12-06 17:28:51.989900] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.175 [2024-12-06 17:28:51.996633] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.175 [2024-12-06 17:28:51.996678] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.175 Running I/O for 5 seconds... 00:13:10.175 [2024-12-06 17:28:52.004680] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.175 [2024-12-06 17:28:52.004701] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.433 [2024-12-06 17:28:52.018957] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.433 [2024-12-06 17:28:52.018988] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.433 [2024-12-06 17:28:52.029878] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.433 [2024-12-06 17:28:52.029907] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.433 [2024-12-06 17:28:52.041014] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.433 [2024-12-06 17:28:52.041042] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.433 [2024-12-06 17:28:52.051845] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.433 [2024-12-06 17:28:52.051872] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.433 [2024-12-06 17:28:52.062914] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.433 [2024-12-06 17:28:52.062942] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.433 [2024-12-06 17:28:52.074091] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.433 [2024-12-06 17:28:52.074120] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.433 [2024-12-06 17:28:52.085222] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.433 [2024-12-06 17:28:52.085249] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.433 [2024-12-06 17:28:52.096138] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.433 [2024-12-06 17:28:52.096164] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.433 [2024-12-06 17:28:52.109044] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.433 [2024-12-06 17:28:52.109086] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.434 [2024-12-06 17:28:52.119332] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.434 [2024-12-06 17:28:52.119374] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.434 [2024-12-06 17:28:52.129929] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.434 [2024-12-06 17:28:52.129956] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.434 [2024-12-06 17:28:52.140612] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.434 [2024-12-06 17:28:52.140637] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.434 [2024-12-06 17:28:52.150852] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.434 [2024-12-06 17:28:52.150879] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.434 [2024-12-06 17:28:52.161368] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.434 [2024-12-06 17:28:52.161395] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.434 [2024-12-06 17:28:52.171621] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.434 [2024-12-06 17:28:52.171662] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.434 [2024-12-06 17:28:52.182255] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.434 [2024-12-06 17:28:52.182288] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.434 [2024-12-06 17:28:52.193023] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.434 [2024-12-06 17:28:52.193049] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.434 [2024-12-06 17:28:52.206222] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.434 [2024-12-06 17:28:52.206248] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.434 [2024-12-06 17:28:52.216582] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.434 [2024-12-06 17:28:52.216608] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.434 [2024-12-06 17:28:52.227013] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.434 [2024-12-06 17:28:52.227039] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.434 [2024-12-06 17:28:52.238045] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.434 [2024-12-06 17:28:52.238071] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.434 [2024-12-06 17:28:52.250577] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.434 [2024-12-06 17:28:52.250604] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.434 [2024-12-06 17:28:52.260425] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.434 [2024-12-06 17:28:52.260451] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.434 [2024-12-06 17:28:52.270612] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.434 [2024-12-06 17:28:52.270638] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.692 [2024-12-06 17:28:52.280918] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.692 [2024-12-06 17:28:52.280946] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.692 [2024-12-06 17:28:52.291469] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.692 [2024-12-06 17:28:52.291495] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.692 [2024-12-06 17:28:52.302248] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.692 [2024-12-06 17:28:52.302274] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.692 [2024-12-06 17:28:52.314741] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.692 [2024-12-06 17:28:52.314769] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.692 [2024-12-06 17:28:52.324605] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.692 [2024-12-06 17:28:52.324631] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.692 [2024-12-06 17:28:52.336145] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.692 [2024-12-06 17:28:52.336171] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.692 [2024-12-06 17:28:52.348811] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.692 [2024-12-06 17:28:52.348839] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.692 [2024-12-06 17:28:52.360246] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.692 [2024-12-06 17:28:52.360273] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.692 [2024-12-06 17:28:52.369548] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.692 [2024-12-06 17:28:52.369574] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.692 [2024-12-06 17:28:52.381125] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.692 [2024-12-06 17:28:52.381151] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.692 [2024-12-06 17:28:52.393928] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.692 [2024-12-06 17:28:52.393963] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.692 [2024-12-06 17:28:52.403412] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.692 [2024-12-06 17:28:52.403438] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.692 [2024-12-06 17:28:52.414419] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.692 [2024-12-06 17:28:52.414445] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.692 [2024-12-06 17:28:52.424961] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.692 [2024-12-06 17:28:52.425002] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.692 [2024-12-06 17:28:52.435313] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.692 [2024-12-06 17:28:52.435339] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.692 [2024-12-06 17:28:52.445605] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.692 [2024-12-06 17:28:52.445631] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.692 [2024-12-06 17:28:52.455813] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.692 [2024-12-06 17:28:52.455840] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.692 [2024-12-06 17:28:52.466851] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.692 [2024-12-06 17:28:52.466878] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.692 [2024-12-06 17:28:52.479462] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.692 [2024-12-06 17:28:52.479487] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.692 [2024-12-06 17:28:52.489796] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.692 [2024-12-06 17:28:52.489823] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.692 [2024-12-06 17:28:52.500162] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.692 [2024-12-06 17:28:52.500188] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.692 [2024-12-06 17:28:52.511219] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.692 [2024-12-06 17:28:52.511260] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.692 [2024-12-06 17:28:52.524683] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.692 [2024-12-06 17:28:52.524710] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.950 [2024-12-06 17:28:52.534958] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.950 [2024-12-06 17:28:52.534993] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.950 [2024-12-06 17:28:52.545872] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.950 [2024-12-06 17:28:52.545898] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.950 [2024-12-06 17:28:52.558437] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.950 [2024-12-06 17:28:52.558463] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.950 [2024-12-06 17:28:52.568963] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.950 [2024-12-06 17:28:52.569006] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.950 [2024-12-06 17:28:52.579936] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.950 [2024-12-06 17:28:52.579962] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.950 [2024-12-06 17:28:52.592564] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.950 [2024-12-06 17:28:52.592590] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.950 [2024-12-06 17:28:52.602634] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.950 [2024-12-06 17:28:52.602690] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.950 [2024-12-06 17:28:52.613134] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.950 [2024-12-06 17:28:52.613160] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.950 [2024-12-06 17:28:52.623493] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.950 [2024-12-06 17:28:52.623519] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.950 [2024-12-06 17:28:52.634223] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.950 [2024-12-06 17:28:52.634249] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.950 [2024-12-06 17:28:52.644968] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.950 [2024-12-06 17:28:52.644994] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.950 [2024-12-06 17:28:52.657374] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.950 [2024-12-06 17:28:52.657401] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.950 [2024-12-06 17:28:52.667403] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.950 [2024-12-06 17:28:52.667429] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.950 [2024-12-06 17:28:52.677688] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.950 [2024-12-06 17:28:52.677715] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.950 [2024-12-06 17:28:52.688085] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.950 [2024-12-06 17:28:52.688111] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.950 [2024-12-06 17:28:52.698707] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.950 [2024-12-06 17:28:52.698734] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.950 [2024-12-06 17:28:52.711141] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.950 [2024-12-06 17:28:52.711168] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.950 [2024-12-06 17:28:52.720270] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.950 [2024-12-06 17:28:52.720296] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.950 [2024-12-06 17:28:52.733282] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.950 [2024-12-06 17:28:52.733308] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.950 [2024-12-06 17:28:52.743093] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.950 [2024-12-06 17:28:52.743120] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.950 [2024-12-06 17:28:52.753313] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.950 [2024-12-06 17:28:52.753339] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.950 [2024-12-06 17:28:52.764025] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.950 [2024-12-06 17:28:52.764051] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.950 [2024-12-06 17:28:52.774703] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.950 [2024-12-06 17:28:52.774732] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:10.950 [2024-12-06 17:28:52.785131] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:10.950 [2024-12-06 17:28:52.785159] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.208 [2024-12-06 17:28:52.795851] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.208 [2024-12-06 17:28:52.795879] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.208 [2024-12-06 17:28:52.806521] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.208 [2024-12-06 17:28:52.806548] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.208 [2024-12-06 17:28:52.818792] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.208 [2024-12-06 17:28:52.818820] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.208 [2024-12-06 17:28:52.829114] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.208 [2024-12-06 17:28:52.829140] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.208 [2024-12-06 17:28:52.839261] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.208 [2024-12-06 17:28:52.839287] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.208 [2024-12-06 17:28:52.849259] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.208 [2024-12-06 17:28:52.849286] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.208 [2024-12-06 17:28:52.859710] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.208 [2024-12-06 17:28:52.859738] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.208 [2024-12-06 17:28:52.870572] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.208 [2024-12-06 17:28:52.870599] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.208 [2024-12-06 17:28:52.883469] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.208 [2024-12-06 17:28:52.883496] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.208 [2024-12-06 17:28:52.893625] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.208 [2024-12-06 17:28:52.893675] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.208 [2024-12-06 17:28:52.903878] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.208 [2024-12-06 17:28:52.903905] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.208 [2024-12-06 17:28:52.914786] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.208 [2024-12-06 17:28:52.914815] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.208 [2024-12-06 17:28:52.926990] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.208 [2024-12-06 17:28:52.927016] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.208 [2024-12-06 17:28:52.936918] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.208 [2024-12-06 17:28:52.936945] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.208 [2024-12-06 17:28:52.949677] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.208 [2024-12-06 17:28:52.949705] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.208 [2024-12-06 17:28:52.961597] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.208 [2024-12-06 17:28:52.961624] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.208 [2024-12-06 17:28:52.970702] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.208 [2024-12-06 17:28:52.970728] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.208 [2024-12-06 17:28:52.982487] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.208 [2024-12-06 17:28:52.982513] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.208 [2024-12-06 17:28:52.995117] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.208 [2024-12-06 17:28:52.995143] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.208 [2024-12-06 17:28:53.007552] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.208 [2024-12-06 17:28:53.007577] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.208 11901.00 IOPS, 92.98 MiB/s [2024-12-06T16:28:53.047Z] [2024-12-06 17:28:53.018075] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.208 [2024-12-06 17:28:53.018101] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.208 [2024-12-06 17:28:53.028759] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.208 [2024-12-06 17:28:53.028788] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.208 [2024-12-06 17:28:53.039222] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.208 [2024-12-06 17:28:53.039248] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.466 [2024-12-06 17:28:53.049939] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.466 [2024-12-06 17:28:53.049966] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.466 [2024-12-06 17:28:53.062219] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.466 [2024-12-06 17:28:53.062245] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.466 [2024-12-06 17:28:53.072589] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.466 [2024-12-06 17:28:53.072615] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.466 [2024-12-06 17:28:53.082985] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.466 [2024-12-06 17:28:53.083027] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.466 [2024-12-06 17:28:53.093472] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.466 [2024-12-06 17:28:53.093498] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.466 [2024-12-06 17:28:53.104111] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.466 [2024-12-06 17:28:53.104137] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.466 [2024-12-06 17:28:53.114741] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.466 [2024-12-06 17:28:53.114770] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.466 [2024-12-06 17:28:53.125539] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.466 [2024-12-06 17:28:53.125566] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.466 [2024-12-06 17:28:53.135779] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.466 [2024-12-06 17:28:53.135806] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.466 [2024-12-06 17:28:53.146235] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.466 [2024-12-06 17:28:53.146261] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.466 [2024-12-06 17:28:53.156794] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.466 [2024-12-06 17:28:53.156821] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.466 [2024-12-06 17:28:53.167801] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.466 [2024-12-06 17:28:53.167828] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.466 [2024-12-06 17:28:53.180557] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.466 [2024-12-06 17:28:53.180583] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.466 [2024-12-06 17:28:53.190861] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.466 [2024-12-06 17:28:53.190888] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.466 [2024-12-06 17:28:53.201679] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.466 [2024-12-06 17:28:53.201706] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.466 [2024-12-06 17:28:53.214082] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.466 [2024-12-06 17:28:53.214117] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.466 [2024-12-06 17:28:53.223897] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.466 [2024-12-06 17:28:53.223924] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.466 [2024-12-06 17:28:53.234341] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.466 [2024-12-06 17:28:53.234367] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.466 [2024-12-06 17:28:53.244931] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.466 [2024-12-06 17:28:53.244957] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.466 [2024-12-06 17:28:53.255863] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.466 [2024-12-06 17:28:53.255889] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.466 [2024-12-06 17:28:53.266510] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.466 [2024-12-06 17:28:53.266536] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.466 [2024-12-06 17:28:53.278594] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.466 [2024-12-06 17:28:53.278620] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.466 [2024-12-06 17:28:53.288398] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.466 [2024-12-06 17:28:53.288423] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.466 [2024-12-06 17:28:53.298681] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.466 [2024-12-06 17:28:53.298707] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.724 [2024-12-06 17:28:53.309267] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.724 [2024-12-06 17:28:53.309294] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.724 [2024-12-06 17:28:53.320027] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.724 [2024-12-06 17:28:53.320054] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.724 [2024-12-06 17:28:53.330955] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.724 [2024-12-06 17:28:53.330997] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.724 [2024-12-06 17:28:53.343603] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.724 [2024-12-06 17:28:53.343629] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.724 [2024-12-06 17:28:53.353465] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.724 [2024-12-06 17:28:53.353491] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.724 [2024-12-06 17:28:53.364184] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.724 [2024-12-06 17:28:53.364210] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.724 [2024-12-06 17:28:53.375024] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.724 [2024-12-06 17:28:53.375050] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.724 [2024-12-06 17:28:53.385787] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.724 [2024-12-06 17:28:53.385814] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.724 [2024-12-06 17:28:53.398095] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.724 [2024-12-06 17:28:53.398121] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.724 [2024-12-06 17:28:53.408163] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.724 [2024-12-06 17:28:53.408190] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.725 [2024-12-06 17:28:53.418401] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.725 [2024-12-06 17:28:53.418435] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.725 [2024-12-06 17:28:53.429006] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.725 [2024-12-06 17:28:53.429032] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.725 [2024-12-06 17:28:53.439514] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.725 [2024-12-06 17:28:53.439541] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.725 [2024-12-06 17:28:53.449934] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.725 [2024-12-06 17:28:53.449961] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.725 [2024-12-06 17:28:53.460412] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.725 [2024-12-06 17:28:53.460438] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.725 [2024-12-06 17:28:53.471147] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.725 [2024-12-06 17:28:53.471173] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.725 [2024-12-06 17:28:53.481680] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.725 [2024-12-06 17:28:53.481706] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.725 [2024-12-06 17:28:53.491948] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.725 [2024-12-06 17:28:53.491990] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.725 [2024-12-06 17:28:53.502389] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.725 [2024-12-06 17:28:53.502414] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.725 [2024-12-06 17:28:53.513086] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.725 [2024-12-06 17:28:53.513112] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.725 [2024-12-06 17:28:53.525820] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.725 [2024-12-06 17:28:53.525848] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.725 [2024-12-06 17:28:53.535804] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.725 [2024-12-06 17:28:53.535832] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.725 [2024-12-06 17:28:53.546793] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.725 [2024-12-06 17:28:53.546819] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.725 [2024-12-06 17:28:53.559296] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.725 [2024-12-06 17:28:53.559323] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.983 [2024-12-06 17:28:53.571354] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.983 [2024-12-06 17:28:53.571382] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.983 [2024-12-06 17:28:53.580167] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.983 [2024-12-06 17:28:53.580193] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.983 [2024-12-06 17:28:53.591493] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.983 [2024-12-06 17:28:53.591519] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.983 [2024-12-06 17:28:53.603922] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.983 [2024-12-06 17:28:53.603949] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.983 [2024-12-06 17:28:53.614041] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.983 [2024-12-06 17:28:53.614067] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.983 [2024-12-06 17:28:53.624892] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.983 [2024-12-06 17:28:53.624926] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.983 [2024-12-06 17:28:53.637805] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.983 [2024-12-06 17:28:53.637832] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.983 [2024-12-06 17:28:53.648047] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.983 [2024-12-06 17:28:53.648073] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.983 [2024-12-06 17:28:53.658635] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.983 [2024-12-06 17:28:53.658688] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.983 [2024-12-06 17:28:53.669364] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.983 [2024-12-06 17:28:53.669390] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.983 [2024-12-06 17:28:53.679708] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.983 [2024-12-06 17:28:53.679735] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.983 [2024-12-06 17:28:53.690417] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.983 [2024-12-06 17:28:53.690443] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.983 [2024-12-06 17:28:53.700997] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.983 [2024-12-06 17:28:53.701023] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.983 [2024-12-06 17:28:53.713519] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.983 [2024-12-06 17:28:53.713545] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.983 [2024-12-06 17:28:53.723560] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.983 [2024-12-06 17:28:53.723586] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.983 [2024-12-06 17:28:53.734016] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.983 [2024-12-06 17:28:53.734042] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.983 [2024-12-06 17:28:53.745395] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.983 [2024-12-06 17:28:53.745421] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.983 [2024-12-06 17:28:53.756259] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.983 [2024-12-06 17:28:53.756285] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.983 [2024-12-06 17:28:53.767529] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.983 [2024-12-06 17:28:53.767555] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.983 [2024-12-06 17:28:53.778375] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.983 [2024-12-06 17:28:53.778401] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.983 [2024-12-06 17:28:53.789194] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.983 [2024-12-06 17:28:53.789220] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.983 [2024-12-06 17:28:53.801807] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.983 [2024-12-06 17:28:53.801835] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:11.983 [2024-12-06 17:28:53.811624] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:11.983 [2024-12-06 17:28:53.811673] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.241 [2024-12-06 17:28:53.822415] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.241 [2024-12-06 17:28:53.822442] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.241 [2024-12-06 17:28:53.835381] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.241 [2024-12-06 17:28:53.835416] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.241 [2024-12-06 17:28:53.845758] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.241 [2024-12-06 17:28:53.845785] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.241 [2024-12-06 17:28:53.856285] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.241 [2024-12-06 17:28:53.856311] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.241 [2024-12-06 17:28:53.868433] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.241 [2024-12-06 17:28:53.868458] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.241 [2024-12-06 17:28:53.878217] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.241 [2024-12-06 17:28:53.878242] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.241 [2024-12-06 17:28:53.889425] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.241 [2024-12-06 17:28:53.889451] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.241 [2024-12-06 17:28:53.900009] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.241 [2024-12-06 17:28:53.900035] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.241 [2024-12-06 17:28:53.910767] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.241 [2024-12-06 17:28:53.910794] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.241 [2024-12-06 17:28:53.921911] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.241 [2024-12-06 17:28:53.921938] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.241 [2024-12-06 17:28:53.934506] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.241 [2024-12-06 17:28:53.934532] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.241 [2024-12-06 17:28:53.944492] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.241 [2024-12-06 17:28:53.944520] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.241 [2024-12-06 17:28:53.954857] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.241 [2024-12-06 17:28:53.954884] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.241 [2024-12-06 17:28:53.965567] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.241 [2024-12-06 17:28:53.965594] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.241 [2024-12-06 17:28:53.976392] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.241 [2024-12-06 17:28:53.976418] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.241 [2024-12-06 17:28:53.987167] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.241 [2024-12-06 17:28:53.987193] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.241 [2024-12-06 17:28:53.999262] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.241 [2024-12-06 17:28:53.999289] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.241 [2024-12-06 17:28:54.008868] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.241 [2024-12-06 17:28:54.008896] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.241 11943.00 IOPS, 93.30 MiB/s [2024-12-06T16:28:54.080Z] [2024-12-06 17:28:54.020118] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.241 [2024-12-06 17:28:54.020145] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.241 [2024-12-06 17:28:54.030298] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.241 [2024-12-06 17:28:54.030325] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.241 [2024-12-06 17:28:54.040908] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.241 [2024-12-06 17:28:54.040935] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.241 [2024-12-06 17:28:54.051616] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.241 [2024-12-06 17:28:54.051642] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.241 [2024-12-06 17:28:54.062011] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.242 [2024-12-06 17:28:54.062037] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.242 [2024-12-06 17:28:54.072761] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.242 [2024-12-06 17:28:54.072789] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.499 [2024-12-06 17:28:54.083476] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.499 [2024-12-06 17:28:54.083502] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.499 [2024-12-06 17:28:54.095902] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.499 [2024-12-06 17:28:54.095929] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.499 [2024-12-06 17:28:54.105913] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.499 [2024-12-06 17:28:54.105940] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.499 [2024-12-06 17:28:54.116293] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.499 [2024-12-06 17:28:54.116319] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.499 [2024-12-06 17:28:54.126467] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.499 [2024-12-06 17:28:54.126493] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.499 [2024-12-06 17:28:54.136760] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.499 [2024-12-06 17:28:54.136787] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.499 [2024-12-06 17:28:54.146895] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.499 [2024-12-06 17:28:54.146921] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.499 [2024-12-06 17:28:54.157507] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.499 [2024-12-06 17:28:54.157532] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.499 [2024-12-06 17:28:54.168094] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.499 [2024-12-06 17:28:54.168120] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.499 [2024-12-06 17:28:54.178524] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.499 [2024-12-06 17:28:54.178550] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.499 [2024-12-06 17:28:54.189210] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.499 [2024-12-06 17:28:54.189236] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.499 [2024-12-06 17:28:54.200026] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.499 [2024-12-06 17:28:54.200052] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.499 [2024-12-06 17:28:54.210370] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.499 [2024-12-06 17:28:54.210395] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.499 [2024-12-06 17:28:54.220815] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.499 [2024-12-06 17:28:54.220843] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.499 [2024-12-06 17:28:54.231380] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.499 [2024-12-06 17:28:54.231406] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.499 [2024-12-06 17:28:54.242208] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.499 [2024-12-06 17:28:54.242234] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.499 [2024-12-06 17:28:54.254812] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.499 [2024-12-06 17:28:54.254839] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.500 [2024-12-06 17:28:54.264605] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.500 [2024-12-06 17:28:54.264631] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.500 [2024-12-06 17:28:54.275233] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.500 [2024-12-06 17:28:54.275260] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.500 [2024-12-06 17:28:54.285908] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.500 [2024-12-06 17:28:54.285935] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.500 [2024-12-06 17:28:54.298637] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.500 [2024-12-06 17:28:54.298689] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.500 [2024-12-06 17:28:54.308793] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.500 [2024-12-06 17:28:54.308820] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.500 [2024-12-06 17:28:54.319472] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.500 [2024-12-06 17:28:54.319498] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.500 [2024-12-06 17:28:54.330357] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.500 [2024-12-06 17:28:54.330384] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.757 [2024-12-06 17:28:54.340801] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.757 [2024-12-06 17:28:54.340828] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.757 [2024-12-06 17:28:54.351427] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.757 [2024-12-06 17:28:54.351454] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.757 [2024-12-06 17:28:54.361834] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.757 [2024-12-06 17:28:54.361860] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.757 [2024-12-06 17:28:54.372537] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.757 [2024-12-06 17:28:54.372563] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.757 [2024-12-06 17:28:54.383167] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.757 [2024-12-06 17:28:54.383193] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.757 [2024-12-06 17:28:54.395732] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.757 [2024-12-06 17:28:54.395759] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.757 [2024-12-06 17:28:54.405615] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.757 [2024-12-06 17:28:54.405657] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.757 [2024-12-06 17:28:54.416166] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.757 [2024-12-06 17:28:54.416192] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.757 [2024-12-06 17:28:54.429916] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.757 [2024-12-06 17:28:54.429942] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.758 [2024-12-06 17:28:54.439956] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.758 [2024-12-06 17:28:54.439997] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.758 [2024-12-06 17:28:54.450287] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.758 [2024-12-06 17:28:54.450312] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.758 [2024-12-06 17:28:54.460534] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.758 [2024-12-06 17:28:54.460559] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.758 [2024-12-06 17:28:54.470886] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.758 [2024-12-06 17:28:54.470912] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.758 [2024-12-06 17:28:54.481400] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.758 [2024-12-06 17:28:54.481425] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.758 [2024-12-06 17:28:54.491919] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.758 [2024-12-06 17:28:54.491946] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.758 [2024-12-06 17:28:54.502780] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.758 [2024-12-06 17:28:54.502807] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.758 [2024-12-06 17:28:54.513585] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.758 [2024-12-06 17:28:54.513612] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.758 [2024-12-06 17:28:54.524419] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.758 [2024-12-06 17:28:54.524445] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.758 [2024-12-06 17:28:54.535345] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.758 [2024-12-06 17:28:54.535372] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.758 [2024-12-06 17:28:54.548044] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.758 [2024-12-06 17:28:54.548071] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.758 [2024-12-06 17:28:54.558454] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.758 [2024-12-06 17:28:54.558481] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.758 [2024-12-06 17:28:54.569249] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.758 [2024-12-06 17:28:54.569276] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.758 [2024-12-06 17:28:54.580057] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.758 [2024-12-06 17:28:54.580084] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:12.758 [2024-12-06 17:28:54.590682] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:12.758 [2024-12-06 17:28:54.590709] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.016 [2024-12-06 17:28:54.603203] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.016 [2024-12-06 17:28:54.603229] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.016 [2024-12-06 17:28:54.613566] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.016 [2024-12-06 17:28:54.613592] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.016 [2024-12-06 17:28:54.624150] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.016 [2024-12-06 17:28:54.624176] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.016 [2024-12-06 17:28:54.634800] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.016 [2024-12-06 17:28:54.634826] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.016 [2024-12-06 17:28:54.645562] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.016 [2024-12-06 17:28:54.645595] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.016 [2024-12-06 17:28:54.656023] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.016 [2024-12-06 17:28:54.656049] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.016 [2024-12-06 17:28:54.666789] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.016 [2024-12-06 17:28:54.666817] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.016 [2024-12-06 17:28:54.677394] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.016 [2024-12-06 17:28:54.677420] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.016 [2024-12-06 17:28:54.688110] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.016 [2024-12-06 17:28:54.688136] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.016 [2024-12-06 17:28:54.698847] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.016 [2024-12-06 17:28:54.698874] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.016 [2024-12-06 17:28:54.709578] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.016 [2024-12-06 17:28:54.709605] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.016 [2024-12-06 17:28:54.720160] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.016 [2024-12-06 17:28:54.720186] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.016 [2024-12-06 17:28:54.730881] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.016 [2024-12-06 17:28:54.730909] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.016 [2024-12-06 17:28:54.743848] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.016 [2024-12-06 17:28:54.743875] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.016 [2024-12-06 17:28:54.754240] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.016 [2024-12-06 17:28:54.754266] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.016 [2024-12-06 17:28:54.765052] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.016 [2024-12-06 17:28:54.765079] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.016 [2024-12-06 17:28:54.777734] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.016 [2024-12-06 17:28:54.777762] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.016 [2024-12-06 17:28:54.788244] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.016 [2024-12-06 17:28:54.788270] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.016 [2024-12-06 17:28:54.799100] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.016 [2024-12-06 17:28:54.799127] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.016 [2024-12-06 17:28:54.809640] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.016 [2024-12-06 17:28:54.809691] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.016 [2024-12-06 17:28:54.820401] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.016 [2024-12-06 17:28:54.820428] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.016 [2024-12-06 17:28:54.832676] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.016 [2024-12-06 17:28:54.832703] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.016 [2024-12-06 17:28:54.842811] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.016 [2024-12-06 17:28:54.842838] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.016 [2024-12-06 17:28:54.853369] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.016 [2024-12-06 17:28:54.853402] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.275 [2024-12-06 17:28:54.864411] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.275 [2024-12-06 17:28:54.864437] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.275 [2024-12-06 17:28:54.875351] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.275 [2024-12-06 17:28:54.875377] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.275 [2024-12-06 17:28:54.888931] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.275 [2024-12-06 17:28:54.888958] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.275 [2024-12-06 17:28:54.899325] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.275 [2024-12-06 17:28:54.899352] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.275 [2024-12-06 17:28:54.909787] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.275 [2024-12-06 17:28:54.909814] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.275 [2024-12-06 17:28:54.920298] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.275 [2024-12-06 17:28:54.920324] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.275 [2024-12-06 17:28:54.930950] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.275 [2024-12-06 17:28:54.930993] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.275 [2024-12-06 17:28:54.943799] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.275 [2024-12-06 17:28:54.943827] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.275 [2024-12-06 17:28:54.954097] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.275 [2024-12-06 17:28:54.954123] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.275 [2024-12-06 17:28:54.964889] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.275 [2024-12-06 17:28:54.964916] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.275 [2024-12-06 17:28:54.977628] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.275 [2024-12-06 17:28:54.977676] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.275 [2024-12-06 17:28:54.987839] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.275 [2024-12-06 17:28:54.987866] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.275 [2024-12-06 17:28:54.998520] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.275 [2024-12-06 17:28:54.998545] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.275 [2024-12-06 17:28:55.010780] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.275 [2024-12-06 17:28:55.010807] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.275 11943.33 IOPS, 93.31 MiB/s [2024-12-06T16:28:55.114Z] [2024-12-06 17:28:55.021129] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.275 [2024-12-06 17:28:55.021155] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.275 [2024-12-06 17:28:55.031617] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.275 [2024-12-06 17:28:55.031643] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.275 [2024-12-06 17:28:55.042376] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.275 [2024-12-06 17:28:55.042401] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.275 [2024-12-06 17:28:55.053238] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.275 [2024-12-06 17:28:55.053263] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.275 [2024-12-06 17:28:55.064196] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.275 [2024-12-06 17:28:55.064233] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.275 [2024-12-06 17:28:55.076849] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.275 [2024-12-06 17:28:55.076877] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.275 [2024-12-06 17:28:55.086886] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.275 [2024-12-06 17:28:55.086913] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.275 [2024-12-06 17:28:55.097448] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.275 [2024-12-06 17:28:55.097475] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.275 [2024-12-06 17:28:55.107809] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.275 [2024-12-06 17:28:55.107836] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.533 [2024-12-06 17:28:55.118184] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.533 [2024-12-06 17:28:55.118210] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.533 [2024-12-06 17:28:55.129286] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.533 [2024-12-06 17:28:55.129313] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.533 [2024-12-06 17:28:55.141484] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.533 [2024-12-06 17:28:55.141510] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.533 [2024-12-06 17:28:55.151456] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.533 [2024-12-06 17:28:55.151484] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.533 [2024-12-06 17:28:55.162259] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.533 [2024-12-06 17:28:55.162285] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.533 [2024-12-06 17:28:55.174642] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.533 [2024-12-06 17:28:55.174696] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.533 [2024-12-06 17:28:55.184806] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.533 [2024-12-06 17:28:55.184833] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.533 [2024-12-06 17:28:55.195700] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.533 [2024-12-06 17:28:55.195727] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.533 [2024-12-06 17:28:55.208112] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.533 [2024-12-06 17:28:55.208140] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.533 [2024-12-06 17:28:55.217798] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.533 [2024-12-06 17:28:55.217825] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.533 [2024-12-06 17:28:55.228805] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.533 [2024-12-06 17:28:55.228835] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.533 [2024-12-06 17:28:55.239621] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.533 [2024-12-06 17:28:55.239663] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.533 [2024-12-06 17:28:55.251769] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.533 [2024-12-06 17:28:55.251796] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.533 [2024-12-06 17:28:55.262098] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.533 [2024-12-06 17:28:55.262125] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.533 [2024-12-06 17:28:55.272814] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.533 [2024-12-06 17:28:55.272841] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.533 [2024-12-06 17:28:55.283586] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.533 [2024-12-06 17:28:55.283612] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.533 [2024-12-06 17:28:55.294175] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.533 [2024-12-06 17:28:55.294201] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.533 [2024-12-06 17:28:55.305356] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.533 [2024-12-06 17:28:55.305382] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.533 [2024-12-06 17:28:55.318629] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.533 [2024-12-06 17:28:55.318679] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.533 [2024-12-06 17:28:55.328924] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.533 [2024-12-06 17:28:55.328950] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.533 [2024-12-06 17:28:55.339437] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.533 [2024-12-06 17:28:55.339464] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.533 [2024-12-06 17:28:55.350425] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.533 [2024-12-06 17:28:55.350453] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.533 [2024-12-06 17:28:55.360822] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.533 [2024-12-06 17:28:55.360850] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.791 [2024-12-06 17:28:55.371546] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.791 [2024-12-06 17:28:55.371573] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.791 [2024-12-06 17:28:55.382117] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.791 [2024-12-06 17:28:55.382144] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.791 [2024-12-06 17:28:55.393118] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.791 [2024-12-06 17:28:55.393145] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.791 [2024-12-06 17:28:55.403893] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.791 [2024-12-06 17:28:55.403921] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.791 [2024-12-06 17:28:55.414916] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.791 [2024-12-06 17:28:55.414942] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.791 [2024-12-06 17:28:55.425604] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.791 [2024-12-06 17:28:55.425631] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.791 [2024-12-06 17:28:55.436072] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.791 [2024-12-06 17:28:55.436098] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.791 [2024-12-06 17:28:55.446626] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.791 [2024-12-06 17:28:55.446676] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.791 [2024-12-06 17:28:55.457240] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.791 [2024-12-06 17:28:55.457282] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.791 [2024-12-06 17:28:55.467913] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.791 [2024-12-06 17:28:55.467940] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.791 [2024-12-06 17:28:55.480462] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.791 [2024-12-06 17:28:55.480489] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.791 [2024-12-06 17:28:55.490350] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.791 [2024-12-06 17:28:55.490391] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.791 [2024-12-06 17:28:55.500910] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.791 [2024-12-06 17:28:55.500949] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.791 [2024-12-06 17:28:55.511627] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.791 [2024-12-06 17:28:55.511654] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.791 [2024-12-06 17:28:55.524457] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.791 [2024-12-06 17:28:55.524483] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.791 [2024-12-06 17:28:55.535060] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.791 [2024-12-06 17:28:55.535086] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.791 [2024-12-06 17:28:55.546161] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.791 [2024-12-06 17:28:55.546187] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.791 [2024-12-06 17:28:55.556987] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.791 [2024-12-06 17:28:55.557014] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.791 [2024-12-06 17:28:55.567731] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.791 [2024-12-06 17:28:55.567759] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.791 [2024-12-06 17:28:55.578394] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.791 [2024-12-06 17:28:55.578421] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.791 [2024-12-06 17:28:55.589308] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.791 [2024-12-06 17:28:55.589334] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.791 [2024-12-06 17:28:55.602158] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.791 [2024-12-06 17:28:55.602185] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.791 [2024-12-06 17:28:55.612409] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.791 [2024-12-06 17:28:55.612435] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:13.791 [2024-12-06 17:28:55.623207] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:13.791 [2024-12-06 17:28:55.623250] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.048 [2024-12-06 17:28:55.636289] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.048 [2024-12-06 17:28:55.636315] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.048 [2024-12-06 17:28:55.654141] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.048 [2024-12-06 17:28:55.654167] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.048 [2024-12-06 17:28:55.664641] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.048 [2024-12-06 17:28:55.664677] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.048 [2024-12-06 17:28:55.675387] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.048 [2024-12-06 17:28:55.675414] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.048 [2024-12-06 17:28:55.686518] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.048 [2024-12-06 17:28:55.686545] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.048 [2024-12-06 17:28:55.697279] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.048 [2024-12-06 17:28:55.697322] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.048 [2024-12-06 17:28:55.708192] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.048 [2024-12-06 17:28:55.708219] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.048 [2024-12-06 17:28:55.719225] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.048 [2024-12-06 17:28:55.719251] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.048 [2024-12-06 17:28:55.732057] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.048 [2024-12-06 17:28:55.732099] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.048 [2024-12-06 17:28:55.742498] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.048 [2024-12-06 17:28:55.742524] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.048 [2024-12-06 17:28:55.753420] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.048 [2024-12-06 17:28:55.753447] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.048 [2024-12-06 17:28:55.766064] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.048 [2024-12-06 17:28:55.766090] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.048 [2024-12-06 17:28:55.776071] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.048 [2024-12-06 17:28:55.776097] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.048 [2024-12-06 17:28:55.786575] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.048 [2024-12-06 17:28:55.786601] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.048 [2024-12-06 17:28:55.797325] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.048 [2024-12-06 17:28:55.797352] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.049 [2024-12-06 17:28:55.810122] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.049 [2024-12-06 17:28:55.810149] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.049 [2024-12-06 17:28:55.820633] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.049 [2024-12-06 17:28:55.820683] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.049 [2024-12-06 17:28:55.831420] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.049 [2024-12-06 17:28:55.831447] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.049 [2024-12-06 17:28:55.843955] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.049 [2024-12-06 17:28:55.843996] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.049 [2024-12-06 17:28:55.854180] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.049 [2024-12-06 17:28:55.854207] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.049 [2024-12-06 17:28:55.865072] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.049 [2024-12-06 17:28:55.865099] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.049 [2024-12-06 17:28:55.875847] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.049 [2024-12-06 17:28:55.875874] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.049 [2024-12-06 17:28:55.886525] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.049 [2024-12-06 17:28:55.886551] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.306 [2024-12-06 17:28:55.899198] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.306 [2024-12-06 17:28:55.899224] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.306 [2024-12-06 17:28:55.909621] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.306 [2024-12-06 17:28:55.909662] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.306 [2024-12-06 17:28:55.920464] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.306 [2024-12-06 17:28:55.920491] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.306 [2024-12-06 17:28:55.932654] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.306 [2024-12-06 17:28:55.932690] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.306 [2024-12-06 17:28:55.942815] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.306 [2024-12-06 17:28:55.942842] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.306 [2024-12-06 17:28:55.953899] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.306 [2024-12-06 17:28:55.953925] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.306 [2024-12-06 17:28:55.966319] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.306 [2024-12-06 17:28:55.966345] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.306 [2024-12-06 17:28:55.976370] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.306 [2024-12-06 17:28:55.976396] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.306 [2024-12-06 17:28:55.986795] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.306 [2024-12-06 17:28:55.986822] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.306 [2024-12-06 17:28:55.997365] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.306 [2024-12-06 17:28:55.997391] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.306 [2024-12-06 17:28:56.007735] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.306 [2024-12-06 17:28:56.007762] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.306 11926.25 IOPS, 93.17 MiB/s [2024-12-06T16:28:56.145Z] [2024-12-06 17:28:56.018540] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.306 [2024-12-06 17:28:56.018567] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.306 [2024-12-06 17:28:56.029155] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.306 [2024-12-06 17:28:56.029181] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.306 [2024-12-06 17:28:56.041443] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.306 [2024-12-06 17:28:56.041468] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.306 [2024-12-06 17:28:56.051530] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.306 [2024-12-06 17:28:56.051557] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.306 [2024-12-06 17:28:56.061907] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.306 [2024-12-06 17:28:56.061934] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.306 [2024-12-06 17:28:56.072315] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.306 [2024-12-06 17:28:56.072341] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.306 [2024-12-06 17:28:56.082834] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.306 [2024-12-06 17:28:56.082860] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.306 [2024-12-06 17:28:56.093465] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.306 [2024-12-06 17:28:56.093491] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.307 [2024-12-06 17:28:56.104331] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.307 [2024-12-06 17:28:56.104363] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.307 [2024-12-06 17:28:56.114714] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.307 [2024-12-06 17:28:56.114741] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.307 [2024-12-06 17:28:56.127096] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.307 [2024-12-06 17:28:56.127122] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.307 [2024-12-06 17:28:56.137180] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.307 [2024-12-06 17:28:56.137205] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.565 [2024-12-06 17:28:56.147808] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.565 [2024-12-06 17:28:56.147836] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.565 [2024-12-06 17:28:56.158544] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.565 [2024-12-06 17:28:56.158570] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.565 [2024-12-06 17:28:56.169370] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.565 [2024-12-06 17:28:56.169396] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.565 [2024-12-06 17:28:56.181829] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.565 [2024-12-06 17:28:56.181855] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.565 [2024-12-06 17:28:56.191206] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.565 [2024-12-06 17:28:56.191232] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.565 [2024-12-06 17:28:56.202254] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.565 [2024-12-06 17:28:56.202281] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.565 [2024-12-06 17:28:56.213386] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.565 [2024-12-06 17:28:56.213412] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.565 [2024-12-06 17:28:56.224129] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.565 [2024-12-06 17:28:56.224156] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.565 [2024-12-06 17:28:56.234886] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.565 [2024-12-06 17:28:56.234913] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.565 [2024-12-06 17:28:56.245598] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.565 [2024-12-06 17:28:56.245624] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.565 [2024-12-06 17:28:56.256266] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.565 [2024-12-06 17:28:56.256292] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.565 [2024-12-06 17:28:56.266783] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.565 [2024-12-06 17:28:56.266810] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.565 [2024-12-06 17:28:56.277241] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.565 [2024-12-06 17:28:56.277267] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.565 [2024-12-06 17:28:56.288020] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.565 [2024-12-06 17:28:56.288047] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.565 [2024-12-06 17:28:56.298912] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.565 [2024-12-06 17:28:56.298940] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.565 [2024-12-06 17:28:56.311084] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.565 [2024-12-06 17:28:56.311126] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.565 [2024-12-06 17:28:56.321029] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.565 [2024-12-06 17:28:56.321057] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.565 [2024-12-06 17:28:56.331493] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.565 [2024-12-06 17:28:56.331520] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.565 [2024-12-06 17:28:56.342046] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.565 [2024-12-06 17:28:56.342072] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.565 [2024-12-06 17:28:56.354951] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.565 [2024-12-06 17:28:56.354992] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.565 [2024-12-06 17:28:56.365096] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.565 [2024-12-06 17:28:56.365123] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.565 [2024-12-06 17:28:56.376044] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.565 [2024-12-06 17:28:56.376070] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.565 [2024-12-06 17:28:56.388279] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.565 [2024-12-06 17:28:56.388304] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.565 [2024-12-06 17:28:56.397541] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.565 [2024-12-06 17:28:56.397566] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.824 [2024-12-06 17:28:56.409184] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.824 [2024-12-06 17:28:56.409210] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.824 [2024-12-06 17:28:56.420088] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.824 [2024-12-06 17:28:56.420114] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.824 [2024-12-06 17:28:56.430629] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.824 [2024-12-06 17:28:56.430679] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.824 [2024-12-06 17:28:56.441065] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.824 [2024-12-06 17:28:56.441090] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.824 [2024-12-06 17:28:56.451425] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.824 [2024-12-06 17:28:56.451451] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.824 [2024-12-06 17:28:56.461766] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.824 [2024-12-06 17:28:56.461793] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.824 [2024-12-06 17:28:56.472311] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.824 [2024-12-06 17:28:56.472337] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.824 [2024-12-06 17:28:56.483168] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.824 [2024-12-06 17:28:56.483194] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.824 [2024-12-06 17:28:56.493904] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.824 [2024-12-06 17:28:56.493931] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.824 [2024-12-06 17:28:56.506177] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.824 [2024-12-06 17:28:56.506205] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.824 [2024-12-06 17:28:56.516114] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.824 [2024-12-06 17:28:56.516148] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.824 [2024-12-06 17:28:56.526846] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.824 [2024-12-06 17:28:56.526874] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.824 [2024-12-06 17:28:56.540242] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.824 [2024-12-06 17:28:56.540269] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.824 [2024-12-06 17:28:56.550762] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.824 [2024-12-06 17:28:56.550790] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.824 [2024-12-06 17:28:56.561606] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.824 [2024-12-06 17:28:56.561633] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.824 [2024-12-06 17:28:56.573946] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.824 [2024-12-06 17:28:56.573974] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.824 [2024-12-06 17:28:56.583749] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.824 [2024-12-06 17:28:56.583777] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.824 [2024-12-06 17:28:56.594753] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.824 [2024-12-06 17:28:56.594781] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.824 [2024-12-06 17:28:56.607079] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.824 [2024-12-06 17:28:56.607105] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.824 [2024-12-06 17:28:56.617479] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.824 [2024-12-06 17:28:56.617505] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.824 [2024-12-06 17:28:56.627867] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.824 [2024-12-06 17:28:56.627894] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.824 [2024-12-06 17:28:56.638311] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.824 [2024-12-06 17:28:56.638347] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.824 [2024-12-06 17:28:56.649192] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.824 [2024-12-06 17:28:56.649218] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:14.824 [2024-12-06 17:28:56.661855] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:14.825 [2024-12-06 17:28:56.661883] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.083 [2024-12-06 17:28:56.671946] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.083 [2024-12-06 17:28:56.671987] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.083 [2024-12-06 17:28:56.682276] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.083 [2024-12-06 17:28:56.682302] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.083 [2024-12-06 17:28:56.692764] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.083 [2024-12-06 17:28:56.692791] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.083 [2024-12-06 17:28:56.703465] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.083 [2024-12-06 17:28:56.703491] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.083 [2024-12-06 17:28:56.714320] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.083 [2024-12-06 17:28:56.714347] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.083 [2024-12-06 17:28:56.726533] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.083 [2024-12-06 17:28:56.726559] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.083 [2024-12-06 17:28:56.736362] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.083 [2024-12-06 17:28:56.736388] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.083 [2024-12-06 17:28:56.747056] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.083 [2024-12-06 17:28:56.747083] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.083 [2024-12-06 17:28:56.757584] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.083 [2024-12-06 17:28:56.757610] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.083 [2024-12-06 17:28:56.768014] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.083 [2024-12-06 17:28:56.768040] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.083 [2024-12-06 17:28:56.779012] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.083 [2024-12-06 17:28:56.779038] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.083 [2024-12-06 17:28:56.791812] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.083 [2024-12-06 17:28:56.791840] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.083 [2024-12-06 17:28:56.802075] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.083 [2024-12-06 17:28:56.802101] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.083 [2024-12-06 17:28:56.812694] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.083 [2024-12-06 17:28:56.812720] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.083 [2024-12-06 17:28:56.823121] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.083 [2024-12-06 17:28:56.823147] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.083 [2024-12-06 17:28:56.843119] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.083 [2024-12-06 17:28:56.843151] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.083 [2024-12-06 17:28:56.853234] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.083 [2024-12-06 17:28:56.853260] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.083 [2024-12-06 17:28:56.863837] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.083 [2024-12-06 17:28:56.863864] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.083 [2024-12-06 17:28:56.874423] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.083 [2024-12-06 17:28:56.874448] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.083 [2024-12-06 17:28:56.885343] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.083 [2024-12-06 17:28:56.885370] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.083 [2024-12-06 17:28:56.898378] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.083 [2024-12-06 17:28:56.898405] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.083 [2024-12-06 17:28:56.911273] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.083 [2024-12-06 17:28:56.911299] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.342 [2024-12-06 17:28:56.921471] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.342 [2024-12-06 17:28:56.921497] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.342 [2024-12-06 17:28:56.932143] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.342 [2024-12-06 17:28:56.932170] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.342 [2024-12-06 17:28:56.945036] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.342 [2024-12-06 17:28:56.945063] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.342 [2024-12-06 17:28:56.955067] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.342 [2024-12-06 17:28:56.955093] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.342 [2024-12-06 17:28:56.966005] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.342 [2024-12-06 17:28:56.966031] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.342 [2024-12-06 17:28:56.976370] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.342 [2024-12-06 17:28:56.976396] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.342 [2024-12-06 17:28:56.986854] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.342 [2024-12-06 17:28:56.986881] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.342 [2024-12-06 17:28:56.996889] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.342 [2024-12-06 17:28:56.996917] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.342 [2024-12-06 17:28:57.007407] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.342 [2024-12-06 17:28:57.007434] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.342 11940.40 IOPS, 93.28 MiB/s [2024-12-06T16:28:57.181Z] [2024-12-06 17:28:57.018180] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.342 [2024-12-06 17:28:57.018206] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.342 [2024-12-06 17:28:57.025824] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.342 [2024-12-06 17:28:57.025851] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.342 00:13:15.342 Latency(us) 00:13:15.342 [2024-12-06T16:28:57.181Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:15.342 Job: Nvme1n1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 128, IO size: 8192) 00:13:15.342 Nvme1n1 : 5.01 11942.31 93.30 0.00 0.00 10704.52 4441.88 21651.15 00:13:15.342 [2024-12-06T16:28:57.181Z] =================================================================================================================== 00:13:15.342 [2024-12-06T16:28:57.181Z] Total : 11942.31 93.30 0.00 0.00 10704.52 4441.88 21651.15 00:13:15.342 [2024-12-06 17:28:57.032750] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.342 [2024-12-06 17:28:57.032776] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.342 [2024-12-06 17:28:57.040767] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.342 [2024-12-06 17:28:57.040790] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.342 [2024-12-06 17:28:57.048827] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.342 [2024-12-06 17:28:57.048874] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.342 [2024-12-06 17:28:57.056856] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.342 [2024-12-06 17:28:57.056917] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.342 [2024-12-06 17:28:57.064870] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.342 [2024-12-06 17:28:57.064921] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.342 [2024-12-06 17:28:57.072900] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.342 [2024-12-06 17:28:57.072950] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.342 [2024-12-06 17:28:57.080910] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.342 [2024-12-06 17:28:57.080974] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.342 [2024-12-06 17:28:57.088945] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.342 [2024-12-06 17:28:57.088998] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.342 [2024-12-06 17:28:57.096962] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.342 [2024-12-06 17:28:57.097010] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.342 [2024-12-06 17:28:57.104981] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.342 [2024-12-06 17:28:57.105028] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.342 [2024-12-06 17:28:57.113007] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.342 [2024-12-06 17:28:57.113057] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.342 [2024-12-06 17:28:57.121038] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.342 [2024-12-06 17:28:57.121093] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.342 [2024-12-06 17:28:57.129045] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.342 [2024-12-06 17:28:57.129095] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.342 [2024-12-06 17:28:57.137065] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.342 [2024-12-06 17:28:57.137115] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.342 [2024-12-06 17:28:57.145089] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.342 [2024-12-06 17:28:57.145138] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.343 [2024-12-06 17:28:57.153105] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.343 [2024-12-06 17:28:57.153153] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.343 [2024-12-06 17:28:57.161090] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.343 [2024-12-06 17:28:57.161115] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.343 [2024-12-06 17:28:57.169106] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.343 [2024-12-06 17:28:57.169133] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.343 [2024-12-06 17:28:57.177174] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.343 [2024-12-06 17:28:57.177223] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.601 [2024-12-06 17:28:57.185197] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.601 [2024-12-06 17:28:57.185245] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.601 [2024-12-06 17:28:57.193155] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.601 [2024-12-06 17:28:57.193176] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.601 [2024-12-06 17:28:57.201173] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.601 [2024-12-06 17:28:57.201192] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.601 [2024-12-06 17:28:57.209209] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:13:15.601 [2024-12-06 17:28:57.209228] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:15.601 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh: line 42: kill: (159024) - No such process 00:13:15.601 17:28:57 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@49 -- # wait 159024 00:13:15.601 17:28:57 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@52 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:13:15.601 17:28:57 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:15.601 17:28:57 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:13:15.601 17:28:57 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:15.601 17:28:57 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@53 -- # rpc_cmd bdev_delay_create -b malloc0 -d delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:13:15.601 17:28:57 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:15.601 17:28:57 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:13:15.601 delay0 00:13:15.601 17:28:57 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:15.601 17:28:57 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@54 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 delay0 -n 1 00:13:15.601 17:28:57 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:15.601 17:28:57 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:13:15.601 17:28:57 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:15.601 17:28:57 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -c 0x1 -t 5 -q 64 -w randrw -M 50 -l warning -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 ns:1' 00:13:15.601 [2024-12-06 17:28:57.376845] nvme_fabric.c: 295:nvme_fabric_discover_probe: *WARNING*: Skipping unsupported current discovery service or discovery service referral 00:13:22.163 [2024-12-06 17:29:03.436183] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2109bd0 is same with the state(6) to be set 00:13:22.163 [2024-12-06 17:29:03.436237] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2109bd0 is same with the state(6) to be set 00:13:22.163 Initializing NVMe Controllers 00:13:22.163 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:13:22.163 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:13:22.163 Initialization complete. Launching workers. 00:13:22.163 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 I/O completed: 320, failed: 94 00:13:22.163 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) abort submitted 381, failed to submit 33 00:13:22.163 success 223, unsuccessful 158, failed 0 00:13:22.163 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@59 -- # trap - SIGINT SIGTERM EXIT 00:13:22.163 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@60 -- # nvmftestfini 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@516 -- # nvmfcleanup 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@121 -- # sync 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@124 -- # set +e 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@125 -- # for i in {1..20} 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:13:22.164 rmmod nvme_tcp 00:13:22.164 rmmod nvme_fabrics 00:13:22.164 rmmod nvme_keyring 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@128 -- # set -e 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@129 -- # return 0 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@517 -- # '[' -n 157798 ']' 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@518 -- # killprocess 157798 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@954 -- # '[' -z 157798 ']' 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@958 -- # kill -0 157798 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@959 -- # uname 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 157798 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@972 -- # echo 'killing process with pid 157798' 00:13:22.164 killing process with pid 157798 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@973 -- # kill 157798 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@978 -- # wait 157798 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@297 -- # iptr 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@791 -- # iptables-save 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@791 -- # iptables-restore 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@302 -- # remove_spdk_ns 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:13:22.164 17:29:03 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:24.071 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:13:24.071 00:13:24.071 real 0m27.710s 00:13:24.071 user 0m41.479s 00:13:24.071 sys 0m7.607s 00:13:24.071 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:24.071 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:13:24.071 ************************************ 00:13:24.071 END TEST nvmf_zcopy 00:13:24.071 ************************************ 00:13:24.071 17:29:05 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@33 -- # run_test nvmf_nmic /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nmic.sh --transport=tcp 00:13:24.071 17:29:05 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:13:24.071 17:29:05 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:24.071 17:29:05 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:13:24.071 ************************************ 00:13:24.071 START TEST nvmf_nmic 00:13:24.071 ************************************ 00:13:24.071 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nmic.sh --transport=tcp 00:13:24.071 * Looking for test storage... 00:13:24.071 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:13:24.071 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:13:24.072 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1711 -- # lcov --version 00:13:24.072 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@333 -- # local ver1 ver1_l 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@334 -- # local ver2 ver2_l 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@336 -- # IFS=.-: 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@336 -- # read -ra ver1 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@337 -- # IFS=.-: 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@337 -- # read -ra ver2 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@338 -- # local 'op=<' 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@340 -- # ver1_l=2 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@341 -- # ver2_l=1 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@344 -- # case "$op" in 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@345 -- # : 1 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@364 -- # (( v = 0 )) 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@365 -- # decimal 1 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@353 -- # local d=1 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@355 -- # echo 1 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@365 -- # ver1[v]=1 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@366 -- # decimal 2 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@353 -- # local d=2 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@355 -- # echo 2 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@366 -- # ver2[v]=2 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@368 -- # return 0 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:13:24.330 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:24.330 --rc genhtml_branch_coverage=1 00:13:24.330 --rc genhtml_function_coverage=1 00:13:24.330 --rc genhtml_legend=1 00:13:24.330 --rc geninfo_all_blocks=1 00:13:24.330 --rc geninfo_unexecuted_blocks=1 00:13:24.330 00:13:24.330 ' 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:13:24.330 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:24.330 --rc genhtml_branch_coverage=1 00:13:24.330 --rc genhtml_function_coverage=1 00:13:24.330 --rc genhtml_legend=1 00:13:24.330 --rc geninfo_all_blocks=1 00:13:24.330 --rc geninfo_unexecuted_blocks=1 00:13:24.330 00:13:24.330 ' 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:13:24.330 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:24.330 --rc genhtml_branch_coverage=1 00:13:24.330 --rc genhtml_function_coverage=1 00:13:24.330 --rc genhtml_legend=1 00:13:24.330 --rc geninfo_all_blocks=1 00:13:24.330 --rc geninfo_unexecuted_blocks=1 00:13:24.330 00:13:24.330 ' 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:13:24.330 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:24.330 --rc genhtml_branch_coverage=1 00:13:24.330 --rc genhtml_function_coverage=1 00:13:24.330 --rc genhtml_legend=1 00:13:24.330 --rc geninfo_all_blocks=1 00:13:24.330 --rc geninfo_unexecuted_blocks=1 00:13:24.330 00:13:24.330 ' 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@7 -- # uname -s 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@15 -- # shopt -s extglob 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- paths/export.sh@5 -- # export PATH 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@51 -- # : 0 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:13:24.330 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:13:24.330 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@55 -- # have_pci_nics=0 00:13:24.331 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@11 -- # MALLOC_BDEV_SIZE=64 00:13:24.331 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:13:24.331 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@14 -- # nvmftestinit 00:13:24.331 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:13:24.331 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:13:24.331 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@476 -- # prepare_net_devs 00:13:24.331 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@438 -- # local -g is_hw=no 00:13:24.331 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@440 -- # remove_spdk_ns 00:13:24.331 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:24.331 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:13:24.331 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:24.331 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:13:24.331 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:13:24.331 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@309 -- # xtrace_disable 00:13:24.331 17:29:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@315 -- # pci_devs=() 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@315 -- # local -a pci_devs 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@316 -- # pci_net_devs=() 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@317 -- # pci_drivers=() 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@317 -- # local -A pci_drivers 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@319 -- # net_devs=() 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@319 -- # local -ga net_devs 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@320 -- # e810=() 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@320 -- # local -ga e810 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@321 -- # x722=() 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@321 -- # local -ga x722 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@322 -- # mlx=() 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@322 -- # local -ga mlx 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:13:26.871 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:13:26.871 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@418 -- # [[ up == up ]] 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:13:26.871 Found net devices under 0000:0a:00.0: cvl_0_0 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@418 -- # [[ up == up ]] 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:13:26.871 Found net devices under 0000:0a:00.1: cvl_0_1 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@442 -- # is_hw=yes 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:13:26.871 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:13:26.872 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:13:26.872 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.207 ms 00:13:26.872 00:13:26.872 --- 10.0.0.2 ping statistics --- 00:13:26.872 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:13:26.872 rtt min/avg/max/mdev = 0.207/0.207/0.207/0.000 ms 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:13:26.872 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:13:26.872 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.066 ms 00:13:26.872 00:13:26.872 --- 10.0.0.1 ping statistics --- 00:13:26.872 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:13:26.872 rtt min/avg/max/mdev = 0.066/0.066/0.066/0.000 ms 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@450 -- # return 0 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@15 -- # nvmfappstart -m 0xF 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@726 -- # xtrace_disable 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@509 -- # nvmfpid=162419 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@510 -- # waitforlisten 162419 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@835 -- # '[' -z 162419 ']' 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:26.872 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:13:26.872 [2024-12-06 17:29:08.444985] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:13:26.872 [2024-12-06 17:29:08.445092] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:26.872 [2024-12-06 17:29:08.516811] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:13:26.872 [2024-12-06 17:29:08.562836] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:13:26.872 [2024-12-06 17:29:08.562898] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:13:26.872 [2024-12-06 17:29:08.562913] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:13:26.872 [2024-12-06 17:29:08.562924] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:13:26.872 [2024-12-06 17:29:08.562933] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:13:26.872 [2024-12-06 17:29:08.564543] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:13:26.872 [2024-12-06 17:29:08.564609] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:13:26.872 [2024-12-06 17:29:08.564740] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:13:26.872 [2024-12-06 17:29:08.564744] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@868 -- # return 0 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@732 -- # xtrace_disable 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:26.872 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:13:27.131 [2024-12-06 17:29:08.709369] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@20 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:13:27.131 Malloc0 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@21 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@22 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@23 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:13:27.131 [2024-12-06 17:29:08.780583] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@25 -- # echo 'test case1: single bdev can'\''t be used in multiple subsystems' 00:13:27.131 test case1: single bdev can't be used in multiple subsystems 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@26 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 -a -s SPDK2 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@28 -- # nmic_status=0 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 Malloc0 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:13:27.131 [2024-12-06 17:29:08.804457] bdev.c:8515:bdev_open: *ERROR*: bdev Malloc0 already claimed: type exclusive_write by module NVMe-oF Target 00:13:27.131 [2024-12-06 17:29:08.804487] subsystem.c:2160:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode2: bdev Malloc0 cannot be opened, error=-1 00:13:27.131 [2024-12-06 17:29:08.804518] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:27.131 request: 00:13:27.131 { 00:13:27.131 "nqn": "nqn.2016-06.io.spdk:cnode2", 00:13:27.131 "namespace": { 00:13:27.131 "bdev_name": "Malloc0", 00:13:27.131 "no_auto_visible": false, 00:13:27.131 "hide_metadata": false 00:13:27.131 }, 00:13:27.131 "method": "nvmf_subsystem_add_ns", 00:13:27.131 "req_id": 1 00:13:27.131 } 00:13:27.131 Got JSON-RPC error response 00:13:27.131 response: 00:13:27.131 { 00:13:27.131 "code": -32602, 00:13:27.131 "message": "Invalid parameters" 00:13:27.131 } 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@29 -- # nmic_status=1 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@31 -- # '[' 1 -eq 0 ']' 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@36 -- # echo ' Adding namespace failed - expected result.' 00:13:27.131 Adding namespace failed - expected result. 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@39 -- # echo 'test case2: host connect to nvmf target in multiple paths' 00:13:27.131 test case2: host connect to nvmf target in multiple paths 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@40 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:13:27.131 [2024-12-06 17:29:08.812561] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:13:27.131 17:29:08 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@41 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:13:27.698 17:29:09 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@42 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4421 00:13:28.642 17:29:10 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@44 -- # waitforserial SPDKISFASTANDAWESOME 00:13:28.642 17:29:10 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1202 -- # local i=0 00:13:28.642 17:29:10 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:13:28.642 17:29:10 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:13:28.642 17:29:10 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1209 -- # sleep 2 00:13:30.537 17:29:12 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:13:30.537 17:29:12 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:13:30.537 17:29:12 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:13:30.537 17:29:12 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:13:30.537 17:29:12 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:13:30.537 17:29:12 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1212 -- # return 0 00:13:30.537 17:29:12 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t write -r 1 -v 00:13:30.537 [global] 00:13:30.537 thread=1 00:13:30.537 invalidate=1 00:13:30.537 rw=write 00:13:30.537 time_based=1 00:13:30.537 runtime=1 00:13:30.537 ioengine=libaio 00:13:30.537 direct=1 00:13:30.537 bs=4096 00:13:30.537 iodepth=1 00:13:30.537 norandommap=0 00:13:30.537 numjobs=1 00:13:30.537 00:13:30.537 verify_dump=1 00:13:30.537 verify_backlog=512 00:13:30.537 verify_state_save=0 00:13:30.537 do_verify=1 00:13:30.537 verify=crc32c-intel 00:13:30.537 [job0] 00:13:30.537 filename=/dev/nvme0n1 00:13:30.537 Could not set queue depth (nvme0n1) 00:13:30.794 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:13:30.794 fio-3.35 00:13:30.794 Starting 1 thread 00:13:32.164 00:13:32.164 job0: (groupid=0, jobs=1): err= 0: pid=163057: Fri Dec 6 17:29:13 2024 00:13:32.164 read: IOPS=22, BW=91.4KiB/s (93.6kB/s)(92.0KiB/1007msec) 00:13:32.164 slat (nsec): min=14415, max=33393, avg=25563.22, stdev=8364.06 00:13:32.164 clat (usec): min=324, max=42054, avg=39788.54, stdev=8617.39 00:13:32.164 lat (usec): min=357, max=42071, avg=39814.10, stdev=8615.77 00:13:32.164 clat percentiles (usec): 00:13:32.164 | 1.00th=[ 326], 5.00th=[40633], 10.00th=[40633], 20.00th=[41157], 00:13:32.164 | 30.00th=[41157], 40.00th=[41681], 50.00th=[41681], 60.00th=[41681], 00:13:32.164 | 70.00th=[42206], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:13:32.164 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:13:32.164 | 99.99th=[42206] 00:13:32.164 write: IOPS=508, BW=2034KiB/s (2083kB/s)(2048KiB/1007msec); 0 zone resets 00:13:32.164 slat (nsec): min=6258, max=49379, avg=14425.14, stdev=6630.17 00:13:32.164 clat (usec): min=136, max=281, avg=159.75, stdev=11.94 00:13:32.164 lat (usec): min=144, max=297, avg=174.18, stdev=14.79 00:13:32.164 clat percentiles (usec): 00:13:32.164 | 1.00th=[ 141], 5.00th=[ 145], 10.00th=[ 147], 20.00th=[ 151], 00:13:32.164 | 30.00th=[ 155], 40.00th=[ 157], 50.00th=[ 159], 60.00th=[ 161], 00:13:32.164 | 70.00th=[ 163], 80.00th=[ 167], 90.00th=[ 172], 95.00th=[ 178], 00:13:32.164 | 99.00th=[ 194], 99.50th=[ 229], 99.90th=[ 281], 99.95th=[ 281], 00:13:32.164 | 99.99th=[ 281] 00:13:32.164 bw ( KiB/s): min= 4096, max= 4096, per=100.00%, avg=4096.00, stdev= 0.00, samples=1 00:13:32.164 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:13:32.164 lat (usec) : 250=95.51%, 500=0.37% 00:13:32.164 lat (msec) : 50=4.11% 00:13:32.164 cpu : usr=0.40%, sys=0.70%, ctx=535, majf=0, minf=1 00:13:32.164 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:13:32.164 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:32.164 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:32.164 issued rwts: total=23,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:32.164 latency : target=0, window=0, percentile=100.00%, depth=1 00:13:32.164 00:13:32.164 Run status group 0 (all jobs): 00:13:32.164 READ: bw=91.4KiB/s (93.6kB/s), 91.4KiB/s-91.4KiB/s (93.6kB/s-93.6kB/s), io=92.0KiB (94.2kB), run=1007-1007msec 00:13:32.164 WRITE: bw=2034KiB/s (2083kB/s), 2034KiB/s-2034KiB/s (2083kB/s-2083kB/s), io=2048KiB (2097kB), run=1007-1007msec 00:13:32.164 00:13:32.164 Disk stats (read/write): 00:13:32.164 nvme0n1: ios=70/512, merge=0/0, ticks=807/82, in_queue=889, util=91.58% 00:13:32.164 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@48 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:13:32.164 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 2 controller(s) 00:13:32.164 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@49 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:13:32.164 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1223 -- # local i=0 00:13:32.164 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:13:32.164 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:13:32.164 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:13:32.164 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:13:32.165 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1235 -- # return 0 00:13:32.165 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@51 -- # trap - SIGINT SIGTERM EXIT 00:13:32.165 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@53 -- # nvmftestfini 00:13:32.165 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@516 -- # nvmfcleanup 00:13:32.165 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@121 -- # sync 00:13:32.165 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:13:32.165 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@124 -- # set +e 00:13:32.165 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@125 -- # for i in {1..20} 00:13:32.165 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:13:32.165 rmmod nvme_tcp 00:13:32.165 rmmod nvme_fabrics 00:13:32.165 rmmod nvme_keyring 00:13:32.165 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:13:32.165 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@128 -- # set -e 00:13:32.165 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@129 -- # return 0 00:13:32.165 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@517 -- # '[' -n 162419 ']' 00:13:32.165 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@518 -- # killprocess 162419 00:13:32.165 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@954 -- # '[' -z 162419 ']' 00:13:32.165 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@958 -- # kill -0 162419 00:13:32.165 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@959 -- # uname 00:13:32.165 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:32.165 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 162419 00:13:32.165 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:32.165 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:32.165 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@972 -- # echo 'killing process with pid 162419' 00:13:32.165 killing process with pid 162419 00:13:32.165 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@973 -- # kill 162419 00:13:32.165 17:29:13 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@978 -- # wait 162419 00:13:32.425 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:13:32.425 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:13:32.425 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:13:32.425 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@297 -- # iptr 00:13:32.425 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@791 -- # iptables-save 00:13:32.425 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:13:32.425 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@791 -- # iptables-restore 00:13:32.425 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:13:32.425 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@302 -- # remove_spdk_ns 00:13:32.425 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:32.425 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:13:32.425 17:29:14 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:13:34.967 00:13:34.967 real 0m10.395s 00:13:34.967 user 0m23.429s 00:13:34.967 sys 0m2.706s 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:13:34.967 ************************************ 00:13:34.967 END TEST nvmf_nmic 00:13:34.967 ************************************ 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@34 -- # run_test nvmf_fio_target /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fio.sh --transport=tcp 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:13:34.967 ************************************ 00:13:34.967 START TEST nvmf_fio_target 00:13:34.967 ************************************ 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fio.sh --transport=tcp 00:13:34.967 * Looking for test storage... 00:13:34.967 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1711 -- # lcov --version 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@333 -- # local ver1 ver1_l 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@334 -- # local ver2 ver2_l 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@336 -- # IFS=.-: 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@336 -- # read -ra ver1 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@337 -- # IFS=.-: 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@337 -- # read -ra ver2 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@338 -- # local 'op=<' 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@340 -- # ver1_l=2 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@341 -- # ver2_l=1 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@344 -- # case "$op" in 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@345 -- # : 1 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@364 -- # (( v = 0 )) 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@365 -- # decimal 1 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@353 -- # local d=1 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@355 -- # echo 1 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@365 -- # ver1[v]=1 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@366 -- # decimal 2 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@353 -- # local d=2 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@355 -- # echo 2 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@366 -- # ver2[v]=2 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@368 -- # return 0 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:13:34.967 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:34.967 --rc genhtml_branch_coverage=1 00:13:34.967 --rc genhtml_function_coverage=1 00:13:34.967 --rc genhtml_legend=1 00:13:34.967 --rc geninfo_all_blocks=1 00:13:34.967 --rc geninfo_unexecuted_blocks=1 00:13:34.967 00:13:34.967 ' 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:13:34.967 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:34.967 --rc genhtml_branch_coverage=1 00:13:34.967 --rc genhtml_function_coverage=1 00:13:34.967 --rc genhtml_legend=1 00:13:34.967 --rc geninfo_all_blocks=1 00:13:34.967 --rc geninfo_unexecuted_blocks=1 00:13:34.967 00:13:34.967 ' 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:13:34.967 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:34.967 --rc genhtml_branch_coverage=1 00:13:34.967 --rc genhtml_function_coverage=1 00:13:34.967 --rc genhtml_legend=1 00:13:34.967 --rc geninfo_all_blocks=1 00:13:34.967 --rc geninfo_unexecuted_blocks=1 00:13:34.967 00:13:34.967 ' 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:13:34.967 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:34.967 --rc genhtml_branch_coverage=1 00:13:34.967 --rc genhtml_function_coverage=1 00:13:34.967 --rc genhtml_legend=1 00:13:34.967 --rc geninfo_all_blocks=1 00:13:34.967 --rc geninfo_unexecuted_blocks=1 00:13:34.967 00:13:34.967 ' 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@7 -- # uname -s 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:13:34.967 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@15 -- # shopt -s extglob 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- paths/export.sh@5 -- # export PATH 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@51 -- # : 0 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:13:34.968 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@55 -- # have_pci_nics=0 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@11 -- # MALLOC_BDEV_SIZE=64 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@14 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@16 -- # nvmftestinit 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@476 -- # prepare_net_devs 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@438 -- # local -g is_hw=no 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@440 -- # remove_spdk_ns 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@309 -- # xtrace_disable 00:13:34.968 17:29:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:13:36.878 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:13:36.878 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@315 -- # pci_devs=() 00:13:36.878 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@315 -- # local -a pci_devs 00:13:36.878 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@316 -- # pci_net_devs=() 00:13:36.878 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:13:36.878 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@317 -- # pci_drivers=() 00:13:36.878 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@317 -- # local -A pci_drivers 00:13:36.878 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@319 -- # net_devs=() 00:13:36.878 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@319 -- # local -ga net_devs 00:13:36.878 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@320 -- # e810=() 00:13:36.878 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@320 -- # local -ga e810 00:13:36.878 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@321 -- # x722=() 00:13:36.878 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@321 -- # local -ga x722 00:13:36.878 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@322 -- # mlx=() 00:13:36.878 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@322 -- # local -ga mlx 00:13:36.878 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:13:36.878 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:13:36.878 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:13:36.878 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:13:36.878 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:13:36.878 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:13:36.878 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:13:36.878 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:13:36.878 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:13:36.878 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:13:36.878 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:13:36.878 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:13:36.879 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:13:36.879 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@418 -- # [[ up == up ]] 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:13:36.879 Found net devices under 0000:0a:00.0: cvl_0_0 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@418 -- # [[ up == up ]] 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:13:36.879 Found net devices under 0000:0a:00.1: cvl_0_1 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@442 -- # is_hw=yes 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:13:36.879 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:13:37.138 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:13:37.138 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.311 ms 00:13:37.138 00:13:37.138 --- 10.0.0.2 ping statistics --- 00:13:37.138 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:13:37.138 rtt min/avg/max/mdev = 0.311/0.311/0.311/0.000 ms 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:13:37.138 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:13:37.138 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.108 ms 00:13:37.138 00:13:37.138 --- 10.0.0.1 ping statistics --- 00:13:37.138 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:13:37.138 rtt min/avg/max/mdev = 0.108/0.108/0.108/0.000 ms 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@450 -- # return 0 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@17 -- # nvmfappstart -m 0xF 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@726 -- # xtrace_disable 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@509 -- # nvmfpid=165147 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@510 -- # waitforlisten 165147 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@835 -- # '[' -z 165147 ']' 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@840 -- # local max_retries=100 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:37.138 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@844 -- # xtrace_disable 00:13:37.138 17:29:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:13:37.138 [2024-12-06 17:29:18.894559] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:13:37.138 [2024-12-06 17:29:18.894644] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:37.138 [2024-12-06 17:29:18.967914] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:13:37.397 [2024-12-06 17:29:19.012563] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:13:37.397 [2024-12-06 17:29:19.012620] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:13:37.397 [2024-12-06 17:29:19.012648] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:13:37.397 [2024-12-06 17:29:19.012659] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:13:37.397 [2024-12-06 17:29:19.012682] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:13:37.397 [2024-12-06 17:29:19.014270] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:13:37.397 [2024-12-06 17:29:19.014379] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:13:37.397 [2024-12-06 17:29:19.014472] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:13:37.397 [2024-12-06 17:29:19.014479] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:37.397 17:29:19 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:13:37.397 17:29:19 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@868 -- # return 0 00:13:37.397 17:29:19 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:13:37.397 17:29:19 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@732 -- # xtrace_disable 00:13:37.397 17:29:19 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:13:37.397 17:29:19 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:13:37.397 17:29:19 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:13:37.655 [2024-12-06 17:29:19.395597] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:13:37.655 17:29:19 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:13:37.914 17:29:19 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@21 -- # malloc_bdevs='Malloc0 ' 00:13:37.914 17:29:19 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@22 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:13:38.173 17:29:20 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@22 -- # malloc_bdevs+=Malloc1 00:13:38.173 17:29:20 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:13:38.740 17:29:20 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@24 -- # raid_malloc_bdevs='Malloc2 ' 00:13:38.740 17:29:20 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:13:38.998 17:29:20 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@25 -- # raid_malloc_bdevs+=Malloc3 00:13:38.998 17:29:20 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n raid0 -z 64 -r 0 -b 'Malloc2 Malloc3' 00:13:39.257 17:29:20 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:13:39.516 17:29:21 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@29 -- # concat_malloc_bdevs='Malloc4 ' 00:13:39.516 17:29:21 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:13:39.775 17:29:21 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@30 -- # concat_malloc_bdevs+='Malloc5 ' 00:13:39.775 17:29:21 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:13:40.033 17:29:21 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@31 -- # concat_malloc_bdevs+=Malloc6 00:13:40.033 17:29:21 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n concat0 -r concat -z 64 -b 'Malloc4 Malloc5 Malloc6' 00:13:40.290 17:29:21 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:13:40.548 17:29:22 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@35 -- # for malloc_bdev in $malloc_bdevs 00:13:40.548 17:29:22 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:13:40.807 17:29:22 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@35 -- # for malloc_bdev in $malloc_bdevs 00:13:40.807 17:29:22 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:13:41.065 17:29:22 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:13:41.323 [2024-12-06 17:29:23.069438] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:13:41.323 17:29:23 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 raid0 00:13:41.582 17:29:23 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 concat0 00:13:41.840 17:29:23 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@46 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:13:42.774 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@48 -- # waitforserial SPDKISFASTANDAWESOME 4 00:13:42.774 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1202 -- # local i=0 00:13:42.774 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:13:42.774 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1204 -- # [[ -n 4 ]] 00:13:42.774 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1205 -- # nvme_device_counter=4 00:13:42.774 17:29:24 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1209 -- # sleep 2 00:13:44.683 17:29:26 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:13:44.683 17:29:26 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:13:44.683 17:29:26 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:13:44.683 17:29:26 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1211 -- # nvme_devices=4 00:13:44.683 17:29:26 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:13:44.683 17:29:26 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1212 -- # return 0 00:13:44.683 17:29:26 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t write -r 1 -v 00:13:44.683 [global] 00:13:44.683 thread=1 00:13:44.683 invalidate=1 00:13:44.683 rw=write 00:13:44.683 time_based=1 00:13:44.683 runtime=1 00:13:44.683 ioengine=libaio 00:13:44.683 direct=1 00:13:44.683 bs=4096 00:13:44.683 iodepth=1 00:13:44.683 norandommap=0 00:13:44.683 numjobs=1 00:13:44.683 00:13:44.683 verify_dump=1 00:13:44.683 verify_backlog=512 00:13:44.683 verify_state_save=0 00:13:44.683 do_verify=1 00:13:44.683 verify=crc32c-intel 00:13:44.683 [job0] 00:13:44.683 filename=/dev/nvme0n1 00:13:44.683 [job1] 00:13:44.683 filename=/dev/nvme0n2 00:13:44.683 [job2] 00:13:44.683 filename=/dev/nvme0n3 00:13:44.683 [job3] 00:13:44.683 filename=/dev/nvme0n4 00:13:44.683 Could not set queue depth (nvme0n1) 00:13:44.683 Could not set queue depth (nvme0n2) 00:13:44.683 Could not set queue depth (nvme0n3) 00:13:44.683 Could not set queue depth (nvme0n4) 00:13:44.940 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:13:44.940 job1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:13:44.940 job2: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:13:44.940 job3: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:13:44.940 fio-3.35 00:13:44.940 Starting 4 threads 00:13:46.315 00:13:46.315 job0: (groupid=0, jobs=1): err= 0: pid=166230: Fri Dec 6 17:29:27 2024 00:13:46.315 read: IOPS=1535, BW=6142KiB/s (6289kB/s)(6148KiB/1001msec) 00:13:46.315 slat (nsec): min=5734, max=57359, avg=11460.63, stdev=5935.26 00:13:46.315 clat (usec): min=168, max=42054, avg=373.32, stdev=2111.29 00:13:46.315 lat (usec): min=186, max=42063, avg=384.78, stdev=2111.48 00:13:46.315 clat percentiles (usec): 00:13:46.315 | 1.00th=[ 184], 5.00th=[ 194], 10.00th=[ 200], 20.00th=[ 210], 00:13:46.315 | 30.00th=[ 219], 40.00th=[ 229], 50.00th=[ 243], 60.00th=[ 255], 00:13:46.315 | 70.00th=[ 281], 80.00th=[ 322], 90.00th=[ 379], 95.00th=[ 400], 00:13:46.315 | 99.00th=[ 486], 99.50th=[ 578], 99.90th=[41681], 99.95th=[42206], 00:13:46.315 | 99.99th=[42206] 00:13:46.315 write: IOPS=2045, BW=8184KiB/s (8380kB/s)(8192KiB/1001msec); 0 zone resets 00:13:46.315 slat (nsec): min=6597, max=67120, avg=17032.58, stdev=7551.03 00:13:46.315 clat (usec): min=133, max=420, avg=175.82, stdev=31.26 00:13:46.315 lat (usec): min=144, max=439, avg=192.86, stdev=35.55 00:13:46.315 clat percentiles (usec): 00:13:46.315 | 1.00th=[ 141], 5.00th=[ 147], 10.00th=[ 149], 20.00th=[ 153], 00:13:46.315 | 30.00th=[ 159], 40.00th=[ 165], 50.00th=[ 169], 60.00th=[ 174], 00:13:46.315 | 70.00th=[ 182], 80.00th=[ 190], 90.00th=[ 215], 95.00th=[ 235], 00:13:46.315 | 99.00th=[ 302], 99.50th=[ 330], 99.90th=[ 408], 99.95th=[ 416], 00:13:46.315 | 99.99th=[ 420] 00:13:46.315 bw ( KiB/s): min= 8520, max= 8520, per=34.39%, avg=8520.00, stdev= 0.00, samples=1 00:13:46.315 iops : min= 2130, max= 2130, avg=2130.00, stdev= 0.00, samples=1 00:13:46.315 lat (usec) : 250=79.61%, 500=20.03%, 750=0.22% 00:13:46.315 lat (msec) : 4=0.03%, 50=0.11% 00:13:46.315 cpu : usr=3.70%, sys=6.30%, ctx=3586, majf=0, minf=1 00:13:46.315 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:13:46.315 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:46.315 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:46.315 issued rwts: total=1537,2048,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:46.315 latency : target=0, window=0, percentile=100.00%, depth=1 00:13:46.315 job1: (groupid=0, jobs=1): err= 0: pid=166231: Fri Dec 6 17:29:27 2024 00:13:46.315 read: IOPS=1017, BW=4071KiB/s (4169kB/s)(4112KiB/1010msec) 00:13:46.315 slat (nsec): min=6503, max=26109, avg=7924.29, stdev=2024.11 00:13:46.315 clat (usec): min=183, max=41607, avg=664.02, stdev=4184.80 00:13:46.315 lat (usec): min=190, max=41614, avg=671.95, stdev=4184.96 00:13:46.315 clat percentiles (usec): 00:13:46.315 | 1.00th=[ 190], 5.00th=[ 196], 10.00th=[ 200], 20.00th=[ 204], 00:13:46.315 | 30.00th=[ 208], 40.00th=[ 212], 50.00th=[ 217], 60.00th=[ 221], 00:13:46.315 | 70.00th=[ 231], 80.00th=[ 255], 90.00th=[ 281], 95.00th=[ 326], 00:13:46.315 | 99.00th=[40633], 99.50th=[41157], 99.90th=[41157], 99.95th=[41681], 00:13:46.315 | 99.99th=[41681] 00:13:46.315 write: IOPS=1520, BW=6083KiB/s (6229kB/s)(6144KiB/1010msec); 0 zone resets 00:13:46.315 slat (nsec): min=7961, max=80484, avg=10625.57, stdev=3736.43 00:13:46.315 clat (usec): min=128, max=387, avg=192.13, stdev=52.44 00:13:46.315 lat (usec): min=137, max=468, avg=202.75, stdev=53.35 00:13:46.315 clat percentiles (usec): 00:13:46.315 | 1.00th=[ 135], 5.00th=[ 141], 10.00th=[ 143], 20.00th=[ 147], 00:13:46.315 | 30.00th=[ 151], 40.00th=[ 157], 50.00th=[ 167], 60.00th=[ 180], 00:13:46.315 | 70.00th=[ 235], 80.00th=[ 249], 90.00th=[ 269], 95.00th=[ 281], 00:13:46.315 | 99.00th=[ 326], 99.50th=[ 347], 99.90th=[ 388], 99.95th=[ 388], 00:13:46.315 | 99.99th=[ 388] 00:13:46.315 bw ( KiB/s): min= 4096, max= 8192, per=24.80%, avg=6144.00, stdev=2896.31, samples=2 00:13:46.315 iops : min= 1024, max= 2048, avg=1536.00, stdev=724.08, samples=2 00:13:46.315 lat (usec) : 250=79.45%, 500=20.12% 00:13:46.315 lat (msec) : 50=0.43% 00:13:46.316 cpu : usr=1.19%, sys=3.57%, ctx=2564, majf=0, minf=1 00:13:46.316 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:13:46.316 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:46.316 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:46.316 issued rwts: total=1028,1536,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:46.316 latency : target=0, window=0, percentile=100.00%, depth=1 00:13:46.316 job2: (groupid=0, jobs=1): err= 0: pid=166232: Fri Dec 6 17:29:27 2024 00:13:46.316 read: IOPS=131, BW=525KiB/s (538kB/s)(540KiB/1028msec) 00:13:46.316 slat (nsec): min=5151, max=43866, avg=10329.81, stdev=8968.09 00:13:46.316 clat (usec): min=185, max=41259, avg=6556.18, stdev=14804.81 00:13:46.316 lat (usec): min=191, max=41292, avg=6566.51, stdev=14808.82 00:13:46.316 clat percentiles (usec): 00:13:46.316 | 1.00th=[ 188], 5.00th=[ 190], 10.00th=[ 194], 20.00th=[ 198], 00:13:46.316 | 30.00th=[ 200], 40.00th=[ 204], 50.00th=[ 210], 60.00th=[ 215], 00:13:46.316 | 70.00th=[ 225], 80.00th=[ 383], 90.00th=[41157], 95.00th=[41157], 00:13:46.316 | 99.00th=[41157], 99.50th=[41157], 99.90th=[41157], 99.95th=[41157], 00:13:46.316 | 99.99th=[41157] 00:13:46.316 write: IOPS=498, BW=1992KiB/s (2040kB/s)(2048KiB/1028msec); 0 zone resets 00:13:46.316 slat (nsec): min=7214, max=51907, avg=19020.01, stdev=8658.85 00:13:46.316 clat (usec): min=160, max=412, avg=251.24, stdev=52.41 00:13:46.316 lat (usec): min=173, max=464, avg=270.26, stdev=51.51 00:13:46.316 clat percentiles (usec): 00:13:46.316 | 1.00th=[ 174], 5.00th=[ 180], 10.00th=[ 186], 20.00th=[ 198], 00:13:46.316 | 30.00th=[ 208], 40.00th=[ 229], 50.00th=[ 247], 60.00th=[ 273], 00:13:46.316 | 70.00th=[ 285], 80.00th=[ 297], 90.00th=[ 322], 95.00th=[ 338], 00:13:46.316 | 99.00th=[ 367], 99.50th=[ 379], 99.90th=[ 412], 99.95th=[ 412], 00:13:46.316 | 99.99th=[ 412] 00:13:46.316 bw ( KiB/s): min= 4096, max= 4096, per=16.53%, avg=4096.00, stdev= 0.00, samples=1 00:13:46.316 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:13:46.316 lat (usec) : 250=55.95%, 500=40.49%, 750=0.31% 00:13:46.316 lat (msec) : 50=3.25% 00:13:46.316 cpu : usr=0.39%, sys=1.17%, ctx=647, majf=0, minf=2 00:13:46.316 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:13:46.316 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:46.316 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:46.316 issued rwts: total=135,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:46.316 latency : target=0, window=0, percentile=100.00%, depth=1 00:13:46.316 job3: (groupid=0, jobs=1): err= 0: pid=166233: Fri Dec 6 17:29:27 2024 00:13:46.316 read: IOPS=2045, BW=8184KiB/s (8380kB/s)(8192KiB/1001msec) 00:13:46.316 slat (nsec): min=5357, max=64627, avg=13509.87, stdev=7989.62 00:13:46.316 clat (usec): min=186, max=535, avg=257.63, stdev=58.76 00:13:46.316 lat (usec): min=195, max=542, avg=271.14, stdev=62.87 00:13:46.316 clat percentiles (usec): 00:13:46.316 | 1.00th=[ 192], 5.00th=[ 200], 10.00th=[ 204], 20.00th=[ 210], 00:13:46.316 | 30.00th=[ 215], 40.00th=[ 221], 50.00th=[ 233], 60.00th=[ 265], 00:13:46.316 | 70.00th=[ 285], 80.00th=[ 306], 90.00th=[ 338], 95.00th=[ 359], 00:13:46.316 | 99.00th=[ 453], 99.50th=[ 478], 99.90th=[ 502], 99.95th=[ 502], 00:13:46.316 | 99.99th=[ 537] 00:13:46.316 write: IOPS=2268, BW=9075KiB/s (9293kB/s)(9084KiB/1001msec); 0 zone resets 00:13:46.316 slat (nsec): min=6703, max=91475, avg=12628.22, stdev=6787.20 00:13:46.316 clat (usec): min=131, max=446, avg=176.18, stdev=35.08 00:13:46.316 lat (usec): min=139, max=471, avg=188.80, stdev=38.32 00:13:46.316 clat percentiles (usec): 00:13:46.316 | 1.00th=[ 139], 5.00th=[ 145], 10.00th=[ 149], 20.00th=[ 153], 00:13:46.316 | 30.00th=[ 157], 40.00th=[ 163], 50.00th=[ 167], 60.00th=[ 174], 00:13:46.316 | 70.00th=[ 182], 80.00th=[ 196], 90.00th=[ 206], 95.00th=[ 225], 00:13:46.316 | 99.00th=[ 334], 99.50th=[ 388], 99.90th=[ 441], 99.95th=[ 441], 00:13:46.316 | 99.99th=[ 449] 00:13:46.316 bw ( KiB/s): min= 8824, max= 8824, per=35.62%, avg=8824.00, stdev= 0.00, samples=1 00:13:46.316 iops : min= 2206, max= 2206, avg=2206.00, stdev= 0.00, samples=1 00:13:46.316 lat (usec) : 250=77.52%, 500=22.39%, 750=0.09% 00:13:46.316 cpu : usr=3.00%, sys=5.90%, ctx=4319, majf=0, minf=2 00:13:46.316 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:13:46.316 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:46.316 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:46.316 issued rwts: total=2048,2271,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:46.316 latency : target=0, window=0, percentile=100.00%, depth=1 00:13:46.316 00:13:46.316 Run status group 0 (all jobs): 00:13:46.316 READ: bw=18.0MiB/s (18.9MB/s), 525KiB/s-8184KiB/s (538kB/s-8380kB/s), io=18.5MiB (19.4MB), run=1001-1028msec 00:13:46.316 WRITE: bw=24.2MiB/s (25.4MB/s), 1992KiB/s-9075KiB/s (2040kB/s-9293kB/s), io=24.9MiB (26.1MB), run=1001-1028msec 00:13:46.316 00:13:46.316 Disk stats (read/write): 00:13:46.316 nvme0n1: ios=1584/1536, merge=0/0, ticks=895/237, in_queue=1132, util=97.80% 00:13:46.316 nvme0n2: ios=1024/1094, merge=0/0, ticks=589/223, in_queue=812, util=86.34% 00:13:46.316 nvme0n3: ios=125/512, merge=0/0, ticks=678/126, in_queue=804, util=88.98% 00:13:46.316 nvme0n4: ios=1720/2048, merge=0/0, ticks=414/342, in_queue=756, util=89.64% 00:13:46.316 17:29:27 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@51 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t randwrite -r 1 -v 00:13:46.316 [global] 00:13:46.316 thread=1 00:13:46.316 invalidate=1 00:13:46.316 rw=randwrite 00:13:46.316 time_based=1 00:13:46.316 runtime=1 00:13:46.316 ioengine=libaio 00:13:46.316 direct=1 00:13:46.316 bs=4096 00:13:46.316 iodepth=1 00:13:46.316 norandommap=0 00:13:46.316 numjobs=1 00:13:46.316 00:13:46.316 verify_dump=1 00:13:46.316 verify_backlog=512 00:13:46.316 verify_state_save=0 00:13:46.316 do_verify=1 00:13:46.316 verify=crc32c-intel 00:13:46.316 [job0] 00:13:46.316 filename=/dev/nvme0n1 00:13:46.316 [job1] 00:13:46.316 filename=/dev/nvme0n2 00:13:46.316 [job2] 00:13:46.316 filename=/dev/nvme0n3 00:13:46.316 [job3] 00:13:46.316 filename=/dev/nvme0n4 00:13:46.316 Could not set queue depth (nvme0n1) 00:13:46.316 Could not set queue depth (nvme0n2) 00:13:46.316 Could not set queue depth (nvme0n3) 00:13:46.316 Could not set queue depth (nvme0n4) 00:13:46.316 job0: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:13:46.316 job1: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:13:46.316 job2: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:13:46.316 job3: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:13:46.316 fio-3.35 00:13:46.316 Starting 4 threads 00:13:47.701 00:13:47.701 job0: (groupid=0, jobs=1): err= 0: pid=166572: Fri Dec 6 17:29:29 2024 00:13:47.701 read: IOPS=1555, BW=6222KiB/s (6371kB/s)(6228KiB/1001msec) 00:13:47.701 slat (nsec): min=5579, max=42153, avg=12377.84, stdev=4647.64 00:13:47.701 clat (usec): min=176, max=41995, avg=370.07, stdev=2369.65 00:13:47.701 lat (usec): min=183, max=42007, avg=382.45, stdev=2369.61 00:13:47.701 clat percentiles (usec): 00:13:47.701 | 1.00th=[ 184], 5.00th=[ 192], 10.00th=[ 198], 20.00th=[ 206], 00:13:47.701 | 30.00th=[ 217], 40.00th=[ 225], 50.00th=[ 231], 60.00th=[ 235], 00:13:47.701 | 70.00th=[ 239], 80.00th=[ 245], 90.00th=[ 251], 95.00th=[ 258], 00:13:47.701 | 99.00th=[ 273], 99.50th=[ 404], 99.90th=[41681], 99.95th=[42206], 00:13:47.701 | 99.99th=[42206] 00:13:47.701 write: IOPS=2045, BW=8184KiB/s (8380kB/s)(8192KiB/1001msec); 0 zone resets 00:13:47.701 slat (nsec): min=5327, max=53916, avg=15838.71, stdev=5767.71 00:13:47.701 clat (usec): min=130, max=294, avg=174.05, stdev=24.95 00:13:47.701 lat (usec): min=137, max=313, avg=189.89, stdev=27.00 00:13:47.701 clat percentiles (usec): 00:13:47.701 | 1.00th=[ 135], 5.00th=[ 139], 10.00th=[ 143], 20.00th=[ 151], 00:13:47.701 | 30.00th=[ 159], 40.00th=[ 167], 50.00th=[ 172], 60.00th=[ 178], 00:13:47.701 | 70.00th=[ 184], 80.00th=[ 196], 90.00th=[ 210], 95.00th=[ 219], 00:13:47.701 | 99.00th=[ 239], 99.50th=[ 247], 99.90th=[ 260], 99.95th=[ 269], 00:13:47.701 | 99.99th=[ 293] 00:13:47.701 bw ( KiB/s): min= 6216, max= 6216, per=34.60%, avg=6216.00, stdev= 0.00, samples=1 00:13:47.701 iops : min= 1554, max= 1554, avg=1554.00, stdev= 0.00, samples=1 00:13:47.701 lat (usec) : 250=94.92%, 500=4.88%, 750=0.03% 00:13:47.701 lat (msec) : 20=0.03%, 50=0.14% 00:13:47.701 cpu : usr=3.90%, sys=6.80%, ctx=3606, majf=0, minf=1 00:13:47.701 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:13:47.701 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:47.701 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:47.701 issued rwts: total=1557,2048,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:47.701 latency : target=0, window=0, percentile=100.00%, depth=1 00:13:47.701 job1: (groupid=0, jobs=1): err= 0: pid=166582: Fri Dec 6 17:29:29 2024 00:13:47.701 read: IOPS=1220, BW=4880KiB/s (4997kB/s)(4968KiB/1018msec) 00:13:47.701 slat (nsec): min=5435, max=48355, avg=12498.26, stdev=5798.21 00:13:47.701 clat (usec): min=172, max=41317, avg=541.27, stdev=3255.11 00:13:47.701 lat (usec): min=179, max=41326, avg=553.77, stdev=3255.21 00:13:47.701 clat percentiles (usec): 00:13:47.701 | 1.00th=[ 227], 5.00th=[ 235], 10.00th=[ 239], 20.00th=[ 243], 00:13:47.701 | 30.00th=[ 247], 40.00th=[ 249], 50.00th=[ 253], 60.00th=[ 255], 00:13:47.701 | 70.00th=[ 262], 80.00th=[ 273], 90.00th=[ 416], 95.00th=[ 429], 00:13:47.701 | 99.00th=[ 486], 99.50th=[40633], 99.90th=[41157], 99.95th=[41157], 00:13:47.701 | 99.99th=[41157] 00:13:47.701 write: IOPS=1508, BW=6035KiB/s (6180kB/s)(6144KiB/1018msec); 0 zone resets 00:13:47.701 slat (nsec): min=7106, max=65046, avg=16337.63, stdev=7855.73 00:13:47.701 clat (usec): min=130, max=501, avg=190.60, stdev=38.43 00:13:47.701 lat (usec): min=138, max=535, avg=206.93, stdev=43.72 00:13:47.701 clat percentiles (usec): 00:13:47.701 | 1.00th=[ 135], 5.00th=[ 141], 10.00th=[ 145], 20.00th=[ 153], 00:13:47.701 | 30.00th=[ 163], 40.00th=[ 180], 50.00th=[ 188], 60.00th=[ 194], 00:13:47.701 | 70.00th=[ 208], 80.00th=[ 229], 90.00th=[ 247], 95.00th=[ 258], 00:13:47.701 | 99.00th=[ 273], 99.50th=[ 277], 99.90th=[ 289], 99.95th=[ 502], 00:13:47.701 | 99.99th=[ 502] 00:13:47.701 bw ( KiB/s): min= 4096, max= 8192, per=34.20%, avg=6144.00, stdev=2896.31, samples=2 00:13:47.701 iops : min= 1024, max= 2048, avg=1536.00, stdev=724.08, samples=2 00:13:47.701 lat (usec) : 250=70.16%, 500=29.45%, 750=0.11% 00:13:47.701 lat (msec) : 50=0.29% 00:13:47.701 cpu : usr=2.95%, sys=5.31%, ctx=2779, majf=0, minf=1 00:13:47.701 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:13:47.701 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:47.701 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:47.701 issued rwts: total=1242,1536,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:47.701 latency : target=0, window=0, percentile=100.00%, depth=1 00:13:47.701 job2: (groupid=0, jobs=1): err= 0: pid=166583: Fri Dec 6 17:29:29 2024 00:13:47.701 read: IOPS=21, BW=85.8KiB/s (87.8kB/s)(88.0KiB/1026msec) 00:13:47.701 slat (nsec): min=13285, max=34056, avg=22308.64, stdev=8861.69 00:13:47.701 clat (usec): min=40890, max=42122, avg=41661.93, stdev=504.40 00:13:47.701 lat (usec): min=40903, max=42138, avg=41684.23, stdev=502.92 00:13:47.701 clat percentiles (usec): 00:13:47.701 | 1.00th=[40633], 5.00th=[41157], 10.00th=[41157], 20.00th=[41157], 00:13:47.701 | 30.00th=[41157], 40.00th=[41681], 50.00th=[42206], 60.00th=[42206], 00:13:47.701 | 70.00th=[42206], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:13:47.701 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:13:47.701 | 99.99th=[42206] 00:13:47.701 write: IOPS=499, BW=1996KiB/s (2044kB/s)(2048KiB/1026msec); 0 zone resets 00:13:47.701 slat (nsec): min=6165, max=50661, avg=14706.10, stdev=4839.07 00:13:47.701 clat (usec): min=162, max=257, avg=194.19, stdev=15.53 00:13:47.701 lat (usec): min=169, max=298, avg=208.90, stdev=16.41 00:13:47.701 clat percentiles (usec): 00:13:47.701 | 1.00th=[ 167], 5.00th=[ 172], 10.00th=[ 176], 20.00th=[ 182], 00:13:47.701 | 30.00th=[ 186], 40.00th=[ 190], 50.00th=[ 194], 60.00th=[ 196], 00:13:47.701 | 70.00th=[ 202], 80.00th=[ 206], 90.00th=[ 217], 95.00th=[ 223], 00:13:47.701 | 99.00th=[ 235], 99.50th=[ 247], 99.90th=[ 258], 99.95th=[ 258], 00:13:47.701 | 99.99th=[ 258] 00:13:47.701 bw ( KiB/s): min= 4096, max= 4096, per=22.80%, avg=4096.00, stdev= 0.00, samples=1 00:13:47.701 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:13:47.701 lat (usec) : 250=95.69%, 500=0.19% 00:13:47.701 lat (msec) : 50=4.12% 00:13:47.701 cpu : usr=0.20%, sys=0.88%, ctx=534, majf=0, minf=1 00:13:47.701 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:13:47.701 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:47.701 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:47.701 issued rwts: total=22,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:47.701 latency : target=0, window=0, percentile=100.00%, depth=1 00:13:47.701 job3: (groupid=0, jobs=1): err= 0: pid=166584: Fri Dec 6 17:29:29 2024 00:13:47.701 read: IOPS=22, BW=89.8KiB/s (91.9kB/s)(92.0KiB/1025msec) 00:13:47.701 slat (nsec): min=7984, max=34103, avg=19595.57, stdev=8304.59 00:13:47.701 clat (usec): min=214, max=42042, avg=39692.12, stdev=8622.94 00:13:47.701 lat (usec): min=230, max=42059, avg=39711.71, stdev=8624.07 00:13:47.701 clat percentiles (usec): 00:13:47.701 | 1.00th=[ 215], 5.00th=[40633], 10.00th=[40633], 20.00th=[41157], 00:13:47.701 | 30.00th=[41157], 40.00th=[41157], 50.00th=[41681], 60.00th=[41681], 00:13:47.701 | 70.00th=[42206], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:13:47.701 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:13:47.701 | 99.99th=[42206] 00:13:47.701 write: IOPS=499, BW=1998KiB/s (2046kB/s)(2048KiB/1025msec); 0 zone resets 00:13:47.701 slat (nsec): min=6134, max=54704, avg=17619.19, stdev=7297.01 00:13:47.701 clat (usec): min=138, max=338, avg=195.32, stdev=21.67 00:13:47.701 lat (usec): min=145, max=376, avg=212.94, stdev=24.54 00:13:47.701 clat percentiles (usec): 00:13:47.701 | 1.00th=[ 147], 5.00th=[ 163], 10.00th=[ 169], 20.00th=[ 178], 00:13:47.701 | 30.00th=[ 184], 40.00th=[ 190], 50.00th=[ 196], 60.00th=[ 200], 00:13:47.701 | 70.00th=[ 206], 80.00th=[ 212], 90.00th=[ 221], 95.00th=[ 227], 00:13:47.701 | 99.00th=[ 243], 99.50th=[ 293], 99.90th=[ 338], 99.95th=[ 338], 00:13:47.701 | 99.99th=[ 338] 00:13:47.701 bw ( KiB/s): min= 4096, max= 4096, per=22.80%, avg=4096.00, stdev= 0.00, samples=1 00:13:47.701 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:13:47.701 lat (usec) : 250=95.14%, 500=0.75% 00:13:47.701 lat (msec) : 50=4.11% 00:13:47.701 cpu : usr=1.17%, sys=0.49%, ctx=535, majf=0, minf=1 00:13:47.701 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:13:47.701 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:47.701 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:47.701 issued rwts: total=23,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:47.701 latency : target=0, window=0, percentile=100.00%, depth=1 00:13:47.701 00:13:47.701 Run status group 0 (all jobs): 00:13:47.701 READ: bw=10.8MiB/s (11.4MB/s), 85.8KiB/s-6222KiB/s (87.8kB/s-6371kB/s), io=11.1MiB (11.6MB), run=1001-1026msec 00:13:47.701 WRITE: bw=17.5MiB/s (18.4MB/s), 1996KiB/s-8184KiB/s (2044kB/s-8380kB/s), io=18.0MiB (18.9MB), run=1001-1026msec 00:13:47.701 00:13:47.701 Disk stats (read/write): 00:13:47.701 nvme0n1: ios=1326/1536, merge=0/0, ticks=515/249, in_queue=764, util=86.67% 00:13:47.701 nvme0n2: ios=1182/1536, merge=0/0, ticks=846/275, in_queue=1121, util=97.86% 00:13:47.702 nvme0n3: ios=17/512, merge=0/0, ticks=709/98, in_queue=807, util=88.94% 00:13:47.702 nvme0n4: ios=56/512, merge=0/0, ticks=737/97, in_queue=834, util=90.43% 00:13:47.702 17:29:29 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 128 -t write -r 1 -v 00:13:47.702 [global] 00:13:47.702 thread=1 00:13:47.702 invalidate=1 00:13:47.702 rw=write 00:13:47.702 time_based=1 00:13:47.702 runtime=1 00:13:47.702 ioengine=libaio 00:13:47.702 direct=1 00:13:47.702 bs=4096 00:13:47.702 iodepth=128 00:13:47.702 norandommap=0 00:13:47.702 numjobs=1 00:13:47.702 00:13:47.702 verify_dump=1 00:13:47.702 verify_backlog=512 00:13:47.702 verify_state_save=0 00:13:47.702 do_verify=1 00:13:47.702 verify=crc32c-intel 00:13:47.702 [job0] 00:13:47.702 filename=/dev/nvme0n1 00:13:47.702 [job1] 00:13:47.702 filename=/dev/nvme0n2 00:13:47.702 [job2] 00:13:47.702 filename=/dev/nvme0n3 00:13:47.702 [job3] 00:13:47.702 filename=/dev/nvme0n4 00:13:47.702 Could not set queue depth (nvme0n1) 00:13:47.702 Could not set queue depth (nvme0n2) 00:13:47.702 Could not set queue depth (nvme0n3) 00:13:47.702 Could not set queue depth (nvme0n4) 00:13:47.960 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:13:47.960 job1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:13:47.960 job2: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:13:47.960 job3: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:13:47.960 fio-3.35 00:13:47.960 Starting 4 threads 00:13:49.334 00:13:49.334 job0: (groupid=0, jobs=1): err= 0: pid=166808: Fri Dec 6 17:29:30 2024 00:13:49.334 read: IOPS=4087, BW=16.0MiB/s (16.7MB/s)(16.0MiB/1002msec) 00:13:49.334 slat (usec): min=2, max=11189, avg=107.46, stdev=639.46 00:13:49.334 clat (usec): min=6628, max=30805, avg=14530.70, stdev=4556.86 00:13:49.334 lat (usec): min=6643, max=30921, avg=14638.17, stdev=4595.34 00:13:49.334 clat percentiles (usec): 00:13:49.335 | 1.00th=[ 7898], 5.00th=[10028], 10.00th=[10552], 20.00th=[11207], 00:13:49.335 | 30.00th=[11863], 40.00th=[12387], 50.00th=[12780], 60.00th=[13435], 00:13:49.335 | 70.00th=[15533], 80.00th=[17695], 90.00th=[20841], 95.00th=[23462], 00:13:49.335 | 99.00th=[30802], 99.50th=[30802], 99.90th=[30802], 99.95th=[30802], 00:13:49.335 | 99.99th=[30802] 00:13:49.335 write: IOPS=4239, BW=16.6MiB/s (17.4MB/s)(16.6MiB/1002msec); 0 zone resets 00:13:49.335 slat (usec): min=3, max=21175, avg=119.27, stdev=787.49 00:13:49.335 clat (usec): min=498, max=60824, avg=15406.29, stdev=9645.25 00:13:49.335 lat (usec): min=2914, max=60836, avg=15525.56, stdev=9710.91 00:13:49.335 clat percentiles (usec): 00:13:49.335 | 1.00th=[ 6194], 5.00th=[ 9241], 10.00th=[10290], 20.00th=[10683], 00:13:49.335 | 30.00th=[10945], 40.00th=[11338], 50.00th=[11994], 60.00th=[12780], 00:13:49.335 | 70.00th=[13304], 80.00th=[16712], 90.00th=[24249], 95.00th=[42730], 00:13:49.335 | 99.00th=[54789], 99.50th=[55313], 99.90th=[61080], 99.95th=[61080], 00:13:49.335 | 99.99th=[61080] 00:13:49.335 bw ( KiB/s): min=12288, max=12288, per=19.81%, avg=12288.00, stdev= 0.00, samples=1 00:13:49.335 iops : min= 3072, max= 3072, avg=3072.00, stdev= 0.00, samples=1 00:13:49.335 lat (usec) : 500=0.01% 00:13:49.335 lat (msec) : 4=0.42%, 10=5.32%, 20=77.98%, 50=15.00%, 100=1.26% 00:13:49.335 cpu : usr=4.50%, sys=8.39%, ctx=386, majf=0, minf=1 00:13:49.335 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.4%, >=64=99.2% 00:13:49.335 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:49.335 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:13:49.335 issued rwts: total=4096,4248,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:49.335 latency : target=0, window=0, percentile=100.00%, depth=128 00:13:49.335 job1: (groupid=0, jobs=1): err= 0: pid=166809: Fri Dec 6 17:29:30 2024 00:13:49.335 read: IOPS=4014, BW=15.7MiB/s (16.4MB/s)(15.8MiB/1006msec) 00:13:49.335 slat (usec): min=2, max=29869, avg=119.03, stdev=986.64 00:13:49.335 clat (usec): min=2150, max=59636, avg=15025.21, stdev=7738.72 00:13:49.335 lat (usec): min=2190, max=59650, avg=15144.24, stdev=7818.30 00:13:49.335 clat percentiles (usec): 00:13:49.335 | 1.00th=[ 3818], 5.00th=[ 7635], 10.00th=[ 8455], 20.00th=[10290], 00:13:49.335 | 30.00th=[10945], 40.00th=[11600], 50.00th=[12518], 60.00th=[13173], 00:13:49.335 | 70.00th=[14615], 80.00th=[19268], 90.00th=[27657], 95.00th=[33817], 00:13:49.335 | 99.00th=[39060], 99.50th=[40109], 99.90th=[44303], 99.95th=[44303], 00:13:49.335 | 99.99th=[59507] 00:13:49.335 write: IOPS=4071, BW=15.9MiB/s (16.7MB/s)(16.0MiB/1006msec); 0 zone resets 00:13:49.335 slat (usec): min=3, max=11779, avg=115.34, stdev=709.37 00:13:49.335 clat (usec): min=573, max=44058, avg=16330.02, stdev=9333.79 00:13:49.335 lat (usec): min=600, max=44067, avg=16445.36, stdev=9406.24 00:13:49.335 clat percentiles (usec): 00:13:49.335 | 1.00th=[ 4555], 5.00th=[ 7177], 10.00th=[ 7963], 20.00th=[ 9503], 00:13:49.335 | 30.00th=[10159], 40.00th=[10683], 50.00th=[11469], 60.00th=[13566], 00:13:49.335 | 70.00th=[21365], 80.00th=[25822], 90.00th=[32900], 95.00th=[35390], 00:13:49.335 | 99.00th=[38536], 99.50th=[39584], 99.90th=[41157], 99.95th=[41157], 00:13:49.335 | 99.99th=[44303] 00:13:49.335 bw ( KiB/s): min=14728, max=18040, per=26.42%, avg=16384.00, stdev=2341.94, samples=2 00:13:49.335 iops : min= 3682, max= 4510, avg=4096.00, stdev=585.48, samples=2 00:13:49.335 lat (usec) : 750=0.05% 00:13:49.335 lat (msec) : 2=0.09%, 4=0.86%, 10=21.55%, 20=53.09%, 50=24.35% 00:13:49.335 lat (msec) : 100=0.01% 00:13:49.335 cpu : usr=3.18%, sys=6.07%, ctx=305, majf=0, minf=1 00:13:49.335 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.4%, >=64=99.2% 00:13:49.335 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:49.335 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:13:49.335 issued rwts: total=4039,4096,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:49.335 latency : target=0, window=0, percentile=100.00%, depth=128 00:13:49.335 job2: (groupid=0, jobs=1): err= 0: pid=166816: Fri Dec 6 17:29:30 2024 00:13:49.335 read: IOPS=2544, BW=9.94MiB/s (10.4MB/s)(10.0MiB/1006msec) 00:13:49.335 slat (usec): min=3, max=43998, avg=229.15, stdev=1837.64 00:13:49.335 clat (msec): min=10, max=128, avg=26.93, stdev=25.87 00:13:49.335 lat (msec): min=10, max=128, avg=27.16, stdev=26.03 00:13:49.335 clat percentiles (msec): 00:13:49.335 | 1.00th=[ 11], 5.00th=[ 12], 10.00th=[ 13], 20.00th=[ 13], 00:13:49.335 | 30.00th=[ 14], 40.00th=[ 17], 50.00th=[ 20], 60.00th=[ 23], 00:13:49.335 | 70.00th=[ 23], 80.00th=[ 27], 90.00th=[ 55], 95.00th=[ 99], 00:13:49.335 | 99.00th=[ 129], 99.50th=[ 129], 99.90th=[ 129], 99.95th=[ 129], 00:13:49.335 | 99.99th=[ 129] 00:13:49.335 write: IOPS=2768, BW=10.8MiB/s (11.3MB/s)(10.9MiB/1006msec); 0 zone resets 00:13:49.335 slat (usec): min=2, max=28877, avg=136.69, stdev=1054.24 00:13:49.335 clat (usec): min=856, max=85654, avg=21068.52, stdev=16437.33 00:13:49.335 lat (usec): min=863, max=85693, avg=21205.20, stdev=16514.29 00:13:49.335 clat percentiles (usec): 00:13:49.335 | 1.00th=[ 3097], 5.00th=[ 9372], 10.00th=[10159], 20.00th=[11207], 00:13:49.335 | 30.00th=[12649], 40.00th=[13435], 50.00th=[14222], 60.00th=[16581], 00:13:49.335 | 70.00th=[16909], 80.00th=[25297], 90.00th=[53740], 95.00th=[60556], 00:13:49.335 | 99.00th=[71828], 99.50th=[71828], 99.90th=[82314], 99.95th=[84411], 00:13:49.335 | 99.99th=[85459] 00:13:49.335 bw ( KiB/s): min= 5384, max=15880, per=17.14%, avg=10632.00, stdev=7421.79, samples=2 00:13:49.335 iops : min= 1346, max= 3970, avg=2658.00, stdev=1855.45, samples=2 00:13:49.335 lat (usec) : 1000=0.13% 00:13:49.335 lat (msec) : 2=0.15%, 4=0.37%, 10=4.27%, 20=59.53%, 50=24.27% 00:13:49.335 lat (msec) : 100=8.92%, 250=2.36% 00:13:49.335 cpu : usr=4.58%, sys=6.07%, ctx=170, majf=0, minf=2 00:13:49.335 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.3%, 32=0.6%, >=64=98.8% 00:13:49.335 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:49.335 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:13:49.335 issued rwts: total=2560,2785,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:49.335 latency : target=0, window=0, percentile=100.00%, depth=128 00:13:49.335 job3: (groupid=0, jobs=1): err= 0: pid=166817: Fri Dec 6 17:29:30 2024 00:13:49.335 read: IOPS=4087, BW=16.0MiB/s (16.7MB/s)(16.0MiB/1002msec) 00:13:49.335 slat (usec): min=3, max=14747, avg=105.81, stdev=628.87 00:13:49.335 clat (usec): min=4593, max=38038, avg=13782.35, stdev=4104.35 00:13:49.335 lat (usec): min=4597, max=38056, avg=13888.16, stdev=4155.13 00:13:49.335 clat percentiles (usec): 00:13:49.335 | 1.00th=[ 8848], 5.00th=[10421], 10.00th=[11076], 20.00th=[11731], 00:13:49.335 | 30.00th=[11994], 40.00th=[12256], 50.00th=[12518], 60.00th=[12780], 00:13:49.335 | 70.00th=[13960], 80.00th=[14877], 90.00th=[17695], 95.00th=[19268], 00:13:49.335 | 99.00th=[33817], 99.50th=[35390], 99.90th=[38011], 99.95th=[38011], 00:13:49.335 | 99.99th=[38011] 00:13:49.335 write: IOPS=4459, BW=17.4MiB/s (18.3MB/s)(17.5MiB/1002msec); 0 zone resets 00:13:49.335 slat (usec): min=4, max=12302, avg=114.62, stdev=568.04 00:13:49.335 clat (usec): min=1180, max=40722, avg=15776.75, stdev=6897.69 00:13:49.335 lat (usec): min=1198, max=40742, avg=15891.37, stdev=6954.16 00:13:49.335 clat percentiles (usec): 00:13:49.335 | 1.00th=[ 5866], 5.00th=[ 9896], 10.00th=[11338], 20.00th=[12125], 00:13:49.335 | 30.00th=[12649], 40.00th=[12780], 50.00th=[13042], 60.00th=[13435], 00:13:49.335 | 70.00th=[13829], 80.00th=[17695], 90.00th=[28443], 95.00th=[30802], 00:13:49.335 | 99.00th=[37487], 99.50th=[39060], 99.90th=[40633], 99.95th=[40633], 00:13:49.335 | 99.99th=[40633] 00:13:49.335 bw ( KiB/s): min=15664, max=15664, per=25.26%, avg=15664.00, stdev= 0.00, samples=1 00:13:49.335 iops : min= 3916, max= 3916, avg=3916.00, stdev= 0.00, samples=1 00:13:49.335 lat (msec) : 2=0.12%, 4=0.16%, 10=4.20%, 20=83.45%, 50=12.06% 00:13:49.335 cpu : usr=8.19%, sys=9.99%, ctx=457, majf=0, minf=1 00:13:49.335 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.4%, >=64=99.3% 00:13:49.335 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:49.335 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:13:49.335 issued rwts: total=4096,4468,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:49.335 latency : target=0, window=0, percentile=100.00%, depth=128 00:13:49.335 00:13:49.335 Run status group 0 (all jobs): 00:13:49.336 READ: bw=57.4MiB/s (60.2MB/s), 9.94MiB/s-16.0MiB/s (10.4MB/s-16.7MB/s), io=57.8MiB (60.6MB), run=1002-1006msec 00:13:49.336 WRITE: bw=60.6MiB/s (63.5MB/s), 10.8MiB/s-17.4MiB/s (11.3MB/s-18.3MB/s), io=60.9MiB (63.9MB), run=1002-1006msec 00:13:49.336 00:13:49.336 Disk stats (read/write): 00:13:49.336 nvme0n1: ios=3202/3584, merge=0/0, ticks=24978/28224, in_queue=53202, util=98.70% 00:13:49.336 nvme0n2: ios=3369/3584, merge=0/0, ticks=49140/54557, in_queue=103697, util=96.65% 00:13:49.336 nvme0n3: ios=2080/2560, merge=0/0, ticks=21036/14930, in_queue=35966, util=88.70% 00:13:49.336 nvme0n4: ios=3485/3584, merge=0/0, ticks=30834/41217, in_queue=72051, util=96.52% 00:13:49.336 17:29:30 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 128 -t randwrite -r 1 -v 00:13:49.336 [global] 00:13:49.336 thread=1 00:13:49.336 invalidate=1 00:13:49.336 rw=randwrite 00:13:49.336 time_based=1 00:13:49.336 runtime=1 00:13:49.336 ioengine=libaio 00:13:49.336 direct=1 00:13:49.336 bs=4096 00:13:49.336 iodepth=128 00:13:49.336 norandommap=0 00:13:49.336 numjobs=1 00:13:49.336 00:13:49.336 verify_dump=1 00:13:49.336 verify_backlog=512 00:13:49.336 verify_state_save=0 00:13:49.336 do_verify=1 00:13:49.336 verify=crc32c-intel 00:13:49.336 [job0] 00:13:49.336 filename=/dev/nvme0n1 00:13:49.336 [job1] 00:13:49.336 filename=/dev/nvme0n2 00:13:49.336 [job2] 00:13:49.336 filename=/dev/nvme0n3 00:13:49.336 [job3] 00:13:49.336 filename=/dev/nvme0n4 00:13:49.336 Could not set queue depth (nvme0n1) 00:13:49.336 Could not set queue depth (nvme0n2) 00:13:49.336 Could not set queue depth (nvme0n3) 00:13:49.336 Could not set queue depth (nvme0n4) 00:13:49.336 job0: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:13:49.336 job1: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:13:49.336 job2: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:13:49.336 job3: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:13:49.336 fio-3.35 00:13:49.336 Starting 4 threads 00:13:50.710 00:13:50.710 job0: (groupid=0, jobs=1): err= 0: pid=167045: Fri Dec 6 17:29:32 2024 00:13:50.710 read: IOPS=5093, BW=19.9MiB/s (20.9MB/s)(20.1MiB/1010msec) 00:13:50.710 slat (usec): min=2, max=11825, avg=90.93, stdev=626.02 00:13:50.710 clat (usec): min=3220, max=38985, avg=11875.09, stdev=3533.88 00:13:50.710 lat (usec): min=3823, max=38994, avg=11966.02, stdev=3573.45 00:13:50.710 clat percentiles (usec): 00:13:50.710 | 1.00th=[ 4752], 5.00th=[ 7701], 10.00th=[ 8586], 20.00th=[10028], 00:13:50.710 | 30.00th=[10421], 40.00th=[10683], 50.00th=[11338], 60.00th=[12125], 00:13:50.710 | 70.00th=[12518], 80.00th=[13304], 90.00th=[14746], 95.00th=[17695], 00:13:50.710 | 99.00th=[27919], 99.50th=[34866], 99.90th=[38536], 99.95th=[39060], 00:13:50.710 | 99.99th=[39060] 00:13:50.710 write: IOPS=5576, BW=21.8MiB/s (22.8MB/s)(22.0MiB/1010msec); 0 zone resets 00:13:50.710 slat (usec): min=3, max=11253, avg=82.21, stdev=513.30 00:13:50.710 clat (usec): min=243, max=38902, avg=11912.54, stdev=4837.00 00:13:50.710 lat (usec): min=500, max=38919, avg=11994.74, stdev=4878.08 00:13:50.710 clat percentiles (usec): 00:13:50.710 | 1.00th=[ 2999], 5.00th=[ 5604], 10.00th=[ 6915], 20.00th=[ 9372], 00:13:50.710 | 30.00th=[10290], 40.00th=[10552], 50.00th=[11076], 60.00th=[11338], 00:13:50.710 | 70.00th=[12256], 80.00th=[13042], 90.00th=[19792], 95.00th=[21890], 00:13:50.710 | 99.00th=[30016], 99.50th=[30540], 99.90th=[32113], 99.95th=[32375], 00:13:50.710 | 99.99th=[39060] 00:13:50.710 bw ( KiB/s): min=19648, max=24576, per=33.43%, avg=22112.00, stdev=3484.62, samples=2 00:13:50.710 iops : min= 4912, max= 6144, avg=5528.00, stdev=871.16, samples=2 00:13:50.710 lat (usec) : 250=0.01%, 500=0.03%, 750=0.01%, 1000=0.03% 00:13:50.710 lat (msec) : 2=0.26%, 4=0.79%, 10=21.71%, 20=71.24%, 50=5.93% 00:13:50.710 cpu : usr=4.26%, sys=9.12%, ctx=519, majf=0, minf=1 00:13:50.710 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.3%, >=64=99.4% 00:13:50.710 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:50.710 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:13:50.710 issued rwts: total=5144,5632,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:50.710 latency : target=0, window=0, percentile=100.00%, depth=128 00:13:50.710 job1: (groupid=0, jobs=1): err= 0: pid=167046: Fri Dec 6 17:29:32 2024 00:13:50.710 read: IOPS=3548, BW=13.9MiB/s (14.5MB/s)(14.0MiB/1010msec) 00:13:50.710 slat (usec): min=2, max=21404, avg=118.83, stdev=897.90 00:13:50.710 clat (usec): min=1913, max=38862, avg=15455.37, stdev=6155.73 00:13:50.710 lat (usec): min=1917, max=39124, avg=15574.20, stdev=6210.94 00:13:50.710 clat percentiles (usec): 00:13:50.710 | 1.00th=[ 3949], 5.00th=[ 8586], 10.00th=[10814], 20.00th=[11731], 00:13:50.710 | 30.00th=[12518], 40.00th=[12911], 50.00th=[13698], 60.00th=[14353], 00:13:50.710 | 70.00th=[15664], 80.00th=[19792], 90.00th=[23200], 95.00th=[28705], 00:13:50.710 | 99.00th=[38011], 99.50th=[38536], 99.90th=[39060], 99.95th=[39060], 00:13:50.710 | 99.99th=[39060] 00:13:50.710 write: IOPS=3863, BW=15.1MiB/s (15.8MB/s)(15.2MiB/1010msec); 0 zone resets 00:13:50.710 slat (usec): min=3, max=15964, avg=137.64, stdev=892.91 00:13:50.710 clat (usec): min=1251, max=55036, avg=18631.59, stdev=9532.11 00:13:50.710 lat (usec): min=1382, max=55042, avg=18769.23, stdev=9583.32 00:13:50.710 clat percentiles (usec): 00:13:50.710 | 1.00th=[ 4817], 5.00th=[ 8586], 10.00th=[ 9896], 20.00th=[10945], 00:13:50.710 | 30.00th=[11994], 40.00th=[13304], 50.00th=[14877], 60.00th=[19530], 00:13:50.710 | 70.00th=[23987], 80.00th=[26084], 90.00th=[28443], 95.00th=[38536], 00:13:50.710 | 99.00th=[51643], 99.50th=[54789], 99.90th=[54789], 99.95th=[54789], 00:13:50.710 | 99.99th=[54789] 00:13:50.710 bw ( KiB/s): min=15064, max=15128, per=22.82%, avg=15096.00, stdev=45.25, samples=2 00:13:50.710 iops : min= 3766, max= 3782, avg=3774.00, stdev=11.31, samples=2 00:13:50.710 lat (msec) : 2=0.27%, 4=0.69%, 10=8.86%, 20=60.29%, 50=29.07% 00:13:50.710 lat (msec) : 100=0.83% 00:13:50.710 cpu : usr=2.87%, sys=5.25%, ctx=316, majf=0, minf=1 00:13:50.711 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.4%, >=64=99.2% 00:13:50.711 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:50.711 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:13:50.711 issued rwts: total=3584,3902,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:50.711 latency : target=0, window=0, percentile=100.00%, depth=128 00:13:50.711 job2: (groupid=0, jobs=1): err= 0: pid=167048: Fri Dec 6 17:29:32 2024 00:13:50.711 read: IOPS=2355, BW=9424KiB/s (9650kB/s)(9452KiB/1003msec) 00:13:50.711 slat (usec): min=3, max=39316, avg=203.50, stdev=1517.40 00:13:50.711 clat (msec): min=2, max=111, avg=24.39, stdev=19.97 00:13:50.711 lat (msec): min=2, max=111, avg=24.60, stdev=20.12 00:13:50.711 clat percentiles (msec): 00:13:50.711 | 1.00th=[ 3], 5.00th=[ 12], 10.00th=[ 13], 20.00th=[ 14], 00:13:50.711 | 30.00th=[ 15], 40.00th=[ 15], 50.00th=[ 16], 60.00th=[ 21], 00:13:50.711 | 70.00th=[ 27], 80.00th=[ 28], 90.00th=[ 46], 95.00th=[ 61], 00:13:50.711 | 99.00th=[ 112], 99.50th=[ 112], 99.90th=[ 112], 99.95th=[ 112], 00:13:50.711 | 99.99th=[ 112] 00:13:50.711 write: IOPS=2552, BW=9.97MiB/s (10.5MB/s)(10.0MiB/1003msec); 0 zone resets 00:13:50.711 slat (usec): min=3, max=19590, avg=192.04, stdev=1190.71 00:13:50.711 clat (msec): min=8, max=108, avg=26.95, stdev=20.91 00:13:50.711 lat (msec): min=9, max=109, avg=27.14, stdev=20.98 00:13:50.711 clat percentiles (msec): 00:13:50.711 | 1.00th=[ 11], 5.00th=[ 11], 10.00th=[ 12], 20.00th=[ 13], 00:13:50.711 | 30.00th=[ 14], 40.00th=[ 14], 50.00th=[ 15], 60.00th=[ 25], 00:13:50.711 | 70.00th=[ 29], 80.00th=[ 45], 90.00th=[ 59], 95.00th=[ 69], 00:13:50.711 | 99.00th=[ 109], 99.50th=[ 109], 99.90th=[ 109], 99.95th=[ 109], 00:13:50.711 | 99.99th=[ 109] 00:13:50.711 bw ( KiB/s): min= 9480, max=11000, per=15.48%, avg=10240.00, stdev=1074.80, samples=2 00:13:50.711 iops : min= 2370, max= 2750, avg=2560.00, stdev=268.70, samples=2 00:13:50.711 lat (msec) : 4=0.55%, 10=1.46%, 20=55.19%, 50=30.10%, 100=10.81% 00:13:50.711 lat (msec) : 250=1.89% 00:13:50.711 cpu : usr=4.49%, sys=4.59%, ctx=250, majf=0, minf=1 00:13:50.711 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.2%, 16=0.3%, 32=0.7%, >=64=98.7% 00:13:50.711 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:50.711 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:13:50.711 issued rwts: total=2363,2560,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:50.711 latency : target=0, window=0, percentile=100.00%, depth=128 00:13:50.711 job3: (groupid=0, jobs=1): err= 0: pid=167050: Fri Dec 6 17:29:32 2024 00:13:50.711 read: IOPS=4460, BW=17.4MiB/s (18.3MB/s)(17.6MiB/1010msec) 00:13:50.711 slat (usec): min=2, max=10555, avg=104.45, stdev=535.42 00:13:50.711 clat (usec): min=1259, max=31395, avg=13236.72, stdev=2550.98 00:13:50.711 lat (usec): min=8041, max=31398, avg=13341.17, stdev=2545.91 00:13:50.711 clat percentiles (usec): 00:13:50.711 | 1.00th=[ 8455], 5.00th=[10159], 10.00th=[10814], 20.00th=[11994], 00:13:50.711 | 30.00th=[12256], 40.00th=[12649], 50.00th=[13042], 60.00th=[13173], 00:13:50.711 | 70.00th=[13566], 80.00th=[14091], 90.00th=[15008], 95.00th=[18220], 00:13:50.711 | 99.00th=[21627], 99.50th=[28443], 99.90th=[31327], 99.95th=[31327], 00:13:50.711 | 99.99th=[31327] 00:13:50.711 write: IOPS=4562, BW=17.8MiB/s (18.7MB/s)(18.0MiB/1010msec); 0 zone resets 00:13:50.711 slat (usec): min=3, max=23040, avg=106.51, stdev=678.62 00:13:50.711 clat (usec): min=7163, max=41952, avg=14726.88, stdev=5824.98 00:13:50.711 lat (usec): min=7197, max=43944, avg=14833.39, stdev=5853.19 00:13:50.711 clat percentiles (usec): 00:13:50.711 | 1.00th=[ 9110], 5.00th=[10028], 10.00th=[10421], 20.00th=[11863], 00:13:50.711 | 30.00th=[12518], 40.00th=[12780], 50.00th=[13042], 60.00th=[13304], 00:13:50.711 | 70.00th=[13698], 80.00th=[15926], 90.00th=[21627], 95.00th=[32113], 00:13:50.711 | 99.00th=[41681], 99.50th=[41681], 99.90th=[41681], 99.95th=[42206], 00:13:50.711 | 99.99th=[42206] 00:13:50.711 bw ( KiB/s): min=18392, max=18472, per=27.87%, avg=18432.00, stdev=56.57, samples=2 00:13:50.711 iops : min= 4598, max= 4618, avg=4608.00, stdev=14.14, samples=2 00:13:50.711 lat (msec) : 2=0.01%, 10=4.56%, 20=88.81%, 50=6.62% 00:13:50.711 cpu : usr=4.96%, sys=10.31%, ctx=458, majf=0, minf=1 00:13:50.711 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.4%, >=64=99.3% 00:13:50.711 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:50.711 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:13:50.711 issued rwts: total=4505,4608,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:50.711 latency : target=0, window=0, percentile=100.00%, depth=128 00:13:50.711 00:13:50.711 Run status group 0 (all jobs): 00:13:50.711 READ: bw=60.3MiB/s (63.2MB/s), 9424KiB/s-19.9MiB/s (9650kB/s-20.9MB/s), io=60.9MiB (63.9MB), run=1003-1010msec 00:13:50.711 WRITE: bw=64.6MiB/s (67.7MB/s), 9.97MiB/s-21.8MiB/s (10.5MB/s-22.8MB/s), io=65.2MiB (68.4MB), run=1003-1010msec 00:13:50.711 00:13:50.711 Disk stats (read/write): 00:13:50.711 nvme0n1: ios=4471/4608, merge=0/0, ticks=41427/42947, in_queue=84374, util=95.89% 00:13:50.711 nvme0n2: ios=3122/3517, merge=0/0, ticks=43023/52600, in_queue=95623, util=96.11% 00:13:50.711 nvme0n3: ios=1536/1884, merge=0/0, ticks=12144/13345, in_queue=25489, util=88.69% 00:13:50.711 nvme0n4: ios=3604/4095, merge=0/0, ticks=15202/16143, in_queue=31345, util=95.91% 00:13:50.711 17:29:32 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@55 -- # sync 00:13:50.711 17:29:32 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@59 -- # fio_pid=167191 00:13:50.711 17:29:32 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t read -r 10 00:13:50.711 17:29:32 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@61 -- # sleep 3 00:13:50.711 [global] 00:13:50.711 thread=1 00:13:50.711 invalidate=1 00:13:50.711 rw=read 00:13:50.711 time_based=1 00:13:50.711 runtime=10 00:13:50.711 ioengine=libaio 00:13:50.711 direct=1 00:13:50.711 bs=4096 00:13:50.711 iodepth=1 00:13:50.711 norandommap=1 00:13:50.711 numjobs=1 00:13:50.711 00:13:50.711 [job0] 00:13:50.711 filename=/dev/nvme0n1 00:13:50.711 [job1] 00:13:50.711 filename=/dev/nvme0n2 00:13:50.711 [job2] 00:13:50.711 filename=/dev/nvme0n3 00:13:50.711 [job3] 00:13:50.711 filename=/dev/nvme0n4 00:13:50.711 Could not set queue depth (nvme0n1) 00:13:50.711 Could not set queue depth (nvme0n2) 00:13:50.711 Could not set queue depth (nvme0n3) 00:13:50.711 Could not set queue depth (nvme0n4) 00:13:50.711 job0: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:13:50.711 job1: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:13:50.711 job2: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:13:50.711 job3: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:13:50.711 fio-3.35 00:13:50.711 Starting 4 threads 00:13:53.987 17:29:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_delete concat0 00:13:53.987 17:29:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_delete raid0 00:13:53.987 fio: io_u error on file /dev/nvme0n4: Operation not supported: read offset=299008, buflen=4096 00:13:53.987 fio: pid=167283, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:13:54.245 17:29:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:13:54.245 17:29:35 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc0 00:13:54.245 fio: io_u error on file /dev/nvme0n3: Operation not supported: read offset=4403200, buflen=4096 00:13:54.245 fio: pid=167282, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:13:54.503 fio: io_u error on file /dev/nvme0n1: Operation not supported: read offset=10231808, buflen=4096 00:13:54.503 fio: pid=167280, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:13:54.503 17:29:36 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:13:54.503 17:29:36 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc1 00:13:54.763 17:29:36 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:13:54.763 17:29:36 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc2 00:13:54.763 fio: io_u error on file /dev/nvme0n2: Operation not supported: read offset=380928, buflen=4096 00:13:54.763 fio: pid=167281, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:13:54.763 00:13:54.763 job0: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=167280: Fri Dec 6 17:29:36 2024 00:13:54.763 read: IOPS=710, BW=2841KiB/s (2909kB/s)(9992KiB/3517msec) 00:13:54.763 slat (usec): min=3, max=34176, avg=28.92, stdev=769.16 00:13:54.763 clat (usec): min=153, max=42002, avg=1373.75, stdev=6842.32 00:13:54.763 lat (usec): min=158, max=42017, avg=1402.68, stdev=6884.74 00:13:54.763 clat percentiles (usec): 00:13:54.763 | 1.00th=[ 159], 5.00th=[ 163], 10.00th=[ 165], 20.00th=[ 169], 00:13:54.763 | 30.00th=[ 174], 40.00th=[ 176], 50.00th=[ 180], 60.00th=[ 184], 00:13:54.763 | 70.00th=[ 190], 80.00th=[ 200], 90.00th=[ 253], 95.00th=[ 383], 00:13:54.763 | 99.00th=[41157], 99.50th=[41157], 99.90th=[42206], 99.95th=[42206], 00:13:54.763 | 99.99th=[42206] 00:13:54.763 bw ( KiB/s): min= 96, max= 696, per=6.28%, avg=246.67, stdev=247.79, samples=6 00:13:54.763 iops : min= 24, max= 174, avg=61.67, stdev=61.95, samples=6 00:13:54.763 lat (usec) : 250=89.48%, 500=7.04%, 750=0.52% 00:13:54.763 lat (msec) : 10=0.04%, 50=2.88% 00:13:54.763 cpu : usr=0.17%, sys=0.46%, ctx=2502, majf=0, minf=1 00:13:54.763 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:13:54.763 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:54.763 complete : 0=0.1%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:54.763 issued rwts: total=2499,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:54.763 latency : target=0, window=0, percentile=100.00%, depth=1 00:13:54.763 job1: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=167281: Fri Dec 6 17:29:36 2024 00:13:54.763 read: IOPS=24, BW=97.4KiB/s (99.7kB/s)(372KiB/3820msec) 00:13:54.763 slat (usec): min=11, max=14891, avg=189.67, stdev=1535.40 00:13:54.763 clat (usec): min=371, max=42043, avg=40616.57, stdev=4227.79 00:13:54.763 lat (usec): min=413, max=56009, avg=40807.92, stdev=4517.77 00:13:54.763 clat percentiles (usec): 00:13:54.763 | 1.00th=[ 371], 5.00th=[40633], 10.00th=[41157], 20.00th=[41157], 00:13:54.763 | 30.00th=[41157], 40.00th=[41157], 50.00th=[41157], 60.00th=[41157], 00:13:54.763 | 70.00th=[41157], 80.00th=[41157], 90.00th=[41157], 95.00th=[42206], 00:13:54.763 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:13:54.763 | 99.99th=[42206] 00:13:54.763 bw ( KiB/s): min= 93, max= 104, per=2.48%, avg=97.86, stdev= 4.34, samples=7 00:13:54.763 iops : min= 23, max= 26, avg=24.43, stdev= 1.13, samples=7 00:13:54.763 lat (usec) : 500=1.06% 00:13:54.763 lat (msec) : 50=97.87% 00:13:54.763 cpu : usr=0.10%, sys=0.00%, ctx=96, majf=0, minf=2 00:13:54.763 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:13:54.763 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:54.763 complete : 0=1.1%, 4=98.9%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:54.763 issued rwts: total=94,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:54.763 latency : target=0, window=0, percentile=100.00%, depth=1 00:13:54.763 job2: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=167282: Fri Dec 6 17:29:36 2024 00:13:54.763 read: IOPS=331, BW=1326KiB/s (1358kB/s)(4300KiB/3243msec) 00:13:54.763 slat (usec): min=7, max=9931, avg=28.27, stdev=302.21 00:13:54.763 clat (usec): min=215, max=42195, avg=2957.15, stdev=10151.49 00:13:54.763 lat (usec): min=240, max=50944, avg=2985.43, stdev=10190.69 00:13:54.763 clat percentiles (usec): 00:13:54.763 | 1.00th=[ 229], 5.00th=[ 237], 10.00th=[ 241], 20.00th=[ 247], 00:13:54.763 | 30.00th=[ 249], 40.00th=[ 253], 50.00th=[ 255], 60.00th=[ 260], 00:13:54.763 | 70.00th=[ 265], 80.00th=[ 269], 90.00th=[ 289], 95.00th=[41157], 00:13:54.763 | 99.00th=[41157], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:13:54.763 | 99.99th=[42206] 00:13:54.763 bw ( KiB/s): min= 104, max= 3960, per=36.42%, avg=1426.67, stdev=1616.90, samples=6 00:13:54.763 iops : min= 26, max= 990, avg=356.67, stdev=404.23, samples=6 00:13:54.763 lat (usec) : 250=34.11%, 500=58.92%, 750=0.19%, 1000=0.09% 00:13:54.763 lat (msec) : 50=6.60% 00:13:54.763 cpu : usr=0.52%, sys=0.83%, ctx=1079, majf=0, minf=2 00:13:54.763 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:13:54.763 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:54.763 complete : 0=0.1%, 4=99.9%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:54.763 issued rwts: total=1076,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:54.763 latency : target=0, window=0, percentile=100.00%, depth=1 00:13:54.763 job3: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=167283: Fri Dec 6 17:29:36 2024 00:13:54.763 read: IOPS=24, BW=97.9KiB/s (100kB/s)(292KiB/2982msec) 00:13:54.763 slat (nsec): min=10058, max=35902, avg=14270.54, stdev=2621.72 00:13:54.763 clat (usec): min=302, max=42022, avg=40523.60, stdev=4782.51 00:13:54.763 lat (usec): min=338, max=42037, avg=40537.85, stdev=4779.95 00:13:54.763 clat percentiles (usec): 00:13:54.763 | 1.00th=[ 302], 5.00th=[41157], 10.00th=[41157], 20.00th=[41157], 00:13:54.763 | 30.00th=[41157], 40.00th=[41157], 50.00th=[41157], 60.00th=[41157], 00:13:54.763 | 70.00th=[41157], 80.00th=[41157], 90.00th=[41157], 95.00th=[42206], 00:13:54.763 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:13:54.763 | 99.99th=[42206] 00:13:54.763 bw ( KiB/s): min= 96, max= 104, per=2.48%, avg=97.60, stdev= 3.58, samples=5 00:13:54.763 iops : min= 24, max= 26, avg=24.40, stdev= 0.89, samples=5 00:13:54.763 lat (usec) : 500=1.35% 00:13:54.764 lat (msec) : 50=97.30% 00:13:54.764 cpu : usr=0.07%, sys=0.00%, ctx=74, majf=0, minf=1 00:13:54.764 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:13:54.764 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:54.764 complete : 0=1.3%, 4=98.7%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:54.764 issued rwts: total=74,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:54.764 latency : target=0, window=0, percentile=100.00%, depth=1 00:13:54.764 00:13:54.764 Run status group 0 (all jobs): 00:13:54.764 READ: bw=3915KiB/s (4009kB/s), 97.4KiB/s-2841KiB/s (99.7kB/s-2909kB/s), io=14.6MiB (15.3MB), run=2982-3820msec 00:13:54.764 00:13:54.764 Disk stats (read/write): 00:13:54.764 nvme0n1: ios=1640/0, merge=0/0, ticks=3276/0, in_queue=3276, util=94.25% 00:13:54.764 nvme0n2: ios=88/0, merge=0/0, ticks=3575/0, in_queue=3575, util=96.11% 00:13:54.764 nvme0n3: ios=1118/0, merge=0/0, ticks=3482/0, in_queue=3482, util=99.09% 00:13:54.764 nvme0n4: ios=70/0, merge=0/0, ticks=2837/0, in_queue=2837, util=96.74% 00:13:55.022 17:29:36 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:13:55.022 17:29:36 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc3 00:13:55.281 17:29:36 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:13:55.281 17:29:36 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc4 00:13:55.539 17:29:37 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:13:55.539 17:29:37 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc5 00:13:55.798 17:29:37 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:13:55.798 17:29:37 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc6 00:13:56.056 17:29:37 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@69 -- # fio_status=0 00:13:56.056 17:29:37 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@70 -- # wait 167191 00:13:56.056 17:29:37 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@70 -- # fio_status=4 00:13:56.056 17:29:37 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@72 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:13:56.315 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:13:56.315 17:29:37 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@73 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:13:56.315 17:29:37 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1223 -- # local i=0 00:13:56.315 17:29:37 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:13:56.315 17:29:37 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:13:56.315 17:29:37 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:13:56.315 17:29:37 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:13:56.315 17:29:37 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1235 -- # return 0 00:13:56.315 17:29:37 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@75 -- # '[' 4 -eq 0 ']' 00:13:56.315 17:29:37 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@80 -- # echo 'nvmf hotplug test: fio failed as expected' 00:13:56.315 nvmf hotplug test: fio failed as expected 00:13:56.315 17:29:37 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:13:56.574 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@85 -- # rm -f ./local-job0-0-verify.state 00:13:56.574 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@86 -- # rm -f ./local-job1-1-verify.state 00:13:56.574 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@87 -- # rm -f ./local-job2-2-verify.state 00:13:56.574 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@89 -- # trap - SIGINT SIGTERM EXIT 00:13:56.574 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@91 -- # nvmftestfini 00:13:56.574 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@516 -- # nvmfcleanup 00:13:56.574 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@121 -- # sync 00:13:56.574 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:13:56.574 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@124 -- # set +e 00:13:56.574 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@125 -- # for i in {1..20} 00:13:56.574 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:13:56.574 rmmod nvme_tcp 00:13:56.574 rmmod nvme_fabrics 00:13:56.574 rmmod nvme_keyring 00:13:56.574 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:13:56.574 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@128 -- # set -e 00:13:56.574 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@129 -- # return 0 00:13:56.574 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@517 -- # '[' -n 165147 ']' 00:13:56.574 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@518 -- # killprocess 165147 00:13:56.574 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@954 -- # '[' -z 165147 ']' 00:13:56.574 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@958 -- # kill -0 165147 00:13:56.574 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@959 -- # uname 00:13:56.574 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:13:56.574 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 165147 00:13:56.574 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:13:56.574 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:13:56.574 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@972 -- # echo 'killing process with pid 165147' 00:13:56.574 killing process with pid 165147 00:13:56.574 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@973 -- # kill 165147 00:13:56.574 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@978 -- # wait 165147 00:13:56.834 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:13:56.834 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:13:56.834 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:13:56.834 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@297 -- # iptr 00:13:56.834 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@791 -- # iptables-save 00:13:56.834 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@791 -- # iptables-restore 00:13:56.834 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:13:56.834 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:13:56.834 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@302 -- # remove_spdk_ns 00:13:56.834 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:56.835 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:13:56.835 17:29:38 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:58.746 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:13:58.746 00:13:58.746 real 0m24.275s 00:13:58.746 user 1m25.515s 00:13:58.746 sys 0m6.509s 00:13:58.746 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1130 -- # xtrace_disable 00:13:58.746 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:13:58.746 ************************************ 00:13:58.746 END TEST nvmf_fio_target 00:13:58.746 ************************************ 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@35 -- # run_test nvmf_bdevio /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1111 -- # xtrace_disable 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:13:59.005 ************************************ 00:13:59.005 START TEST nvmf_bdevio 00:13:59.005 ************************************ 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp 00:13:59.005 * Looking for test storage... 00:13:59.005 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1711 -- # lcov --version 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@333 -- # local ver1 ver1_l 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@334 -- # local ver2 ver2_l 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@336 -- # IFS=.-: 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@336 -- # read -ra ver1 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@337 -- # IFS=.-: 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@337 -- # read -ra ver2 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@338 -- # local 'op=<' 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@340 -- # ver1_l=2 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@341 -- # ver2_l=1 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@344 -- # case "$op" in 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@345 -- # : 1 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@364 -- # (( v = 0 )) 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@365 -- # decimal 1 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@353 -- # local d=1 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@355 -- # echo 1 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@365 -- # ver1[v]=1 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@366 -- # decimal 2 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@353 -- # local d=2 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@355 -- # echo 2 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@366 -- # ver2[v]=2 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@368 -- # return 0 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:13:59.005 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:59.005 --rc genhtml_branch_coverage=1 00:13:59.005 --rc genhtml_function_coverage=1 00:13:59.005 --rc genhtml_legend=1 00:13:59.005 --rc geninfo_all_blocks=1 00:13:59.005 --rc geninfo_unexecuted_blocks=1 00:13:59.005 00:13:59.005 ' 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:13:59.005 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:59.005 --rc genhtml_branch_coverage=1 00:13:59.005 --rc genhtml_function_coverage=1 00:13:59.005 --rc genhtml_legend=1 00:13:59.005 --rc geninfo_all_blocks=1 00:13:59.005 --rc geninfo_unexecuted_blocks=1 00:13:59.005 00:13:59.005 ' 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:13:59.005 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:59.005 --rc genhtml_branch_coverage=1 00:13:59.005 --rc genhtml_function_coverage=1 00:13:59.005 --rc genhtml_legend=1 00:13:59.005 --rc geninfo_all_blocks=1 00:13:59.005 --rc geninfo_unexecuted_blocks=1 00:13:59.005 00:13:59.005 ' 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:13:59.005 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:59.005 --rc genhtml_branch_coverage=1 00:13:59.005 --rc genhtml_function_coverage=1 00:13:59.005 --rc genhtml_legend=1 00:13:59.005 --rc geninfo_all_blocks=1 00:13:59.005 --rc geninfo_unexecuted_blocks=1 00:13:59.005 00:13:59.005 ' 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@7 -- # uname -s 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@15 -- # shopt -s extglob 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:13:59.005 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- paths/export.sh@5 -- # export PATH 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@51 -- # : 0 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:13:59.006 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@55 -- # have_pci_nics=0 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@11 -- # MALLOC_BDEV_SIZE=64 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@14 -- # nvmftestinit 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@476 -- # prepare_net_devs 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@438 -- # local -g is_hw=no 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@440 -- # remove_spdk_ns 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@309 -- # xtrace_disable 00:13:59.006 17:29:40 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:14:01.542 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:14:01.542 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@315 -- # pci_devs=() 00:14:01.542 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@315 -- # local -a pci_devs 00:14:01.542 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@316 -- # pci_net_devs=() 00:14:01.542 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:14:01.542 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@317 -- # pci_drivers=() 00:14:01.542 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@317 -- # local -A pci_drivers 00:14:01.542 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@319 -- # net_devs=() 00:14:01.542 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@319 -- # local -ga net_devs 00:14:01.542 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@320 -- # e810=() 00:14:01.542 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@320 -- # local -ga e810 00:14:01.542 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@321 -- # x722=() 00:14:01.542 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@321 -- # local -ga x722 00:14:01.542 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@322 -- # mlx=() 00:14:01.542 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@322 -- # local -ga mlx 00:14:01.542 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:14:01.542 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:14:01.542 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:14:01.542 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:14:01.542 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:14:01.542 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:14:01.542 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:14:01.543 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:14:01.543 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@418 -- # [[ up == up ]] 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:14:01.543 Found net devices under 0000:0a:00.0: cvl_0_0 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@418 -- # [[ up == up ]] 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:14:01.543 Found net devices under 0000:0a:00.1: cvl_0_1 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@442 -- # is_hw=yes 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:14:01.543 17:29:42 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:14:01.543 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:14:01.543 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.233 ms 00:14:01.543 00:14:01.543 --- 10.0.0.2 ping statistics --- 00:14:01.543 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:14:01.543 rtt min/avg/max/mdev = 0.233/0.233/0.233/0.000 ms 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:14:01.543 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:14:01.543 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.096 ms 00:14:01.543 00:14:01.543 --- 10.0.0.1 ping statistics --- 00:14:01.543 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:14:01.543 rtt min/avg/max/mdev = 0.096/0.096/0.096/0.000 ms 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@450 -- # return 0 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@16 -- # nvmfappstart -m 0x78 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@726 -- # xtrace_disable 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@509 -- # nvmfpid=170030 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x78 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@510 -- # waitforlisten 170030 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@835 -- # '[' -z 170030 ']' 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:01.543 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:01.543 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:14:01.543 [2024-12-06 17:29:43.198494] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:14:01.544 [2024-12-06 17:29:43.198580] [ DPDK EAL parameters: nvmf -c 0x78 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:01.544 [2024-12-06 17:29:43.272680] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:14:01.544 [2024-12-06 17:29:43.317231] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:14:01.544 [2024-12-06 17:29:43.317288] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:14:01.544 [2024-12-06 17:29:43.317316] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:14:01.544 [2024-12-06 17:29:43.317327] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:14:01.544 [2024-12-06 17:29:43.317336] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:14:01.544 [2024-12-06 17:29:43.318887] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:14:01.544 [2024-12-06 17:29:43.318984] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 6 00:14:01.544 [2024-12-06 17:29:43.318917] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 5 00:14:01.544 [2024-12-06 17:29:43.318988] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@868 -- # return 0 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@732 -- # xtrace_disable 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:14:01.802 [2024-12-06 17:29:43.465477] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:14:01.802 Malloc0 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@20 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@21 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@22 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:14:01.802 [2024-12-06 17:29:43.524002] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/bdev/bdevio/bdevio --json /dev/fd/62 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@24 -- # gen_nvmf_target_json 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@560 -- # config=() 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@560 -- # local subsystem config 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:14:01.802 { 00:14:01.802 "params": { 00:14:01.802 "name": "Nvme$subsystem", 00:14:01.802 "trtype": "$TEST_TRANSPORT", 00:14:01.802 "traddr": "$NVMF_FIRST_TARGET_IP", 00:14:01.802 "adrfam": "ipv4", 00:14:01.802 "trsvcid": "$NVMF_PORT", 00:14:01.802 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:14:01.802 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:14:01.802 "hdgst": ${hdgst:-false}, 00:14:01.802 "ddgst": ${ddgst:-false} 00:14:01.802 }, 00:14:01.802 "method": "bdev_nvme_attach_controller" 00:14:01.802 } 00:14:01.802 EOF 00:14:01.802 )") 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@582 -- # cat 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@584 -- # jq . 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@585 -- # IFS=, 00:14:01.802 17:29:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:14:01.802 "params": { 00:14:01.803 "name": "Nvme1", 00:14:01.803 "trtype": "tcp", 00:14:01.803 "traddr": "10.0.0.2", 00:14:01.803 "adrfam": "ipv4", 00:14:01.803 "trsvcid": "4420", 00:14:01.803 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:14:01.803 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:14:01.803 "hdgst": false, 00:14:01.803 "ddgst": false 00:14:01.803 }, 00:14:01.803 "method": "bdev_nvme_attach_controller" 00:14:01.803 }' 00:14:01.803 [2024-12-06 17:29:43.571639] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:14:01.803 [2024-12-06 17:29:43.571744] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid170066 ] 00:14:02.061 [2024-12-06 17:29:43.643392] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:14:02.061 [2024-12-06 17:29:43.693629] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:14:02.061 [2024-12-06 17:29:43.693687] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:14:02.061 [2024-12-06 17:29:43.693692] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:02.320 I/O targets: 00:14:02.320 Nvme1n1: 131072 blocks of 512 bytes (64 MiB) 00:14:02.320 00:14:02.320 00:14:02.320 CUnit - A unit testing framework for C - Version 2.1-3 00:14:02.320 http://cunit.sourceforge.net/ 00:14:02.320 00:14:02.320 00:14:02.320 Suite: bdevio tests on: Nvme1n1 00:14:02.320 Test: blockdev write read block ...passed 00:14:02.320 Test: blockdev write zeroes read block ...passed 00:14:02.320 Test: blockdev write zeroes read no split ...passed 00:14:02.320 Test: blockdev write zeroes read split ...passed 00:14:02.320 Test: blockdev write zeroes read split partial ...passed 00:14:02.320 Test: blockdev reset ...[2024-12-06 17:29:44.027906] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 1] resetting controller 00:14:02.320 [2024-12-06 17:29:44.028026] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x135a700 (9): Bad file descriptor 00:14:02.320 [2024-12-06 17:29:44.085725] bdev_nvme.c:2286:bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller successful. 00:14:02.320 passed 00:14:02.320 Test: blockdev write read 8 blocks ...passed 00:14:02.320 Test: blockdev write read size > 128k ...passed 00:14:02.320 Test: blockdev write read invalid size ...passed 00:14:02.320 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:14:02.320 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:14:02.320 Test: blockdev write read max offset ...passed 00:14:02.579 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:14:02.579 Test: blockdev writev readv 8 blocks ...passed 00:14:02.579 Test: blockdev writev readv 30 x 1block ...passed 00:14:02.579 Test: blockdev writev readv block ...passed 00:14:02.579 Test: blockdev writev readv size > 128k ...passed 00:14:02.579 Test: blockdev writev readv size > 128k in two iovs ...passed 00:14:02.579 Test: blockdev comparev and writev ...[2024-12-06 17:29:44.300888] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:14:02.579 [2024-12-06 17:29:44.300922] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:14:02.579 [2024-12-06 17:29:44.300951] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:14:02.579 [2024-12-06 17:29:44.300967] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:14:02.579 [2024-12-06 17:29:44.301281] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:14:02.579 [2024-12-06 17:29:44.301306] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:14:02.579 [2024-12-06 17:29:44.301337] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:14:02.579 [2024-12-06 17:29:44.301354] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:14:02.579 [2024-12-06 17:29:44.301658] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:14:02.579 [2024-12-06 17:29:44.301691] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:14:02.579 [2024-12-06 17:29:44.301713] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:14:02.579 [2024-12-06 17:29:44.301729] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:14:02.579 [2024-12-06 17:29:44.302029] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:14:02.579 [2024-12-06 17:29:44.302053] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:14:02.579 [2024-12-06 17:29:44.302074] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:14:02.579 [2024-12-06 17:29:44.302089] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:14:02.579 passed 00:14:02.579 Test: blockdev nvme passthru rw ...passed 00:14:02.579 Test: blockdev nvme passthru vendor specific ...[2024-12-06 17:29:44.383907] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:14:02.579 [2024-12-06 17:29:44.383933] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:14:02.579 [2024-12-06 17:29:44.384079] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:14:02.579 [2024-12-06 17:29:44.384103] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:14:02.579 [2024-12-06 17:29:44.384239] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:14:02.579 [2024-12-06 17:29:44.384262] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:14:02.579 [2024-12-06 17:29:44.384401] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:14:02.579 [2024-12-06 17:29:44.384423] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:14:02.579 passed 00:14:02.579 Test: blockdev nvme admin passthru ...passed 00:14:02.839 Test: blockdev copy ...passed 00:14:02.839 00:14:02.839 Run Summary: Type Total Ran Passed Failed Inactive 00:14:02.839 suites 1 1 n/a 0 0 00:14:02.839 tests 23 23 23 0 0 00:14:02.839 asserts 152 152 152 0 n/a 00:14:02.839 00:14:02.839 Elapsed time = 1.048 seconds 00:14:02.839 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@26 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:14:02.839 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:02.839 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:14:02.839 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:02.839 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@28 -- # trap - SIGINT SIGTERM EXIT 00:14:02.839 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@30 -- # nvmftestfini 00:14:02.839 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@516 -- # nvmfcleanup 00:14:02.839 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@121 -- # sync 00:14:02.839 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:14:02.839 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@124 -- # set +e 00:14:02.839 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@125 -- # for i in {1..20} 00:14:02.839 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:14:02.839 rmmod nvme_tcp 00:14:02.839 rmmod nvme_fabrics 00:14:02.839 rmmod nvme_keyring 00:14:03.097 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:14:03.097 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@128 -- # set -e 00:14:03.097 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@129 -- # return 0 00:14:03.097 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@517 -- # '[' -n 170030 ']' 00:14:03.097 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@518 -- # killprocess 170030 00:14:03.097 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@954 -- # '[' -z 170030 ']' 00:14:03.097 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@958 -- # kill -0 170030 00:14:03.097 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@959 -- # uname 00:14:03.097 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:03.097 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 170030 00:14:03.097 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@960 -- # process_name=reactor_3 00:14:03.097 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@964 -- # '[' reactor_3 = sudo ']' 00:14:03.097 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@972 -- # echo 'killing process with pid 170030' 00:14:03.097 killing process with pid 170030 00:14:03.097 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@973 -- # kill 170030 00:14:03.097 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@978 -- # wait 170030 00:14:03.357 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:14:03.357 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:14:03.357 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:14:03.357 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@297 -- # iptr 00:14:03.357 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@791 -- # iptables-save 00:14:03.357 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:14:03.357 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@791 -- # iptables-restore 00:14:03.357 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:14:03.357 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@302 -- # remove_spdk_ns 00:14:03.357 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:14:03.357 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:14:03.357 17:29:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:14:05.264 17:29:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:14:05.264 00:14:05.264 real 0m6.365s 00:14:05.264 user 0m9.413s 00:14:05.264 sys 0m2.240s 00:14:05.264 17:29:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:05.264 17:29:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:14:05.264 ************************************ 00:14:05.264 END TEST nvmf_bdevio 00:14:05.264 ************************************ 00:14:05.264 17:29:47 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:14:05.264 00:14:05.264 real 3m55.082s 00:14:05.264 user 10m14.945s 00:14:05.264 sys 1m5.808s 00:14:05.264 17:29:47 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:05.264 17:29:47 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:14:05.264 ************************************ 00:14:05.264 END TEST nvmf_target_core 00:14:05.264 ************************************ 00:14:05.264 17:29:47 nvmf_tcp -- nvmf/nvmf.sh@15 -- # run_test nvmf_target_extra /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_target_extra.sh --transport=tcp 00:14:05.264 17:29:47 nvmf_tcp -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:14:05.264 17:29:47 nvmf_tcp -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:05.264 17:29:47 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:14:05.264 ************************************ 00:14:05.264 START TEST nvmf_target_extra 00:14:05.264 ************************************ 00:14:05.264 17:29:47 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_target_extra.sh --transport=tcp 00:14:05.523 * Looking for test storage... 00:14:05.523 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1711 -- # lcov --version 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@333 -- # local ver1 ver1_l 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@334 -- # local ver2 ver2_l 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@336 -- # IFS=.-: 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@336 -- # read -ra ver1 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@337 -- # IFS=.-: 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@337 -- # read -ra ver2 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@338 -- # local 'op=<' 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@340 -- # ver1_l=2 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@341 -- # ver2_l=1 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@344 -- # case "$op" in 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@345 -- # : 1 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@364 -- # (( v = 0 )) 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@365 -- # decimal 1 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@353 -- # local d=1 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@355 -- # echo 1 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@365 -- # ver1[v]=1 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@366 -- # decimal 2 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@353 -- # local d=2 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@355 -- # echo 2 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@366 -- # ver2[v]=2 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@368 -- # return 0 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:14:05.523 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:05.523 --rc genhtml_branch_coverage=1 00:14:05.523 --rc genhtml_function_coverage=1 00:14:05.523 --rc genhtml_legend=1 00:14:05.523 --rc geninfo_all_blocks=1 00:14:05.523 --rc geninfo_unexecuted_blocks=1 00:14:05.523 00:14:05.523 ' 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:14:05.523 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:05.523 --rc genhtml_branch_coverage=1 00:14:05.523 --rc genhtml_function_coverage=1 00:14:05.523 --rc genhtml_legend=1 00:14:05.523 --rc geninfo_all_blocks=1 00:14:05.523 --rc geninfo_unexecuted_blocks=1 00:14:05.523 00:14:05.523 ' 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:14:05.523 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:05.523 --rc genhtml_branch_coverage=1 00:14:05.523 --rc genhtml_function_coverage=1 00:14:05.523 --rc genhtml_legend=1 00:14:05.523 --rc geninfo_all_blocks=1 00:14:05.523 --rc geninfo_unexecuted_blocks=1 00:14:05.523 00:14:05.523 ' 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:14:05.523 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:05.523 --rc genhtml_branch_coverage=1 00:14:05.523 --rc genhtml_function_coverage=1 00:14:05.523 --rc genhtml_legend=1 00:14:05.523 --rc geninfo_all_blocks=1 00:14:05.523 --rc geninfo_unexecuted_blocks=1 00:14:05.523 00:14:05.523 ' 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@7 -- # uname -s 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@15 -- # shopt -s extglob 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- paths/export.sh@5 -- # export PATH 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@51 -- # : 0 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:14:05.523 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:14:05.524 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:14:05.524 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:14:05.524 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:14:05.524 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:14:05.524 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:14:05.524 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:14:05.524 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@55 -- # have_pci_nics=0 00:14:05.524 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@11 -- # trap 'exit 1' SIGINT SIGTERM EXIT 00:14:05.524 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@13 -- # TEST_ARGS=("$@") 00:14:05.524 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@15 -- # [[ 0 -eq 0 ]] 00:14:05.524 17:29:47 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@16 -- # run_test nvmf_example /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_example.sh --transport=tcp 00:14:05.524 17:29:47 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:14:05.524 17:29:47 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:05.524 17:29:47 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:14:05.524 ************************************ 00:14:05.524 START TEST nvmf_example 00:14:05.524 ************************************ 00:14:05.524 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_example.sh --transport=tcp 00:14:05.524 * Looking for test storage... 00:14:05.524 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:14:05.524 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:14:05.524 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1711 -- # lcov --version 00:14:05.524 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@333 -- # local ver1 ver1_l 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@334 -- # local ver2 ver2_l 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@336 -- # IFS=.-: 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@336 -- # read -ra ver1 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@337 -- # IFS=.-: 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@337 -- # read -ra ver2 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@338 -- # local 'op=<' 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@340 -- # ver1_l=2 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@341 -- # ver2_l=1 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@344 -- # case "$op" in 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@345 -- # : 1 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@364 -- # (( v = 0 )) 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@365 -- # decimal 1 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@353 -- # local d=1 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@355 -- # echo 1 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@365 -- # ver1[v]=1 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@366 -- # decimal 2 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@353 -- # local d=2 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@355 -- # echo 2 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@366 -- # ver2[v]=2 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@368 -- # return 0 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:14:05.785 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:05.785 --rc genhtml_branch_coverage=1 00:14:05.785 --rc genhtml_function_coverage=1 00:14:05.785 --rc genhtml_legend=1 00:14:05.785 --rc geninfo_all_blocks=1 00:14:05.785 --rc geninfo_unexecuted_blocks=1 00:14:05.785 00:14:05.785 ' 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:14:05.785 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:05.785 --rc genhtml_branch_coverage=1 00:14:05.785 --rc genhtml_function_coverage=1 00:14:05.785 --rc genhtml_legend=1 00:14:05.785 --rc geninfo_all_blocks=1 00:14:05.785 --rc geninfo_unexecuted_blocks=1 00:14:05.785 00:14:05.785 ' 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:14:05.785 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:05.785 --rc genhtml_branch_coverage=1 00:14:05.785 --rc genhtml_function_coverage=1 00:14:05.785 --rc genhtml_legend=1 00:14:05.785 --rc geninfo_all_blocks=1 00:14:05.785 --rc geninfo_unexecuted_blocks=1 00:14:05.785 00:14:05.785 ' 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:14:05.785 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:05.785 --rc genhtml_branch_coverage=1 00:14:05.785 --rc genhtml_function_coverage=1 00:14:05.785 --rc genhtml_legend=1 00:14:05.785 --rc geninfo_all_blocks=1 00:14:05.785 --rc geninfo_unexecuted_blocks=1 00:14:05.785 00:14:05.785 ' 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@7 -- # uname -s 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@15 -- # shopt -s extglob 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:05.785 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- paths/export.sh@5 -- # export PATH 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@51 -- # : 0 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:14:05.786 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@55 -- # have_pci_nics=0 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@11 -- # NVMF_EXAMPLE=("$SPDK_EXAMPLE_DIR/nvmf") 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@13 -- # MALLOC_BDEV_SIZE=64 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@14 -- # MALLOC_BLOCK_SIZE=512 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@24 -- # build_nvmf_example_args 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@17 -- # '[' 0 -eq 1 ']' 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@20 -- # NVMF_EXAMPLE+=(-i "$NVMF_APP_SHM_ID" -g 10000) 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@21 -- # NVMF_EXAMPLE+=("${NO_HUGE[@]}") 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@40 -- # timing_enter nvmf_example_test 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@726 -- # xtrace_disable 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@41 -- # nvmftestinit 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@476 -- # prepare_net_devs 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@438 -- # local -g is_hw=no 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@440 -- # remove_spdk_ns 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@309 -- # xtrace_disable 00:14:05.786 17:29:47 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@315 -- # pci_devs=() 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@315 -- # local -a pci_devs 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@316 -- # pci_net_devs=() 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@317 -- # pci_drivers=() 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@317 -- # local -A pci_drivers 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@319 -- # net_devs=() 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@319 -- # local -ga net_devs 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@320 -- # e810=() 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@320 -- # local -ga e810 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@321 -- # x722=() 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@321 -- # local -ga x722 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@322 -- # mlx=() 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@322 -- # local -ga mlx 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:14:08.322 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:14:08.322 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@418 -- # [[ up == up ]] 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:14:08.322 Found net devices under 0000:0a:00.0: cvl_0_0 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@418 -- # [[ up == up ]] 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:14:08.322 Found net devices under 0000:0a:00.1: cvl_0_1 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@442 -- # is_hw=yes 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:14:08.322 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:14:08.323 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:14:08.323 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.257 ms 00:14:08.323 00:14:08.323 --- 10.0.0.2 ping statistics --- 00:14:08.323 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:14:08.323 rtt min/avg/max/mdev = 0.257/0.257/0.257/0.000 ms 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:14:08.323 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:14:08.323 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.121 ms 00:14:08.323 00:14:08.323 --- 10.0.0.1 ping statistics --- 00:14:08.323 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:14:08.323 rtt min/avg/max/mdev = 0.121/0.121/0.121/0.000 ms 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@450 -- # return 0 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@42 -- # nvmfexamplestart '-m 0xF' 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@27 -- # timing_enter start_nvmf_example 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@726 -- # xtrace_disable 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@29 -- # '[' tcp == tcp ']' 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@30 -- # NVMF_EXAMPLE=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_EXAMPLE[@]}") 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@34 -- # nvmfpid=172318 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@35 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@36 -- # waitforlisten 172318 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@835 -- # '[' -z 172318 ']' 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@33 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/nvmf -i 0 -g 10000 -m 0xF 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:08.323 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:08.323 17:29:49 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:14:09.257 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:09.257 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@868 -- # return 0 00:14:09.258 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@37 -- # timing_exit start_nvmf_example 00:14:09.258 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@732 -- # xtrace_disable 00:14:09.258 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:14:09.258 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@45 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:14:09.258 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:09.258 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:14:09.258 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:09.258 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@47 -- # rpc_cmd bdev_malloc_create 64 512 00:14:09.258 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:09.258 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:14:09.258 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:09.258 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@47 -- # malloc_bdevs='Malloc0 ' 00:14:09.258 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@49 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:14:09.258 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:09.258 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:14:09.258 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:09.258 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@52 -- # for malloc_bdev in $malloc_bdevs 00:14:09.258 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@53 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:14:09.258 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:09.258 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:14:09.258 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:09.258 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@57 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:14:09.258 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:09.258 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:14:09.258 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:09.258 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@59 -- # perf=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf 00:14:09.258 17:29:50 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 64 -o 4096 -w randrw -M 30 -t 10 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:14:19.230 Initializing NVMe Controllers 00:14:19.230 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:14:19.230 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:14:19.230 Initialization complete. Launching workers. 00:14:19.230 ======================================================== 00:14:19.230 Latency(us) 00:14:19.230 Device Information : IOPS MiB/s Average min max 00:14:19.230 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 14869.51 58.08 4306.72 849.60 15225.99 00:14:19.230 ======================================================== 00:14:19.230 Total : 14869.51 58.08 4306.72 849.60 15225.99 00:14:19.230 00:14:19.230 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@65 -- # trap - SIGINT SIGTERM EXIT 00:14:19.230 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@66 -- # nvmftestfini 00:14:19.230 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@516 -- # nvmfcleanup 00:14:19.230 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@121 -- # sync 00:14:19.230 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:14:19.230 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@124 -- # set +e 00:14:19.230 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@125 -- # for i in {1..20} 00:14:19.230 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:14:19.230 rmmod nvme_tcp 00:14:19.230 rmmod nvme_fabrics 00:14:19.487 rmmod nvme_keyring 00:14:19.487 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:14:19.487 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@128 -- # set -e 00:14:19.487 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@129 -- # return 0 00:14:19.487 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@517 -- # '[' -n 172318 ']' 00:14:19.487 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@518 -- # killprocess 172318 00:14:19.487 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@954 -- # '[' -z 172318 ']' 00:14:19.487 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@958 -- # kill -0 172318 00:14:19.487 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@959 -- # uname 00:14:19.487 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:19.487 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 172318 00:14:19.487 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@960 -- # process_name=nvmf 00:14:19.487 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@964 -- # '[' nvmf = sudo ']' 00:14:19.487 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@972 -- # echo 'killing process with pid 172318' 00:14:19.487 killing process with pid 172318 00:14:19.487 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@973 -- # kill 172318 00:14:19.487 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@978 -- # wait 172318 00:14:19.744 nvmf threads initialize successfully 00:14:19.744 bdev subsystem init successfully 00:14:19.744 created a nvmf target service 00:14:19.744 create targets's poll groups done 00:14:19.744 all subsystems of target started 00:14:19.744 nvmf target is running 00:14:19.744 all subsystems of target stopped 00:14:19.744 destroy targets's poll groups done 00:14:19.744 destroyed the nvmf target service 00:14:19.744 bdev subsystem finish successfully 00:14:19.744 nvmf threads destroy successfully 00:14:19.744 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:14:19.744 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:14:19.744 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:14:19.744 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@297 -- # iptr 00:14:19.744 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@791 -- # iptables-save 00:14:19.744 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:14:19.744 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@791 -- # iptables-restore 00:14:19.744 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:14:19.744 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@302 -- # remove_spdk_ns 00:14:19.744 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:14:19.744 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:14:19.744 17:30:01 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:14:21.652 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:14:21.652 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@67 -- # timing_exit nvmf_example_test 00:14:21.652 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@732 -- # xtrace_disable 00:14:21.652 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:14:21.652 00:14:21.652 real 0m16.167s 00:14:21.652 user 0m45.419s 00:14:21.652 sys 0m3.365s 00:14:21.652 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:21.652 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:14:21.652 ************************************ 00:14:21.652 END TEST nvmf_example 00:14:21.652 ************************************ 00:14:21.652 17:30:03 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@17 -- # run_test nvmf_filesystem /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/filesystem.sh --transport=tcp 00:14:21.652 17:30:03 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:14:21.652 17:30:03 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:21.652 17:30:03 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:14:21.652 ************************************ 00:14:21.652 START TEST nvmf_filesystem 00:14:21.652 ************************************ 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/filesystem.sh --transport=tcp 00:14:21.915 * Looking for test storage... 00:14:21.915 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1711 -- # lcov --version 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@333 -- # local ver1 ver1_l 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@334 -- # local ver2 ver2_l 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@336 -- # IFS=.-: 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@336 -- # read -ra ver1 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@337 -- # IFS=.-: 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@337 -- # read -ra ver2 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@338 -- # local 'op=<' 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@340 -- # ver1_l=2 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@341 -- # ver2_l=1 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@344 -- # case "$op" in 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@345 -- # : 1 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@364 -- # (( v = 0 )) 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@365 -- # decimal 1 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@353 -- # local d=1 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@355 -- # echo 1 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@365 -- # ver1[v]=1 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@366 -- # decimal 2 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@353 -- # local d=2 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@355 -- # echo 2 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@366 -- # ver2[v]=2 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@368 -- # return 0 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:14:21.915 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:14:21.915 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:21.915 --rc genhtml_branch_coverage=1 00:14:21.915 --rc genhtml_function_coverage=1 00:14:21.915 --rc genhtml_legend=1 00:14:21.915 --rc geninfo_all_blocks=1 00:14:21.915 --rc geninfo_unexecuted_blocks=1 00:14:21.915 00:14:21.915 ' 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:14:21.916 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:21.916 --rc genhtml_branch_coverage=1 00:14:21.916 --rc genhtml_function_coverage=1 00:14:21.916 --rc genhtml_legend=1 00:14:21.916 --rc geninfo_all_blocks=1 00:14:21.916 --rc geninfo_unexecuted_blocks=1 00:14:21.916 00:14:21.916 ' 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:14:21.916 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:21.916 --rc genhtml_branch_coverage=1 00:14:21.916 --rc genhtml_function_coverage=1 00:14:21.916 --rc genhtml_legend=1 00:14:21.916 --rc geninfo_all_blocks=1 00:14:21.916 --rc geninfo_unexecuted_blocks=1 00:14:21.916 00:14:21.916 ' 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:14:21.916 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:21.916 --rc genhtml_branch_coverage=1 00:14:21.916 --rc genhtml_function_coverage=1 00:14:21.916 --rc genhtml_legend=1 00:14:21.916 --rc geninfo_all_blocks=1 00:14:21.916 --rc geninfo_unexecuted_blocks=1 00:14:21.916 00:14:21.916 ' 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@7 -- # rpc_py=rpc_cmd 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@34 -- # set -e 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@35 -- # shopt -s nullglob 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@36 -- # shopt -s extglob 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@37 -- # shopt -s inherit_errexit 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@39 -- # '[' -z /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output ']' 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@44 -- # [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/build_config.sh ]] 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/build_config.sh 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@1 -- # CONFIG_WPDK_DIR= 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@2 -- # CONFIG_ASAN=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@3 -- # CONFIG_VBDEV_COMPRESS=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@4 -- # CONFIG_HAVE_EXECINFO_H=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@5 -- # CONFIG_USDT=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@6 -- # CONFIG_CUSTOMOCF=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@7 -- # CONFIG_PREFIX=/usr/local 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@8 -- # CONFIG_RBD=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@9 -- # CONFIG_LIBDIR= 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@10 -- # CONFIG_IDXD=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@11 -- # CONFIG_NVME_CUSE=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@12 -- # CONFIG_SMA=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@13 -- # CONFIG_VTUNE=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@14 -- # CONFIG_TSAN=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@15 -- # CONFIG_RDMA_SEND_WITH_INVAL=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@16 -- # CONFIG_VFIO_USER_DIR= 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@17 -- # CONFIG_MAX_NUMA_NODES=1 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@18 -- # CONFIG_PGO_CAPTURE=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@19 -- # CONFIG_HAVE_UUID_GENERATE_SHA1=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@20 -- # CONFIG_ENV=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@21 -- # CONFIG_LTO=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@22 -- # CONFIG_ISCSI_INITIATOR=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@23 -- # CONFIG_CET=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@24 -- # CONFIG_VBDEV_COMPRESS_MLX5=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@25 -- # CONFIG_OCF_PATH= 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@26 -- # CONFIG_RDMA_SET_TOS=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@27 -- # CONFIG_AIO_FSDEV=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@28 -- # CONFIG_HAVE_ARC4RANDOM=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@29 -- # CONFIG_HAVE_LIBARCHIVE=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@30 -- # CONFIG_UBLK=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@31 -- # CONFIG_ISAL_CRYPTO=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@32 -- # CONFIG_OPENSSL_PATH= 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@33 -- # CONFIG_OCF=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@34 -- # CONFIG_FUSE=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@35 -- # CONFIG_VTUNE_DIR= 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@36 -- # CONFIG_FUZZER_LIB= 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@37 -- # CONFIG_FUZZER=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@38 -- # CONFIG_FSDEV=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@39 -- # CONFIG_DPDK_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@40 -- # CONFIG_CRYPTO=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@41 -- # CONFIG_PGO_USE=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@42 -- # CONFIG_VHOST=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@43 -- # CONFIG_DAOS=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@44 -- # CONFIG_DPDK_INC_DIR=//var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@45 -- # CONFIG_DAOS_DIR= 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@46 -- # CONFIG_UNIT_TESTS=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@47 -- # CONFIG_RDMA_SET_ACK_TIMEOUT=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@48 -- # CONFIG_VIRTIO=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@49 -- # CONFIG_DPDK_UADK=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@50 -- # CONFIG_COVERAGE=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@51 -- # CONFIG_RDMA=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@52 -- # CONFIG_HAVE_STRUCT_STAT_ST_ATIM=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@53 -- # CONFIG_HAVE_LZ4=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@54 -- # CONFIG_FIO_SOURCE_DIR=/usr/src/fio 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@55 -- # CONFIG_URING_PATH= 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@56 -- # CONFIG_XNVME=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@57 -- # CONFIG_VFIO_USER=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@58 -- # CONFIG_ARCH=native 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@59 -- # CONFIG_HAVE_EVP_MAC=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@60 -- # CONFIG_URING_ZNS=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@61 -- # CONFIG_WERROR=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@62 -- # CONFIG_HAVE_LIBBSD=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@63 -- # CONFIG_UBSAN=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@64 -- # CONFIG_HAVE_STRUCT_STAT_ST_ATIMESPEC=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@65 -- # CONFIG_IPSEC_MB_DIR= 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@66 -- # CONFIG_GOLANG=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@67 -- # CONFIG_ISAL=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@68 -- # CONFIG_IDXD_KERNEL=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@69 -- # CONFIG_DPDK_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@70 -- # CONFIG_RDMA_PROV=verbs 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@71 -- # CONFIG_APPS=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@72 -- # CONFIG_SHARED=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@73 -- # CONFIG_HAVE_KEYUTILS=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@74 -- # CONFIG_FC_PATH= 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@75 -- # CONFIG_DPDK_PKG_CONFIG=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@76 -- # CONFIG_FC=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@77 -- # CONFIG_AVAHI=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@78 -- # CONFIG_FIO_PLUGIN=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@79 -- # CONFIG_RAID5F=n 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@80 -- # CONFIG_EXAMPLES=y 00:14:21.916 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@81 -- # CONFIG_TESTS=y 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@82 -- # CONFIG_CRYPTO_MLX5=n 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@83 -- # CONFIG_MAX_LCORES=128 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@84 -- # CONFIG_IPSEC_MB=n 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@85 -- # CONFIG_PGO_DIR= 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@86 -- # CONFIG_DEBUG=y 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@87 -- # CONFIG_DPDK_COMPRESSDEV=n 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@88 -- # CONFIG_CROSS_PREFIX= 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@89 -- # CONFIG_COPY_FILE_RANGE=y 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@90 -- # CONFIG_URING=n 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@54 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/applications.sh 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@8 -- # dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/applications.sh 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@8 -- # readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@8 -- # _root=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@9 -- # _root=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@10 -- # _app_dir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@11 -- # _test_app_dir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@12 -- # _examples_dir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@14 -- # VHOST_FUZZ_APP=("$_test_app_dir/fuzz/vhost_fuzz/vhost_fuzz") 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@15 -- # ISCSI_APP=("$_app_dir/iscsi_tgt") 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@16 -- # NVMF_APP=("$_app_dir/nvmf_tgt") 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@17 -- # VHOST_APP=("$_app_dir/vhost") 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@18 -- # DD_APP=("$_app_dir/spdk_dd") 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@19 -- # SPDK_APP=("$_app_dir/spdk_tgt") 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@22 -- # [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/config.h ]] 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@23 -- # [[ #ifndef SPDK_CONFIG_H 00:14:21.917 #define SPDK_CONFIG_H 00:14:21.917 #define SPDK_CONFIG_AIO_FSDEV 1 00:14:21.917 #define SPDK_CONFIG_APPS 1 00:14:21.917 #define SPDK_CONFIG_ARCH native 00:14:21.917 #undef SPDK_CONFIG_ASAN 00:14:21.917 #undef SPDK_CONFIG_AVAHI 00:14:21.917 #undef SPDK_CONFIG_CET 00:14:21.917 #define SPDK_CONFIG_COPY_FILE_RANGE 1 00:14:21.917 #define SPDK_CONFIG_COVERAGE 1 00:14:21.917 #define SPDK_CONFIG_CROSS_PREFIX 00:14:21.917 #undef SPDK_CONFIG_CRYPTO 00:14:21.917 #undef SPDK_CONFIG_CRYPTO_MLX5 00:14:21.917 #undef SPDK_CONFIG_CUSTOMOCF 00:14:21.917 #undef SPDK_CONFIG_DAOS 00:14:21.917 #define SPDK_CONFIG_DAOS_DIR 00:14:21.917 #define SPDK_CONFIG_DEBUG 1 00:14:21.917 #undef SPDK_CONFIG_DPDK_COMPRESSDEV 00:14:21.917 #define SPDK_CONFIG_DPDK_DIR /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build 00:14:21.917 #define SPDK_CONFIG_DPDK_INC_DIR //var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/include 00:14:21.917 #define SPDK_CONFIG_DPDK_LIB_DIR /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:14:21.917 #undef SPDK_CONFIG_DPDK_PKG_CONFIG 00:14:21.917 #undef SPDK_CONFIG_DPDK_UADK 00:14:21.917 #define SPDK_CONFIG_ENV /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk 00:14:21.917 #define SPDK_CONFIG_EXAMPLES 1 00:14:21.917 #undef SPDK_CONFIG_FC 00:14:21.917 #define SPDK_CONFIG_FC_PATH 00:14:21.917 #define SPDK_CONFIG_FIO_PLUGIN 1 00:14:21.917 #define SPDK_CONFIG_FIO_SOURCE_DIR /usr/src/fio 00:14:21.917 #define SPDK_CONFIG_FSDEV 1 00:14:21.917 #undef SPDK_CONFIG_FUSE 00:14:21.917 #undef SPDK_CONFIG_FUZZER 00:14:21.917 #define SPDK_CONFIG_FUZZER_LIB 00:14:21.917 #undef SPDK_CONFIG_GOLANG 00:14:21.917 #define SPDK_CONFIG_HAVE_ARC4RANDOM 1 00:14:21.917 #define SPDK_CONFIG_HAVE_EVP_MAC 1 00:14:21.917 #define SPDK_CONFIG_HAVE_EXECINFO_H 1 00:14:21.917 #define SPDK_CONFIG_HAVE_KEYUTILS 1 00:14:21.917 #undef SPDK_CONFIG_HAVE_LIBARCHIVE 00:14:21.917 #undef SPDK_CONFIG_HAVE_LIBBSD 00:14:21.917 #undef SPDK_CONFIG_HAVE_LZ4 00:14:21.917 #define SPDK_CONFIG_HAVE_STRUCT_STAT_ST_ATIM 1 00:14:21.917 #undef SPDK_CONFIG_HAVE_STRUCT_STAT_ST_ATIMESPEC 00:14:21.917 #define SPDK_CONFIG_HAVE_UUID_GENERATE_SHA1 1 00:14:21.917 #define SPDK_CONFIG_IDXD 1 00:14:21.917 #define SPDK_CONFIG_IDXD_KERNEL 1 00:14:21.917 #undef SPDK_CONFIG_IPSEC_MB 00:14:21.917 #define SPDK_CONFIG_IPSEC_MB_DIR 00:14:21.917 #define SPDK_CONFIG_ISAL 1 00:14:21.917 #define SPDK_CONFIG_ISAL_CRYPTO 1 00:14:21.917 #define SPDK_CONFIG_ISCSI_INITIATOR 1 00:14:21.917 #define SPDK_CONFIG_LIBDIR 00:14:21.917 #undef SPDK_CONFIG_LTO 00:14:21.917 #define SPDK_CONFIG_MAX_LCORES 128 00:14:21.917 #define SPDK_CONFIG_MAX_NUMA_NODES 1 00:14:21.917 #define SPDK_CONFIG_NVME_CUSE 1 00:14:21.917 #undef SPDK_CONFIG_OCF 00:14:21.917 #define SPDK_CONFIG_OCF_PATH 00:14:21.917 #define SPDK_CONFIG_OPENSSL_PATH 00:14:21.917 #undef SPDK_CONFIG_PGO_CAPTURE 00:14:21.917 #define SPDK_CONFIG_PGO_DIR 00:14:21.917 #undef SPDK_CONFIG_PGO_USE 00:14:21.917 #define SPDK_CONFIG_PREFIX /usr/local 00:14:21.917 #undef SPDK_CONFIG_RAID5F 00:14:21.917 #undef SPDK_CONFIG_RBD 00:14:21.917 #define SPDK_CONFIG_RDMA 1 00:14:21.917 #define SPDK_CONFIG_RDMA_PROV verbs 00:14:21.917 #define SPDK_CONFIG_RDMA_SEND_WITH_INVAL 1 00:14:21.917 #define SPDK_CONFIG_RDMA_SET_ACK_TIMEOUT 1 00:14:21.917 #define SPDK_CONFIG_RDMA_SET_TOS 1 00:14:21.917 #define SPDK_CONFIG_SHARED 1 00:14:21.917 #undef SPDK_CONFIG_SMA 00:14:21.917 #define SPDK_CONFIG_TESTS 1 00:14:21.917 #undef SPDK_CONFIG_TSAN 00:14:21.917 #define SPDK_CONFIG_UBLK 1 00:14:21.917 #define SPDK_CONFIG_UBSAN 1 00:14:21.917 #undef SPDK_CONFIG_UNIT_TESTS 00:14:21.917 #undef SPDK_CONFIG_URING 00:14:21.917 #define SPDK_CONFIG_URING_PATH 00:14:21.917 #undef SPDK_CONFIG_URING_ZNS 00:14:21.917 #undef SPDK_CONFIG_USDT 00:14:21.917 #undef SPDK_CONFIG_VBDEV_COMPRESS 00:14:21.917 #undef SPDK_CONFIG_VBDEV_COMPRESS_MLX5 00:14:21.917 #define SPDK_CONFIG_VFIO_USER 1 00:14:21.917 #define SPDK_CONFIG_VFIO_USER_DIR 00:14:21.917 #define SPDK_CONFIG_VHOST 1 00:14:21.917 #define SPDK_CONFIG_VIRTIO 1 00:14:21.917 #undef SPDK_CONFIG_VTUNE 00:14:21.917 #define SPDK_CONFIG_VTUNE_DIR 00:14:21.917 #define SPDK_CONFIG_WERROR 1 00:14:21.917 #define SPDK_CONFIG_WPDK_DIR 00:14:21.917 #undef SPDK_CONFIG_XNVME 00:14:21.917 #endif /* SPDK_CONFIG_H */ == *\#\d\e\f\i\n\e\ \S\P\D\K\_\C\O\N\F\I\G\_\D\E\B\U\G* ]] 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@24 -- # (( SPDK_AUTOTEST_DEBUG_APPS )) 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@55 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@15 -- # shopt -s extglob 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@5 -- # export PATH 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@56 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/common 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@6 -- # dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/common 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@6 -- # readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@6 -- # _pmdir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm 00:14:21.917 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@7 -- # readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/../../../ 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@7 -- # _pmrootdir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@64 -- # TEST_TAG=N/A 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@65 -- # TEST_TAG_FILE=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/.run_test_name 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@67 -- # PM_OUTPUTDIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@68 -- # uname -s 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@68 -- # PM_OS=Linux 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@70 -- # MONITOR_RESOURCES_SUDO=() 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@70 -- # declare -A MONITOR_RESOURCES_SUDO 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@71 -- # MONITOR_RESOURCES_SUDO["collect-bmc-pm"]=1 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@72 -- # MONITOR_RESOURCES_SUDO["collect-cpu-load"]=0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@73 -- # MONITOR_RESOURCES_SUDO["collect-cpu-temp"]=0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@74 -- # MONITOR_RESOURCES_SUDO["collect-vmstat"]=0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@76 -- # SUDO[0]= 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@76 -- # SUDO[1]='sudo -E' 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@78 -- # MONITOR_RESOURCES=(collect-cpu-load collect-vmstat) 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@79 -- # [[ Linux == FreeBSD ]] 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@81 -- # [[ Linux == Linux ]] 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@81 -- # [[ ............................... != QEMU ]] 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@81 -- # [[ ! -e /.dockerenv ]] 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@84 -- # MONITOR_RESOURCES+=(collect-cpu-temp) 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@85 -- # MONITOR_RESOURCES+=(collect-bmc-pm) 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@88 -- # [[ ! -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power ]] 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@58 -- # : 1 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@59 -- # export RUN_NIGHTLY 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@62 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@63 -- # export SPDK_AUTOTEST_DEBUG_APPS 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@64 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@65 -- # export SPDK_RUN_VALGRIND 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@66 -- # : 1 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@67 -- # export SPDK_RUN_FUNCTIONAL_TEST 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@68 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@69 -- # export SPDK_TEST_UNITTEST 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@70 -- # : 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@71 -- # export SPDK_TEST_AUTOBUILD 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@72 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@73 -- # export SPDK_TEST_RELEASE_BUILD 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@74 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@75 -- # export SPDK_TEST_ISAL 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@76 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@77 -- # export SPDK_TEST_ISCSI 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@78 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@79 -- # export SPDK_TEST_ISCSI_INITIATOR 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@80 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@81 -- # export SPDK_TEST_NVME 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@82 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@83 -- # export SPDK_TEST_NVME_PMR 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@84 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@85 -- # export SPDK_TEST_NVME_BP 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@86 -- # : 1 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@87 -- # export SPDK_TEST_NVME_CLI 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@88 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@89 -- # export SPDK_TEST_NVME_CUSE 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@90 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@91 -- # export SPDK_TEST_NVME_FDP 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@92 -- # : 1 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@93 -- # export SPDK_TEST_NVMF 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@94 -- # : 1 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@95 -- # export SPDK_TEST_VFIOUSER 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@96 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@97 -- # export SPDK_TEST_VFIOUSER_QEMU 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@98 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@99 -- # export SPDK_TEST_FUZZER 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@100 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@101 -- # export SPDK_TEST_FUZZER_SHORT 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@102 -- # : tcp 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@103 -- # export SPDK_TEST_NVMF_TRANSPORT 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@104 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@105 -- # export SPDK_TEST_RBD 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@106 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@107 -- # export SPDK_TEST_VHOST 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@108 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@109 -- # export SPDK_TEST_BLOCKDEV 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@110 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@111 -- # export SPDK_TEST_RAID 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@112 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@113 -- # export SPDK_TEST_IOAT 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@114 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@115 -- # export SPDK_TEST_BLOBFS 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@116 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@117 -- # export SPDK_TEST_VHOST_INIT 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@118 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@119 -- # export SPDK_TEST_LVOL 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@120 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@121 -- # export SPDK_TEST_VBDEV_COMPRESS 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@122 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@123 -- # export SPDK_RUN_ASAN 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@124 -- # : 1 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@125 -- # export SPDK_RUN_UBSAN 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@126 -- # : /var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@127 -- # export SPDK_RUN_EXTERNAL_DPDK 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@128 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@129 -- # export SPDK_RUN_NON_ROOT 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@130 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@131 -- # export SPDK_TEST_CRYPTO 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@132 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@133 -- # export SPDK_TEST_FTL 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@134 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@135 -- # export SPDK_TEST_OCF 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@136 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@137 -- # export SPDK_TEST_VMD 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@138 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@139 -- # export SPDK_TEST_OPAL 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@140 -- # : v23.11 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@141 -- # export SPDK_TEST_NATIVE_DPDK 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@142 -- # : true 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@143 -- # export SPDK_AUTOTEST_X 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@144 -- # : 0 00:14:21.918 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@145 -- # export SPDK_TEST_URING 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@146 -- # : 0 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@147 -- # export SPDK_TEST_USDT 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@148 -- # : 0 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@149 -- # export SPDK_TEST_USE_IGB_UIO 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@150 -- # : 0 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@151 -- # export SPDK_TEST_SCHEDULER 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@152 -- # : 0 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@153 -- # export SPDK_TEST_SCANBUILD 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@154 -- # : e810 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@155 -- # export SPDK_TEST_NVMF_NICS 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@156 -- # : 0 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@157 -- # export SPDK_TEST_SMA 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@158 -- # : 0 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@159 -- # export SPDK_TEST_DAOS 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@160 -- # : 0 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@161 -- # export SPDK_TEST_XNVME 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@162 -- # : 0 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@163 -- # export SPDK_TEST_ACCEL 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@164 -- # : 0 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@165 -- # export SPDK_TEST_ACCEL_DSA 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@166 -- # : 0 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@167 -- # export SPDK_TEST_ACCEL_IAA 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@169 -- # : 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@170 -- # export SPDK_TEST_FUZZER_TARGET 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@171 -- # : 0 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@172 -- # export SPDK_TEST_NVMF_MDNS 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@173 -- # : 0 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@174 -- # export SPDK_JSONRPC_GO_CLIENT 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@175 -- # : 0 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@176 -- # export SPDK_TEST_SETUP 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@177 -- # : 0 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@178 -- # export SPDK_TEST_NVME_INTERRUPT 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@181 -- # export SPDK_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@181 -- # SPDK_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@182 -- # export DPDK_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@182 -- # DPDK_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@183 -- # export VFIO_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@183 -- # VFIO_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@184 -- # export LD_LIBRARY_PATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@184 -- # LD_LIBRARY_PATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@187 -- # export PCI_BLOCK_SYNC_ON_RESET=yes 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@187 -- # PCI_BLOCK_SYNC_ON_RESET=yes 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@191 -- # export PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@191 -- # PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@195 -- # export PYTHONDONTWRITEBYTECODE=1 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@195 -- # PYTHONDONTWRITEBYTECODE=1 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@199 -- # export ASAN_OPTIONS=new_delete_type_mismatch=0:disable_coredump=0:abort_on_error=1:use_sigaltstack=0 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@199 -- # ASAN_OPTIONS=new_delete_type_mismatch=0:disable_coredump=0:abort_on_error=1:use_sigaltstack=0 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@200 -- # export UBSAN_OPTIONS=halt_on_error=1:print_stacktrace=1:abort_on_error=1:disable_coredump=0:exitcode=134 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@200 -- # UBSAN_OPTIONS=halt_on_error=1:print_stacktrace=1:abort_on_error=1:disable_coredump=0:exitcode=134 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@204 -- # asan_suppression_file=/var/tmp/asan_suppression_file 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@205 -- # rm -rf /var/tmp/asan_suppression_file 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@206 -- # cat 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@242 -- # echo leak:libfuse3.so 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@244 -- # export LSAN_OPTIONS=suppressions=/var/tmp/asan_suppression_file 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@244 -- # LSAN_OPTIONS=suppressions=/var/tmp/asan_suppression_file 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@246 -- # export DEFAULT_RPC_ADDR=/var/tmp/spdk.sock 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@246 -- # DEFAULT_RPC_ADDR=/var/tmp/spdk.sock 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@248 -- # '[' -z /var/spdk/dependencies ']' 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@251 -- # export DEPENDENCY_DIR 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@255 -- # export SPDK_BIN_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@255 -- # SPDK_BIN_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@256 -- # export SPDK_EXAMPLE_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@256 -- # SPDK_EXAMPLE_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@259 -- # export QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@259 -- # QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@260 -- # export VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@260 -- # VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:14:21.919 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@262 -- # export AR_TOOL=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/ar-xnvme-fixer 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@262 -- # AR_TOOL=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/ar-xnvme-fixer 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@265 -- # export UNBIND_ENTIRE_IOMMU_GROUP=yes 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@265 -- # UNBIND_ENTIRE_IOMMU_GROUP=yes 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@267 -- # _LCOV_MAIN=0 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@268 -- # _LCOV_LLVM=1 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@269 -- # _LCOV= 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@270 -- # [[ '' == *clang* ]] 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@270 -- # [[ 0 -eq 1 ]] 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@272 -- # _lcov_opt[_LCOV_LLVM]='--gcov-tool /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/fuzz/llvm/llvm-gcov.sh' 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@273 -- # _lcov_opt[_LCOV_MAIN]= 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@275 -- # lcov_opt= 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@278 -- # '[' 0 -eq 0 ']' 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@279 -- # export valgrind= 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@279 -- # valgrind= 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@285 -- # uname -s 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@285 -- # '[' Linux = Linux ']' 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@286 -- # HUGEMEM=4096 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@287 -- # export CLEAR_HUGE=yes 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@287 -- # CLEAR_HUGE=yes 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@289 -- # MAKE=make 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@290 -- # MAKEFLAGS=-j48 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@306 -- # export HUGEMEM=4096 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@306 -- # HUGEMEM=4096 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@308 -- # NO_HUGE=() 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@309 -- # TEST_MODE= 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@310 -- # for i in "$@" 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@311 -- # case "$i" in 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@316 -- # TEST_TRANSPORT=tcp 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@331 -- # [[ -z 174132 ]] 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@331 -- # kill -0 174132 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1696 -- # set_test_storage 2147483648 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@341 -- # [[ -v testdir ]] 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@343 -- # local requested_size=2147483648 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@344 -- # local mount target_dir 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@346 -- # local -A mounts fss sizes avails uses 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@347 -- # local source fs size avail mount use 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@349 -- # local storage_fallback storage_candidates 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@351 -- # mktemp -udt spdk.XXXXXX 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@351 -- # storage_fallback=/tmp/spdk.Jq95fo 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@356 -- # storage_candidates=("$testdir" "$storage_fallback/tests/${testdir##*/}" "$storage_fallback") 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@358 -- # [[ -n '' ]] 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@363 -- # [[ -n '' ]] 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@368 -- # mkdir -p /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target /tmp/spdk.Jq95fo/tests/target /tmp/spdk.Jq95fo 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@371 -- # requested_size=2214592512 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # read -r source fs size use avail _ mount 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@340 -- # df -T 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@340 -- # grep -v Filesystem 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # mounts["$mount"]=spdk_devtmpfs 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # fss["$mount"]=devtmpfs 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # avails["$mount"]=67108864 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # sizes["$mount"]=67108864 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@376 -- # uses["$mount"]=0 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # read -r source fs size use avail _ mount 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # mounts["$mount"]=/dev/pmem0 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # fss["$mount"]=ext2 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # avails["$mount"]=4096 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # sizes["$mount"]=5284429824 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@376 -- # uses["$mount"]=5284425728 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # read -r source fs size use avail _ mount 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # mounts["$mount"]=spdk_root 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # fss["$mount"]=overlay 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # avails["$mount"]=53617934336 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # sizes["$mount"]=61988507648 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@376 -- # uses["$mount"]=8370573312 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # read -r source fs size use avail _ mount 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # mounts["$mount"]=tmpfs 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # fss["$mount"]=tmpfs 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # avails["$mount"]=30984220672 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # sizes["$mount"]=30994251776 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@376 -- # uses["$mount"]=10031104 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # read -r source fs size use avail _ mount 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # mounts["$mount"]=tmpfs 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # fss["$mount"]=tmpfs 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # avails["$mount"]=12375273472 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # sizes["$mount"]=12397703168 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@376 -- # uses["$mount"]=22429696 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # read -r source fs size use avail _ mount 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # mounts["$mount"]=tmpfs 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # fss["$mount"]=tmpfs 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # avails["$mount"]=30994075648 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # sizes["$mount"]=30994255872 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@376 -- # uses["$mount"]=180224 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # read -r source fs size use avail _ mount 00:14:21.920 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # mounts["$mount"]=tmpfs 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # fss["$mount"]=tmpfs 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # avails["$mount"]=6198837248 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@375 -- # sizes["$mount"]=6198849536 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@376 -- # uses["$mount"]=12288 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # read -r source fs size use avail _ mount 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@379 -- # printf '* Looking for test storage...\n' 00:14:21.921 * Looking for test storage... 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@381 -- # local target_space new_size 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@382 -- # for target_dir in "${storage_candidates[@]}" 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@385 -- # df /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@385 -- # awk '$1 !~ /Filesystem/{print $6}' 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@385 -- # mount=/ 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@387 -- # target_space=53617934336 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@388 -- # (( target_space == 0 || target_space < requested_size )) 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@391 -- # (( target_space >= requested_size )) 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@393 -- # [[ overlay == tmpfs ]] 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@393 -- # [[ overlay == ramfs ]] 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@393 -- # [[ / == / ]] 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@394 -- # new_size=10585165824 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@395 -- # (( new_size * 100 / sizes[/] > 95 )) 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@400 -- # export SPDK_TEST_STORAGE=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@400 -- # SPDK_TEST_STORAGE=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@401 -- # printf '* Found test storage at %s\n' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:14:21.921 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@402 -- # return 0 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1698 -- # set -o errtrace 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1699 -- # shopt -s extdebug 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1700 -- # trap 'trap - ERR; print_backtrace >&2' ERR 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1702 -- # PS4=' \t ${test_domain:-} -- ${BASH_SOURCE#${BASH_SOURCE%/*/*}/}@${LINENO} -- \$ ' 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1703 -- # true 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1705 -- # xtrace_fd 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@25 -- # [[ -n 15 ]] 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@25 -- # [[ -e /proc/self/fd/15 ]] 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@27 -- # exec 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@29 -- # exec 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@31 -- # xtrace_restore 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@16 -- # unset -v 'X_STACK[0 - 1 < 0 ? 0 : 0 - 1]' 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@17 -- # (( 0 == 0 )) 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@18 -- # set -x 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1711 -- # lcov --version 00:14:21.921 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@333 -- # local ver1 ver1_l 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@334 -- # local ver2 ver2_l 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@336 -- # IFS=.-: 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@336 -- # read -ra ver1 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@337 -- # IFS=.-: 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@337 -- # read -ra ver2 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@338 -- # local 'op=<' 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@340 -- # ver1_l=2 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@341 -- # ver2_l=1 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@344 -- # case "$op" in 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@345 -- # : 1 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@364 -- # (( v = 0 )) 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@365 -- # decimal 1 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@353 -- # local d=1 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@355 -- # echo 1 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@365 -- # ver1[v]=1 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@366 -- # decimal 2 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@353 -- # local d=2 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@355 -- # echo 2 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@366 -- # ver2[v]=2 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@368 -- # return 0 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:14:22.181 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:22.181 --rc genhtml_branch_coverage=1 00:14:22.181 --rc genhtml_function_coverage=1 00:14:22.181 --rc genhtml_legend=1 00:14:22.181 --rc geninfo_all_blocks=1 00:14:22.181 --rc geninfo_unexecuted_blocks=1 00:14:22.181 00:14:22.181 ' 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:14:22.181 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:22.181 --rc genhtml_branch_coverage=1 00:14:22.181 --rc genhtml_function_coverage=1 00:14:22.181 --rc genhtml_legend=1 00:14:22.181 --rc geninfo_all_blocks=1 00:14:22.181 --rc geninfo_unexecuted_blocks=1 00:14:22.181 00:14:22.181 ' 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:14:22.181 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:22.181 --rc genhtml_branch_coverage=1 00:14:22.181 --rc genhtml_function_coverage=1 00:14:22.181 --rc genhtml_legend=1 00:14:22.181 --rc geninfo_all_blocks=1 00:14:22.181 --rc geninfo_unexecuted_blocks=1 00:14:22.181 00:14:22.181 ' 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:14:22.181 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:22.181 --rc genhtml_branch_coverage=1 00:14:22.181 --rc genhtml_function_coverage=1 00:14:22.181 --rc genhtml_legend=1 00:14:22.181 --rc geninfo_all_blocks=1 00:14:22.181 --rc geninfo_unexecuted_blocks=1 00:14:22.181 00:14:22.181 ' 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@7 -- # uname -s 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@15 -- # shopt -s extglob 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:14:22.181 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@5 -- # export PATH 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@51 -- # : 0 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:14:22.182 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@55 -- # have_pci_nics=0 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@12 -- # MALLOC_BDEV_SIZE=512 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@15 -- # nvmftestinit 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@476 -- # prepare_net_devs 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@438 -- # local -g is_hw=no 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@440 -- # remove_spdk_ns 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@309 -- # xtrace_disable 00:14:22.182 17:30:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@10 -- # set +x 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@315 -- # pci_devs=() 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@315 -- # local -a pci_devs 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@316 -- # pci_net_devs=() 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@317 -- # pci_drivers=() 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@317 -- # local -A pci_drivers 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@319 -- # net_devs=() 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@319 -- # local -ga net_devs 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@320 -- # e810=() 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@320 -- # local -ga e810 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@321 -- # x722=() 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@321 -- # local -ga x722 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@322 -- # mlx=() 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@322 -- # local -ga mlx 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:14:24.717 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:14:24.718 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:14:24.718 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@418 -- # [[ up == up ]] 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:14:24.718 Found net devices under 0000:0a:00.0: cvl_0_0 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@418 -- # [[ up == up ]] 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:14:24.718 Found net devices under 0000:0a:00.1: cvl_0_1 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@442 -- # is_hw=yes 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:14:24.718 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:14:24.718 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.186 ms 00:14:24.718 00:14:24.718 --- 10.0.0.2 ping statistics --- 00:14:24.718 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:14:24.718 rtt min/avg/max/mdev = 0.186/0.186/0.186/0.000 ms 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:14:24.718 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:14:24.718 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.052 ms 00:14:24.718 00:14:24.718 --- 10.0.0.1 ping statistics --- 00:14:24.718 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:14:24.718 rtt min/avg/max/mdev = 0.052/0.052/0.052/0.000 ms 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@450 -- # return 0 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@105 -- # run_test nvmf_filesystem_no_in_capsule nvmf_filesystem_part 0 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@10 -- # set +x 00:14:24.718 ************************************ 00:14:24.718 START TEST nvmf_filesystem_no_in_capsule 00:14:24.718 ************************************ 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1129 -- # nvmf_filesystem_part 0 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@47 -- # in_capsule=0 00:14:24.718 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@49 -- # nvmfappstart -m 0xF 00:14:24.719 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:14:24.719 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@726 -- # xtrace_disable 00:14:24.719 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:24.719 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@509 -- # nvmfpid=175779 00:14:24.719 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:14:24.719 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@510 -- # waitforlisten 175779 00:14:24.719 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@835 -- # '[' -z 175779 ']' 00:14:24.719 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:24.719 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:24.719 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:24.719 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:24.719 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:24.719 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:24.719 [2024-12-06 17:30:06.441369] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:14:24.719 [2024-12-06 17:30:06.441479] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:24.719 [2024-12-06 17:30:06.516782] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:14:24.978 [2024-12-06 17:30:06.569833] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:14:24.978 [2024-12-06 17:30:06.569890] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:14:24.978 [2024-12-06 17:30:06.569928] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:14:24.978 [2024-12-06 17:30:06.569939] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:14:24.978 [2024-12-06 17:30:06.569949] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:14:24.978 [2024-12-06 17:30:06.571592] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:14:24.978 [2024-12-06 17:30:06.571674] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:14:24.978 [2024-12-06 17:30:06.571720] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:14:24.978 [2024-12-06 17:30:06.571722] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:24.978 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:24.978 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@868 -- # return 0 00:14:24.978 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:14:24.978 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@732 -- # xtrace_disable 00:14:24.978 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:24.978 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:14:24.978 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@50 -- # malloc_name=Malloc1 00:14:24.978 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@52 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -c 0 00:14:24.978 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:24.978 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:24.978 [2024-12-06 17:30:06.743582] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:14:24.978 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:24.978 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@53 -- # rpc_cmd bdev_malloc_create 512 512 -b Malloc1 00:14:24.978 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:24.978 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:25.237 Malloc1 00:14:25.237 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:25.237 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@54 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:14:25.237 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:25.237 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:25.237 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:25.237 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@55 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:14:25.237 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:25.237 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:25.238 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:25.238 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@56 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:14:25.238 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:25.238 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:25.238 [2024-12-06 17:30:06.926108] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:14:25.238 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:25.238 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@58 -- # get_bdev_size Malloc1 00:14:25.238 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1382 -- # local bdev_name=Malloc1 00:14:25.238 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1383 -- # local bdev_info 00:14:25.238 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1384 -- # local bs 00:14:25.238 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1385 -- # local nb 00:14:25.238 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1386 -- # rpc_cmd bdev_get_bdevs -b Malloc1 00:14:25.238 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:25.238 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:25.238 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:25.238 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1386 -- # bdev_info='[ 00:14:25.238 { 00:14:25.238 "name": "Malloc1", 00:14:25.238 "aliases": [ 00:14:25.238 "6d3e52e8-66a1-4bb4-8256-c6d343596c58" 00:14:25.238 ], 00:14:25.238 "product_name": "Malloc disk", 00:14:25.238 "block_size": 512, 00:14:25.238 "num_blocks": 1048576, 00:14:25.238 "uuid": "6d3e52e8-66a1-4bb4-8256-c6d343596c58", 00:14:25.238 "assigned_rate_limits": { 00:14:25.238 "rw_ios_per_sec": 0, 00:14:25.238 "rw_mbytes_per_sec": 0, 00:14:25.238 "r_mbytes_per_sec": 0, 00:14:25.238 "w_mbytes_per_sec": 0 00:14:25.238 }, 00:14:25.238 "claimed": true, 00:14:25.238 "claim_type": "exclusive_write", 00:14:25.238 "zoned": false, 00:14:25.238 "supported_io_types": { 00:14:25.238 "read": true, 00:14:25.238 "write": true, 00:14:25.238 "unmap": true, 00:14:25.238 "flush": true, 00:14:25.238 "reset": true, 00:14:25.238 "nvme_admin": false, 00:14:25.238 "nvme_io": false, 00:14:25.238 "nvme_io_md": false, 00:14:25.238 "write_zeroes": true, 00:14:25.238 "zcopy": true, 00:14:25.238 "get_zone_info": false, 00:14:25.238 "zone_management": false, 00:14:25.238 "zone_append": false, 00:14:25.238 "compare": false, 00:14:25.238 "compare_and_write": false, 00:14:25.238 "abort": true, 00:14:25.238 "seek_hole": false, 00:14:25.238 "seek_data": false, 00:14:25.238 "copy": true, 00:14:25.238 "nvme_iov_md": false 00:14:25.238 }, 00:14:25.238 "memory_domains": [ 00:14:25.238 { 00:14:25.238 "dma_device_id": "system", 00:14:25.238 "dma_device_type": 1 00:14:25.238 }, 00:14:25.238 { 00:14:25.238 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:25.238 "dma_device_type": 2 00:14:25.238 } 00:14:25.238 ], 00:14:25.238 "driver_specific": {} 00:14:25.238 } 00:14:25.238 ]' 00:14:25.238 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1387 -- # jq '.[] .block_size' 00:14:25.238 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1387 -- # bs=512 00:14:25.238 17:30:06 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1388 -- # jq '.[] .num_blocks' 00:14:25.238 17:30:07 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1388 -- # nb=1048576 00:14:25.238 17:30:07 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1391 -- # bdev_size=512 00:14:25.238 17:30:07 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1392 -- # echo 512 00:14:25.238 17:30:07 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@58 -- # malloc_size=536870912 00:14:25.238 17:30:07 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@60 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:14:25.804 17:30:07 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@62 -- # waitforserial SPDKISFASTANDAWESOME 00:14:25.804 17:30:07 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1202 -- # local i=0 00:14:25.804 17:30:07 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:14:25.804 17:30:07 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:14:25.804 17:30:07 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1209 -- # sleep 2 00:14:28.336 17:30:09 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:14:28.336 17:30:09 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:14:28.336 17:30:09 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:14:28.336 17:30:09 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:14:28.336 17:30:09 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:14:28.336 17:30:09 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1212 -- # return 0 00:14:28.336 17:30:09 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@63 -- # lsblk -l -o NAME,SERIAL 00:14:28.336 17:30:09 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@63 -- # grep -oP '([\w]*)(?=\s+SPDKISFASTANDAWESOME)' 00:14:28.336 17:30:09 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@63 -- # nvme_name=nvme0n1 00:14:28.336 17:30:09 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@64 -- # sec_size_to_bytes nvme0n1 00:14:28.336 17:30:09 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- setup/common.sh@76 -- # local dev=nvme0n1 00:14:28.336 17:30:09 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- setup/common.sh@78 -- # [[ -e /sys/block/nvme0n1 ]] 00:14:28.336 17:30:09 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- setup/common.sh@80 -- # echo 536870912 00:14:28.336 17:30:09 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@64 -- # nvme_size=536870912 00:14:28.336 17:30:09 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@66 -- # mkdir -p /mnt/device 00:14:28.336 17:30:09 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@67 -- # (( nvme_size == malloc_size )) 00:14:28.336 17:30:09 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@68 -- # parted -s /dev/nvme0n1 mklabel gpt mkpart SPDK_TEST 0% 100% 00:14:28.336 17:30:09 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@69 -- # partprobe 00:14:28.595 17:30:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@70 -- # sleep 1 00:14:29.971 17:30:11 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@76 -- # '[' 0 -eq 0 ']' 00:14:29.971 17:30:11 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@77 -- # run_test filesystem_ext4 nvmf_filesystem_create ext4 nvme0n1 00:14:29.971 17:30:11 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:14:29.971 17:30:11 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:29.971 17:30:11 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:29.971 ************************************ 00:14:29.971 START TEST filesystem_ext4 00:14:29.971 ************************************ 00:14:29.971 17:30:11 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@1129 -- # nvmf_filesystem_create ext4 nvme0n1 00:14:29.972 17:30:11 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@18 -- # fstype=ext4 00:14:29.972 17:30:11 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:14:29.972 17:30:11 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@21 -- # make_filesystem ext4 /dev/nvme0n1p1 00:14:29.972 17:30:11 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@930 -- # local fstype=ext4 00:14:29.972 17:30:11 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@931 -- # local dev_name=/dev/nvme0n1p1 00:14:29.972 17:30:11 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@932 -- # local i=0 00:14:29.972 17:30:11 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@933 -- # local force 00:14:29.972 17:30:11 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@935 -- # '[' ext4 = ext4 ']' 00:14:29.972 17:30:11 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@936 -- # force=-F 00:14:29.972 17:30:11 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@941 -- # mkfs.ext4 -F /dev/nvme0n1p1 00:14:29.972 mke2fs 1.47.0 (5-Feb-2023) 00:14:29.972 Discarding device blocks: 0/522240 done 00:14:29.972 Creating filesystem with 522240 1k blocks and 130560 inodes 00:14:29.972 Filesystem UUID: 1358a6de-bb62-4d18-b520-a45126abc0ac 00:14:29.972 Superblock backups stored on blocks: 00:14:29.972 8193, 24577, 40961, 57345, 73729, 204801, 221185, 401409 00:14:29.972 00:14:29.972 Allocating group tables: 0/64 done 00:14:29.972 Writing inode tables: 0/64 done 00:14:29.972 Creating journal (8192 blocks): done 00:14:29.972 Writing superblocks and filesystem accounting information: 0/64 done 00:14:29.972 00:14:29.972 17:30:11 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@949 -- # return 0 00:14:29.972 17:30:11 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:14:36.528 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:14:36.528 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@25 -- # sync 00:14:36.528 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:14:36.528 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@27 -- # sync 00:14:36.528 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@29 -- # i=0 00:14:36.528 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@30 -- # umount /mnt/device 00:14:36.528 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@37 -- # kill -0 175779 00:14:36.528 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:14:36.528 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:14:36.528 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:14:36.528 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:14:36.528 00:14:36.528 real 0m5.821s 00:14:36.528 user 0m0.018s 00:14:36.528 sys 0m0.094s 00:14:36.528 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:36.528 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@10 -- # set +x 00:14:36.528 ************************************ 00:14:36.528 END TEST filesystem_ext4 00:14:36.528 ************************************ 00:14:36.528 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@78 -- # run_test filesystem_btrfs nvmf_filesystem_create btrfs nvme0n1 00:14:36.528 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:14:36.528 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:36.528 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:36.528 ************************************ 00:14:36.528 START TEST filesystem_btrfs 00:14:36.528 ************************************ 00:14:36.528 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@1129 -- # nvmf_filesystem_create btrfs nvme0n1 00:14:36.528 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@18 -- # fstype=btrfs 00:14:36.528 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:14:36.528 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@21 -- # make_filesystem btrfs /dev/nvme0n1p1 00:14:36.528 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@930 -- # local fstype=btrfs 00:14:36.528 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@931 -- # local dev_name=/dev/nvme0n1p1 00:14:36.528 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@932 -- # local i=0 00:14:36.528 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@933 -- # local force 00:14:36.528 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@935 -- # '[' btrfs = ext4 ']' 00:14:36.528 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@938 -- # force=-f 00:14:36.528 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@941 -- # mkfs.btrfs -f /dev/nvme0n1p1 00:14:36.528 btrfs-progs v6.8.1 00:14:36.528 See https://btrfs.readthedocs.io for more information. 00:14:36.528 00:14:36.528 Performing full device TRIM /dev/nvme0n1p1 (510.00MiB) ... 00:14:36.528 NOTE: several default settings have changed in version 5.15, please make sure 00:14:36.528 this does not affect your deployments: 00:14:36.528 - DUP for metadata (-m dup) 00:14:36.528 - enabled no-holes (-O no-holes) 00:14:36.528 - enabled free-space-tree (-R free-space-tree) 00:14:36.528 00:14:36.528 Label: (null) 00:14:36.528 UUID: 41fc8f96-23eb-45b3-b649-9a579f1fa685 00:14:36.528 Node size: 16384 00:14:36.528 Sector size: 4096 (CPU page size: 4096) 00:14:36.528 Filesystem size: 510.00MiB 00:14:36.528 Block group profiles: 00:14:36.528 Data: single 8.00MiB 00:14:36.528 Metadata: DUP 32.00MiB 00:14:36.528 System: DUP 8.00MiB 00:14:36.528 SSD detected: yes 00:14:36.528 Zoned device: no 00:14:36.528 Features: extref, skinny-metadata, no-holes, free-space-tree 00:14:36.528 Checksum: crc32c 00:14:36.528 Number of devices: 1 00:14:36.528 Devices: 00:14:36.529 ID SIZE PATH 00:14:36.529 1 510.00MiB /dev/nvme0n1p1 00:14:36.529 00:14:36.529 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@949 -- # return 0 00:14:36.529 17:30:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:14:36.787 17:30:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:14:36.788 17:30:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@25 -- # sync 00:14:36.788 17:30:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:14:36.788 17:30:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@27 -- # sync 00:14:36.788 17:30:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@29 -- # i=0 00:14:36.788 17:30:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@30 -- # umount /mnt/device 00:14:36.788 17:30:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@37 -- # kill -0 175779 00:14:36.788 17:30:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:14:36.788 17:30:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:14:36.788 17:30:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:14:36.788 17:30:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:14:36.788 00:14:36.788 real 0m1.174s 00:14:36.788 user 0m0.015s 00:14:36.788 sys 0m0.132s 00:14:36.788 17:30:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:36.788 17:30:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@10 -- # set +x 00:14:36.788 ************************************ 00:14:36.788 END TEST filesystem_btrfs 00:14:36.788 ************************************ 00:14:36.788 17:30:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@79 -- # run_test filesystem_xfs nvmf_filesystem_create xfs nvme0n1 00:14:36.788 17:30:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:14:36.788 17:30:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:36.788 17:30:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:36.788 ************************************ 00:14:36.788 START TEST filesystem_xfs 00:14:36.788 ************************************ 00:14:36.788 17:30:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@1129 -- # nvmf_filesystem_create xfs nvme0n1 00:14:36.788 17:30:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@18 -- # fstype=xfs 00:14:36.788 17:30:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:14:36.788 17:30:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@21 -- # make_filesystem xfs /dev/nvme0n1p1 00:14:36.788 17:30:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@930 -- # local fstype=xfs 00:14:36.788 17:30:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@931 -- # local dev_name=/dev/nvme0n1p1 00:14:36.788 17:30:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@932 -- # local i=0 00:14:36.788 17:30:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@933 -- # local force 00:14:36.788 17:30:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@935 -- # '[' xfs = ext4 ']' 00:14:36.788 17:30:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@938 -- # force=-f 00:14:36.788 17:30:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@941 -- # mkfs.xfs -f /dev/nvme0n1p1 00:14:37.047 meta-data=/dev/nvme0n1p1 isize=512 agcount=4, agsize=32640 blks 00:14:37.047 = sectsz=512 attr=2, projid32bit=1 00:14:37.047 = crc=1 finobt=1, sparse=1, rmapbt=0 00:14:37.047 = reflink=1 bigtime=1 inobtcount=1 nrext64=0 00:14:37.047 data = bsize=4096 blocks=130560, imaxpct=25 00:14:37.047 = sunit=0 swidth=0 blks 00:14:37.047 naming =version 2 bsize=4096 ascii-ci=0, ftype=1 00:14:37.047 log =internal log bsize=4096 blocks=16384, version=2 00:14:37.047 = sectsz=512 sunit=0 blks, lazy-count=1 00:14:37.047 realtime =none extsz=4096 blocks=0, rtextents=0 00:14:37.981 Discarding blocks...Done. 00:14:37.981 17:30:19 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@949 -- # return 0 00:14:37.981 17:30:19 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:14:41.264 17:30:22 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:14:41.264 17:30:22 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@25 -- # sync 00:14:41.264 17:30:22 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:14:41.264 17:30:22 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@27 -- # sync 00:14:41.264 17:30:22 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@29 -- # i=0 00:14:41.264 17:30:22 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@30 -- # umount /mnt/device 00:14:41.264 17:30:22 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@37 -- # kill -0 175779 00:14:41.264 17:30:22 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:14:41.265 17:30:22 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:14:41.265 17:30:22 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:14:41.265 17:30:22 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:14:41.265 00:14:41.265 real 0m4.316s 00:14:41.265 user 0m0.023s 00:14:41.265 sys 0m0.083s 00:14:41.265 17:30:22 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:41.265 17:30:22 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@10 -- # set +x 00:14:41.265 ************************************ 00:14:41.265 END TEST filesystem_xfs 00:14:41.265 ************************************ 00:14:41.265 17:30:22 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@91 -- # flock /dev/nvme0n1 parted -s /dev/nvme0n1 rm 1 00:14:41.523 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@93 -- # sync 00:14:41.523 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@94 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:14:41.523 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:14:41.523 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@95 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:14:41.523 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1223 -- # local i=0 00:14:41.523 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:14:41.523 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:14:41.523 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:14:41.523 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:14:41.523 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1235 -- # return 0 00:14:41.523 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@97 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:14:41.523 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:41.523 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:41.523 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:41.523 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@99 -- # trap - SIGINT SIGTERM EXIT 00:14:41.523 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@101 -- # killprocess 175779 00:14:41.523 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@954 -- # '[' -z 175779 ']' 00:14:41.524 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@958 -- # kill -0 175779 00:14:41.524 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@959 -- # uname 00:14:41.524 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:41.524 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 175779 00:14:41.524 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:41.524 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:41.524 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@972 -- # echo 'killing process with pid 175779' 00:14:41.524 killing process with pid 175779 00:14:41.524 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@973 -- # kill 175779 00:14:41.524 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@978 -- # wait 175779 00:14:42.091 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@102 -- # nvmfpid= 00:14:42.091 00:14:42.091 real 0m17.304s 00:14:42.091 user 1m7.082s 00:14:42.091 sys 0m2.265s 00:14:42.091 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:42.091 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:42.091 ************************************ 00:14:42.091 END TEST nvmf_filesystem_no_in_capsule 00:14:42.091 ************************************ 00:14:42.091 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@106 -- # run_test nvmf_filesystem_in_capsule nvmf_filesystem_part 4096 00:14:42.091 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:14:42.091 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:42.091 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@10 -- # set +x 00:14:42.091 ************************************ 00:14:42.091 START TEST nvmf_filesystem_in_capsule 00:14:42.091 ************************************ 00:14:42.091 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1129 -- # nvmf_filesystem_part 4096 00:14:42.091 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@47 -- # in_capsule=4096 00:14:42.091 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@49 -- # nvmfappstart -m 0xF 00:14:42.091 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:14:42.091 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@726 -- # xtrace_disable 00:14:42.091 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:42.091 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@509 -- # nvmfpid=178523 00:14:42.091 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:14:42.091 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@510 -- # waitforlisten 178523 00:14:42.091 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@835 -- # '[' -z 178523 ']' 00:14:42.091 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:42.091 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@840 -- # local max_retries=100 00:14:42.091 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:42.091 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:42.091 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@844 -- # xtrace_disable 00:14:42.091 17:30:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:42.091 [2024-12-06 17:30:23.800169] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:14:42.091 [2024-12-06 17:30:23.800250] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:42.091 [2024-12-06 17:30:23.875235] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:14:42.091 [2024-12-06 17:30:23.924244] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:14:42.091 [2024-12-06 17:30:23.924296] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:14:42.091 [2024-12-06 17:30:23.924310] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:14:42.091 [2024-12-06 17:30:23.924321] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:14:42.091 [2024-12-06 17:30:23.924331] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:14:42.091 [2024-12-06 17:30:23.925910] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:14:42.091 [2024-12-06 17:30:23.926035] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:14:42.091 [2024-12-06 17:30:23.926102] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:14:42.091 [2024-12-06 17:30:23.926105] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:42.350 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:14:42.350 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@868 -- # return 0 00:14:42.350 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:14:42.350 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@732 -- # xtrace_disable 00:14:42.350 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:42.350 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:14:42.350 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@50 -- # malloc_name=Malloc1 00:14:42.350 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@52 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -c 4096 00:14:42.350 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.350 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:42.350 [2024-12-06 17:30:24.074275] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:14:42.350 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.350 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@53 -- # rpc_cmd bdev_malloc_create 512 512 -b Malloc1 00:14:42.350 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.350 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:42.609 Malloc1 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@54 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@55 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@56 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:42.609 [2024-12-06 17:30:24.284199] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@58 -- # get_bdev_size Malloc1 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1382 -- # local bdev_name=Malloc1 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1383 -- # local bdev_info 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1384 -- # local bs 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1385 -- # local nb 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1386 -- # rpc_cmd bdev_get_bdevs -b Malloc1 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1386 -- # bdev_info='[ 00:14:42.609 { 00:14:42.609 "name": "Malloc1", 00:14:42.609 "aliases": [ 00:14:42.609 "30cd3760-18dc-44ac-80f1-4f288379c629" 00:14:42.609 ], 00:14:42.609 "product_name": "Malloc disk", 00:14:42.609 "block_size": 512, 00:14:42.609 "num_blocks": 1048576, 00:14:42.609 "uuid": "30cd3760-18dc-44ac-80f1-4f288379c629", 00:14:42.609 "assigned_rate_limits": { 00:14:42.609 "rw_ios_per_sec": 0, 00:14:42.609 "rw_mbytes_per_sec": 0, 00:14:42.609 "r_mbytes_per_sec": 0, 00:14:42.609 "w_mbytes_per_sec": 0 00:14:42.609 }, 00:14:42.609 "claimed": true, 00:14:42.609 "claim_type": "exclusive_write", 00:14:42.609 "zoned": false, 00:14:42.609 "supported_io_types": { 00:14:42.609 "read": true, 00:14:42.609 "write": true, 00:14:42.609 "unmap": true, 00:14:42.609 "flush": true, 00:14:42.609 "reset": true, 00:14:42.609 "nvme_admin": false, 00:14:42.609 "nvme_io": false, 00:14:42.609 "nvme_io_md": false, 00:14:42.609 "write_zeroes": true, 00:14:42.609 "zcopy": true, 00:14:42.609 "get_zone_info": false, 00:14:42.609 "zone_management": false, 00:14:42.609 "zone_append": false, 00:14:42.609 "compare": false, 00:14:42.609 "compare_and_write": false, 00:14:42.609 "abort": true, 00:14:42.609 "seek_hole": false, 00:14:42.609 "seek_data": false, 00:14:42.609 "copy": true, 00:14:42.609 "nvme_iov_md": false 00:14:42.609 }, 00:14:42.609 "memory_domains": [ 00:14:42.609 { 00:14:42.609 "dma_device_id": "system", 00:14:42.609 "dma_device_type": 1 00:14:42.609 }, 00:14:42.609 { 00:14:42.609 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:42.609 "dma_device_type": 2 00:14:42.609 } 00:14:42.609 ], 00:14:42.609 "driver_specific": {} 00:14:42.609 } 00:14:42.609 ]' 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1387 -- # jq '.[] .block_size' 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1387 -- # bs=512 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1388 -- # jq '.[] .num_blocks' 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1388 -- # nb=1048576 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1391 -- # bdev_size=512 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1392 -- # echo 512 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@58 -- # malloc_size=536870912 00:14:42.609 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@60 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:14:43.178 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@62 -- # waitforserial SPDKISFASTANDAWESOME 00:14:43.178 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1202 -- # local i=0 00:14:43.178 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:14:43.178 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:14:43.178 17:30:24 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1209 -- # sleep 2 00:14:45.708 17:30:26 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:14:45.708 17:30:26 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:14:45.708 17:30:26 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:14:45.708 17:30:26 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:14:45.708 17:30:26 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:14:45.708 17:30:26 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1212 -- # return 0 00:14:45.708 17:30:26 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@63 -- # lsblk -l -o NAME,SERIAL 00:14:45.708 17:30:26 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@63 -- # grep -oP '([\w]*)(?=\s+SPDKISFASTANDAWESOME)' 00:14:45.708 17:30:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@63 -- # nvme_name=nvme0n1 00:14:45.708 17:30:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@64 -- # sec_size_to_bytes nvme0n1 00:14:45.708 17:30:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- setup/common.sh@76 -- # local dev=nvme0n1 00:14:45.708 17:30:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- setup/common.sh@78 -- # [[ -e /sys/block/nvme0n1 ]] 00:14:45.708 17:30:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- setup/common.sh@80 -- # echo 536870912 00:14:45.708 17:30:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@64 -- # nvme_size=536870912 00:14:45.708 17:30:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@66 -- # mkdir -p /mnt/device 00:14:45.708 17:30:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@67 -- # (( nvme_size == malloc_size )) 00:14:45.708 17:30:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@68 -- # parted -s /dev/nvme0n1 mklabel gpt mkpart SPDK_TEST 0% 100% 00:14:45.708 17:30:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@69 -- # partprobe 00:14:45.966 17:30:27 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@70 -- # sleep 1 00:14:46.902 17:30:28 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@76 -- # '[' 4096 -eq 0 ']' 00:14:46.902 17:30:28 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@81 -- # run_test filesystem_in_capsule_ext4 nvmf_filesystem_create ext4 nvme0n1 00:14:46.902 17:30:28 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:14:46.902 17:30:28 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:46.902 17:30:28 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:46.902 ************************************ 00:14:46.902 START TEST filesystem_in_capsule_ext4 00:14:46.902 ************************************ 00:14:46.902 17:30:28 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@1129 -- # nvmf_filesystem_create ext4 nvme0n1 00:14:46.902 17:30:28 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@18 -- # fstype=ext4 00:14:46.902 17:30:28 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:14:46.902 17:30:28 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@21 -- # make_filesystem ext4 /dev/nvme0n1p1 00:14:46.902 17:30:28 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@930 -- # local fstype=ext4 00:14:46.902 17:30:28 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@931 -- # local dev_name=/dev/nvme0n1p1 00:14:46.902 17:30:28 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@932 -- # local i=0 00:14:46.902 17:30:28 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@933 -- # local force 00:14:46.902 17:30:28 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@935 -- # '[' ext4 = ext4 ']' 00:14:46.902 17:30:28 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@936 -- # force=-F 00:14:46.902 17:30:28 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@941 -- # mkfs.ext4 -F /dev/nvme0n1p1 00:14:46.902 mke2fs 1.47.0 (5-Feb-2023) 00:14:47.160 Discarding device blocks: 0/522240 done 00:14:47.160 Creating filesystem with 522240 1k blocks and 130560 inodes 00:14:47.160 Filesystem UUID: d35674e3-52fb-40e3-97fb-13c40bf57adf 00:14:47.160 Superblock backups stored on blocks: 00:14:47.160 8193, 24577, 40961, 57345, 73729, 204801, 221185, 401409 00:14:47.160 00:14:47.160 Allocating group tables: 0/64 done 00:14:47.160 Writing inode tables: 0/64 done 00:14:47.160 Creating journal (8192 blocks): done 00:14:48.535 Writing superblocks and filesystem accounting information: 0/64 done 00:14:48.535 00:14:48.535 17:30:30 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@949 -- # return 0 00:14:48.535 17:30:30 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:14:55.095 17:30:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:14:55.095 17:30:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@25 -- # sync 00:14:55.095 17:30:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:14:55.095 17:30:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@27 -- # sync 00:14:55.095 17:30:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@29 -- # i=0 00:14:55.095 17:30:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@30 -- # umount /mnt/device 00:14:55.095 17:30:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@37 -- # kill -0 178523 00:14:55.095 17:30:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:14:55.095 17:30:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:14:55.095 17:30:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:14:55.095 17:30:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:14:55.095 00:14:55.095 real 0m7.198s 00:14:55.095 user 0m0.024s 00:14:55.095 sys 0m0.057s 00:14:55.095 17:30:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:55.095 17:30:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@10 -- # set +x 00:14:55.095 ************************************ 00:14:55.095 END TEST filesystem_in_capsule_ext4 00:14:55.095 ************************************ 00:14:55.095 17:30:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@82 -- # run_test filesystem_in_capsule_btrfs nvmf_filesystem_create btrfs nvme0n1 00:14:55.095 17:30:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:14:55.095 17:30:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:55.095 17:30:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:55.095 ************************************ 00:14:55.095 START TEST filesystem_in_capsule_btrfs 00:14:55.095 ************************************ 00:14:55.095 17:30:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@1129 -- # nvmf_filesystem_create btrfs nvme0n1 00:14:55.095 17:30:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@18 -- # fstype=btrfs 00:14:55.095 17:30:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:14:55.095 17:30:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@21 -- # make_filesystem btrfs /dev/nvme0n1p1 00:14:55.095 17:30:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@930 -- # local fstype=btrfs 00:14:55.095 17:30:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@931 -- # local dev_name=/dev/nvme0n1p1 00:14:55.095 17:30:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@932 -- # local i=0 00:14:55.095 17:30:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@933 -- # local force 00:14:55.095 17:30:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@935 -- # '[' btrfs = ext4 ']' 00:14:55.095 17:30:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@938 -- # force=-f 00:14:55.095 17:30:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@941 -- # mkfs.btrfs -f /dev/nvme0n1p1 00:14:55.095 btrfs-progs v6.8.1 00:14:55.095 See https://btrfs.readthedocs.io for more information. 00:14:55.095 00:14:55.095 Performing full device TRIM /dev/nvme0n1p1 (510.00MiB) ... 00:14:55.095 NOTE: several default settings have changed in version 5.15, please make sure 00:14:55.095 this does not affect your deployments: 00:14:55.095 - DUP for metadata (-m dup) 00:14:55.095 - enabled no-holes (-O no-holes) 00:14:55.095 - enabled free-space-tree (-R free-space-tree) 00:14:55.095 00:14:55.095 Label: (null) 00:14:55.095 UUID: 1ec35c40-66e1-49e8-ad30-60dc480668fa 00:14:55.095 Node size: 16384 00:14:55.095 Sector size: 4096 (CPU page size: 4096) 00:14:55.095 Filesystem size: 510.00MiB 00:14:55.095 Block group profiles: 00:14:55.095 Data: single 8.00MiB 00:14:55.095 Metadata: DUP 32.00MiB 00:14:55.095 System: DUP 8.00MiB 00:14:55.095 SSD detected: yes 00:14:55.095 Zoned device: no 00:14:55.095 Features: extref, skinny-metadata, no-holes, free-space-tree 00:14:55.095 Checksum: crc32c 00:14:55.095 Number of devices: 1 00:14:55.095 Devices: 00:14:55.095 ID SIZE PATH 00:14:55.095 1 510.00MiB /dev/nvme0n1p1 00:14:55.095 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@949 -- # return 0 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@25 -- # sync 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@27 -- # sync 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@29 -- # i=0 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@30 -- # umount /mnt/device 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@37 -- # kill -0 178523 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:14:55.095 00:14:55.095 real 0m0.696s 00:14:55.095 user 0m0.017s 00:14:55.095 sys 0m0.097s 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@10 -- # set +x 00:14:55.095 ************************************ 00:14:55.095 END TEST filesystem_in_capsule_btrfs 00:14:55.095 ************************************ 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@83 -- # run_test filesystem_in_capsule_xfs nvmf_filesystem_create xfs nvme0n1 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1111 -- # xtrace_disable 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:55.095 ************************************ 00:14:55.095 START TEST filesystem_in_capsule_xfs 00:14:55.095 ************************************ 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@1129 -- # nvmf_filesystem_create xfs nvme0n1 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@18 -- # fstype=xfs 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@21 -- # make_filesystem xfs /dev/nvme0n1p1 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@930 -- # local fstype=xfs 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@931 -- # local dev_name=/dev/nvme0n1p1 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@932 -- # local i=0 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@933 -- # local force 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@935 -- # '[' xfs = ext4 ']' 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@938 -- # force=-f 00:14:55.095 17:30:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@941 -- # mkfs.xfs -f /dev/nvme0n1p1 00:14:55.095 meta-data=/dev/nvme0n1p1 isize=512 agcount=4, agsize=32640 blks 00:14:55.095 = sectsz=512 attr=2, projid32bit=1 00:14:55.095 = crc=1 finobt=1, sparse=1, rmapbt=0 00:14:55.095 = reflink=1 bigtime=1 inobtcount=1 nrext64=0 00:14:55.095 data = bsize=4096 blocks=130560, imaxpct=25 00:14:55.095 = sunit=0 swidth=0 blks 00:14:55.095 naming =version 2 bsize=4096 ascii-ci=0, ftype=1 00:14:55.095 log =internal log bsize=4096 blocks=16384, version=2 00:14:55.095 = sectsz=512 sunit=0 blks, lazy-count=1 00:14:55.095 realtime =none extsz=4096 blocks=0, rtextents=0 00:14:55.661 Discarding blocks...Done. 00:14:55.661 17:30:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@949 -- # return 0 00:14:55.661 17:30:37 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:14:58.187 17:30:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:14:58.187 17:30:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@25 -- # sync 00:14:58.187 17:30:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:14:58.187 17:30:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@27 -- # sync 00:14:58.187 17:30:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@29 -- # i=0 00:14:58.187 17:30:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@30 -- # umount /mnt/device 00:14:58.187 17:30:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@37 -- # kill -0 178523 00:14:58.187 17:30:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:14:58.187 17:30:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:14:58.187 17:30:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:14:58.187 17:30:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:14:58.187 00:14:58.187 real 0m3.273s 00:14:58.187 user 0m0.011s 00:14:58.187 sys 0m0.065s 00:14:58.187 17:30:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:58.187 17:30:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@10 -- # set +x 00:14:58.187 ************************************ 00:14:58.187 END TEST filesystem_in_capsule_xfs 00:14:58.187 ************************************ 00:14:58.187 17:30:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@91 -- # flock /dev/nvme0n1 parted -s /dev/nvme0n1 rm 1 00:14:58.187 17:30:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@93 -- # sync 00:14:58.187 17:30:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@94 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:14:58.187 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:14:58.187 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@95 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:14:58.187 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1223 -- # local i=0 00:14:58.187 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:14:58.187 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:14:58.445 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:14:58.445 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:14:58.445 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1235 -- # return 0 00:14:58.445 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@97 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:14:58.445 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@563 -- # xtrace_disable 00:14:58.445 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:58.445 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:14:58.445 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@99 -- # trap - SIGINT SIGTERM EXIT 00:14:58.445 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@101 -- # killprocess 178523 00:14:58.445 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@954 -- # '[' -z 178523 ']' 00:14:58.445 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@958 -- # kill -0 178523 00:14:58.445 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@959 -- # uname 00:14:58.445 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:14:58.445 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 178523 00:14:58.445 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:14:58.445 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:14:58.445 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@972 -- # echo 'killing process with pid 178523' 00:14:58.445 killing process with pid 178523 00:14:58.445 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@973 -- # kill 178523 00:14:58.445 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@978 -- # wait 178523 00:14:58.705 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@102 -- # nvmfpid= 00:14:58.705 00:14:58.705 real 0m16.749s 00:14:58.705 user 1m4.994s 00:14:58.705 sys 0m2.030s 00:14:58.705 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1130 -- # xtrace_disable 00:14:58.705 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:58.705 ************************************ 00:14:58.705 END TEST nvmf_filesystem_in_capsule 00:14:58.705 ************************************ 00:14:58.705 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@108 -- # nvmftestfini 00:14:58.705 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@516 -- # nvmfcleanup 00:14:58.705 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@121 -- # sync 00:14:58.705 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:14:58.705 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@124 -- # set +e 00:14:58.705 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@125 -- # for i in {1..20} 00:14:58.705 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:14:58.705 rmmod nvme_tcp 00:14:58.705 rmmod nvme_fabrics 00:14:58.965 rmmod nvme_keyring 00:14:58.965 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:14:58.965 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@128 -- # set -e 00:14:58.965 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@129 -- # return 0 00:14:58.965 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@517 -- # '[' -n '' ']' 00:14:58.965 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:14:58.965 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:14:58.965 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:14:58.965 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@297 -- # iptr 00:14:58.965 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@791 -- # iptables-save 00:14:58.965 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:14:58.965 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@791 -- # iptables-restore 00:14:58.965 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:14:58.965 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@302 -- # remove_spdk_ns 00:14:58.965 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:14:58.965 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:14:58.965 17:30:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:15:00.876 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:15:00.876 00:15:00.876 real 0m39.132s 00:15:00.876 user 2m13.209s 00:15:00.876 sys 0m6.151s 00:15:00.876 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:00.876 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@10 -- # set +x 00:15:00.876 ************************************ 00:15:00.876 END TEST nvmf_filesystem 00:15:00.876 ************************************ 00:15:00.876 17:30:42 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@18 -- # run_test nvmf_target_discovery /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/discovery.sh --transport=tcp 00:15:00.876 17:30:42 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:15:00.876 17:30:42 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:00.876 17:30:42 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:15:00.876 ************************************ 00:15:00.876 START TEST nvmf_target_discovery 00:15:00.876 ************************************ 00:15:00.876 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/discovery.sh --transport=tcp 00:15:00.876 * Looking for test storage... 00:15:01.136 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1711 -- # lcov --version 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@333 -- # local ver1 ver1_l 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@334 -- # local ver2 ver2_l 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@336 -- # IFS=.-: 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@336 -- # read -ra ver1 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@337 -- # IFS=.-: 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@337 -- # read -ra ver2 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@338 -- # local 'op=<' 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@340 -- # ver1_l=2 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@341 -- # ver2_l=1 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@344 -- # case "$op" in 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@345 -- # : 1 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@364 -- # (( v = 0 )) 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@365 -- # decimal 1 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@353 -- # local d=1 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@355 -- # echo 1 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@365 -- # ver1[v]=1 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@366 -- # decimal 2 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@353 -- # local d=2 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@355 -- # echo 2 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@366 -- # ver2[v]=2 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@368 -- # return 0 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:15:01.136 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:01.136 --rc genhtml_branch_coverage=1 00:15:01.136 --rc genhtml_function_coverage=1 00:15:01.136 --rc genhtml_legend=1 00:15:01.136 --rc geninfo_all_blocks=1 00:15:01.136 --rc geninfo_unexecuted_blocks=1 00:15:01.136 00:15:01.136 ' 00:15:01.136 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:15:01.136 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:01.137 --rc genhtml_branch_coverage=1 00:15:01.137 --rc genhtml_function_coverage=1 00:15:01.137 --rc genhtml_legend=1 00:15:01.137 --rc geninfo_all_blocks=1 00:15:01.137 --rc geninfo_unexecuted_blocks=1 00:15:01.137 00:15:01.137 ' 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:15:01.137 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:01.137 --rc genhtml_branch_coverage=1 00:15:01.137 --rc genhtml_function_coverage=1 00:15:01.137 --rc genhtml_legend=1 00:15:01.137 --rc geninfo_all_blocks=1 00:15:01.137 --rc geninfo_unexecuted_blocks=1 00:15:01.137 00:15:01.137 ' 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:15:01.137 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:01.137 --rc genhtml_branch_coverage=1 00:15:01.137 --rc genhtml_function_coverage=1 00:15:01.137 --rc genhtml_legend=1 00:15:01.137 --rc geninfo_all_blocks=1 00:15:01.137 --rc geninfo_unexecuted_blocks=1 00:15:01.137 00:15:01.137 ' 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@7 -- # uname -s 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@15 -- # shopt -s extglob 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- paths/export.sh@5 -- # export PATH 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@51 -- # : 0 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:15:01.137 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@55 -- # have_pci_nics=0 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@11 -- # NULL_BDEV_SIZE=102400 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@12 -- # NULL_BLOCK_SIZE=512 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@13 -- # NVMF_PORT_REFERRAL=4430 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@15 -- # hash nvme 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@20 -- # nvmftestinit 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@476 -- # prepare_net_devs 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@438 -- # local -g is_hw=no 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@440 -- # remove_spdk_ns 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@309 -- # xtrace_disable 00:15:01.137 17:30:42 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@315 -- # pci_devs=() 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@315 -- # local -a pci_devs 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@316 -- # pci_net_devs=() 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@317 -- # pci_drivers=() 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@317 -- # local -A pci_drivers 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@319 -- # net_devs=() 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@319 -- # local -ga net_devs 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@320 -- # e810=() 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@320 -- # local -ga e810 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@321 -- # x722=() 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@321 -- # local -ga x722 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@322 -- # mlx=() 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@322 -- # local -ga mlx 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:15:03.673 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:15:03.673 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@418 -- # [[ up == up ]] 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:15:03.673 Found net devices under 0000:0a:00.0: cvl_0_0 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@418 -- # [[ up == up ]] 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:15:03.673 Found net devices under 0000:0a:00.1: cvl_0_1 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@442 -- # is_hw=yes 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:15:03.673 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:15:03.674 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:15:03.674 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.283 ms 00:15:03.674 00:15:03.674 --- 10.0.0.2 ping statistics --- 00:15:03.674 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:03.674 rtt min/avg/max/mdev = 0.283/0.283/0.283/0.000 ms 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:15:03.674 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:15:03.674 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.157 ms 00:15:03.674 00:15:03.674 --- 10.0.0.1 ping statistics --- 00:15:03.674 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:03.674 rtt min/avg/max/mdev = 0.157/0.157/0.157/0.000 ms 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@450 -- # return 0 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@21 -- # nvmfappstart -m 0xF 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@726 -- # xtrace_disable 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@509 -- # nvmfpid=182680 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@510 -- # waitforlisten 182680 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@835 -- # '[' -z 182680 ']' 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:03.674 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:03.674 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:03.674 [2024-12-06 17:30:45.304797] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:15:03.674 [2024-12-06 17:30:45.304904] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:03.674 [2024-12-06 17:30:45.377263] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:15:03.674 [2024-12-06 17:30:45.427814] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:15:03.674 [2024-12-06 17:30:45.427879] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:15:03.674 [2024-12-06 17:30:45.427908] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:15:03.674 [2024-12-06 17:30:45.427920] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:15:03.674 [2024-12-06 17:30:45.427930] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:15:03.674 [2024-12-06 17:30:45.429495] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:15:03.674 [2024-12-06 17:30:45.429557] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:15:03.674 [2024-12-06 17:30:45.429619] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:15:03.674 [2024-12-06 17:30:45.429622] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@868 -- # return 0 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@732 -- # xtrace_disable 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@23 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:03.933 [2024-12-06 17:30:45.581888] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@26 -- # seq 1 4 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@26 -- # for i in $(seq 1 4) 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@27 -- # rpc_cmd bdev_null_create Null1 102400 512 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:03.933 Null1 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Null1 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@30 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:03.933 [2024-12-06 17:30:45.641840] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@26 -- # for i in $(seq 1 4) 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@27 -- # rpc_cmd bdev_null_create Null2 102400 512 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:03.933 Null2 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 -a -s SPDK00000000000002 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 Null2 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@30 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@26 -- # for i in $(seq 1 4) 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@27 -- # rpc_cmd bdev_null_create Null3 102400 512 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:03.933 Null3 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode3 -a -s SPDK00000000000003 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.933 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode3 Null3 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@30 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode3 -t tcp -a 10.0.0.2 -s 4420 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@26 -- # for i in $(seq 1 4) 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@27 -- # rpc_cmd bdev_null_create Null4 102400 512 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:03.934 Null4 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode4 -a -s SPDK00000000000004 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode4 Null4 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@30 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode4 -t tcp -a 10.0.0.2 -s 4420 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@32 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@35 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 10.0.0.2 -s 4430 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:03.934 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@37 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -a 10.0.0.2 -s 4420 00:15:04.192 00:15:04.192 Discovery Log Number of Records 6, Generation counter 6 00:15:04.192 =====Discovery Log Entry 0====== 00:15:04.192 trtype: tcp 00:15:04.192 adrfam: ipv4 00:15:04.192 subtype: current discovery subsystem 00:15:04.192 treq: not required 00:15:04.192 portid: 0 00:15:04.192 trsvcid: 4420 00:15:04.192 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:15:04.192 traddr: 10.0.0.2 00:15:04.192 eflags: explicit discovery connections, duplicate discovery information 00:15:04.192 sectype: none 00:15:04.192 =====Discovery Log Entry 1====== 00:15:04.192 trtype: tcp 00:15:04.192 adrfam: ipv4 00:15:04.192 subtype: nvme subsystem 00:15:04.192 treq: not required 00:15:04.192 portid: 0 00:15:04.192 trsvcid: 4420 00:15:04.192 subnqn: nqn.2016-06.io.spdk:cnode1 00:15:04.192 traddr: 10.0.0.2 00:15:04.192 eflags: none 00:15:04.192 sectype: none 00:15:04.192 =====Discovery Log Entry 2====== 00:15:04.192 trtype: tcp 00:15:04.192 adrfam: ipv4 00:15:04.192 subtype: nvme subsystem 00:15:04.192 treq: not required 00:15:04.192 portid: 0 00:15:04.192 trsvcid: 4420 00:15:04.192 subnqn: nqn.2016-06.io.spdk:cnode2 00:15:04.192 traddr: 10.0.0.2 00:15:04.192 eflags: none 00:15:04.192 sectype: none 00:15:04.192 =====Discovery Log Entry 3====== 00:15:04.192 trtype: tcp 00:15:04.192 adrfam: ipv4 00:15:04.192 subtype: nvme subsystem 00:15:04.192 treq: not required 00:15:04.192 portid: 0 00:15:04.192 trsvcid: 4420 00:15:04.192 subnqn: nqn.2016-06.io.spdk:cnode3 00:15:04.192 traddr: 10.0.0.2 00:15:04.192 eflags: none 00:15:04.192 sectype: none 00:15:04.192 =====Discovery Log Entry 4====== 00:15:04.192 trtype: tcp 00:15:04.192 adrfam: ipv4 00:15:04.192 subtype: nvme subsystem 00:15:04.192 treq: not required 00:15:04.192 portid: 0 00:15:04.192 trsvcid: 4420 00:15:04.192 subnqn: nqn.2016-06.io.spdk:cnode4 00:15:04.192 traddr: 10.0.0.2 00:15:04.192 eflags: none 00:15:04.192 sectype: none 00:15:04.192 =====Discovery Log Entry 5====== 00:15:04.192 trtype: tcp 00:15:04.192 adrfam: ipv4 00:15:04.192 subtype: discovery subsystem referral 00:15:04.192 treq: not required 00:15:04.192 portid: 0 00:15:04.192 trsvcid: 4430 00:15:04.192 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:15:04.193 traddr: 10.0.0.2 00:15:04.193 eflags: none 00:15:04.193 sectype: none 00:15:04.193 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@39 -- # echo 'Perform nvmf subsystem discovery via RPC' 00:15:04.193 Perform nvmf subsystem discovery via RPC 00:15:04.193 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@40 -- # rpc_cmd nvmf_get_subsystems 00:15:04.193 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.193 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:04.193 [ 00:15:04.193 { 00:15:04.193 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:15:04.193 "subtype": "Discovery", 00:15:04.193 "listen_addresses": [ 00:15:04.193 { 00:15:04.193 "trtype": "TCP", 00:15:04.193 "adrfam": "IPv4", 00:15:04.193 "traddr": "10.0.0.2", 00:15:04.193 "trsvcid": "4420" 00:15:04.193 } 00:15:04.193 ], 00:15:04.193 "allow_any_host": true, 00:15:04.193 "hosts": [] 00:15:04.193 }, 00:15:04.193 { 00:15:04.193 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:15:04.193 "subtype": "NVMe", 00:15:04.193 "listen_addresses": [ 00:15:04.193 { 00:15:04.193 "trtype": "TCP", 00:15:04.193 "adrfam": "IPv4", 00:15:04.193 "traddr": "10.0.0.2", 00:15:04.193 "trsvcid": "4420" 00:15:04.193 } 00:15:04.193 ], 00:15:04.193 "allow_any_host": true, 00:15:04.193 "hosts": [], 00:15:04.193 "serial_number": "SPDK00000000000001", 00:15:04.193 "model_number": "SPDK bdev Controller", 00:15:04.193 "max_namespaces": 32, 00:15:04.193 "min_cntlid": 1, 00:15:04.193 "max_cntlid": 65519, 00:15:04.193 "namespaces": [ 00:15:04.193 { 00:15:04.193 "nsid": 1, 00:15:04.193 "bdev_name": "Null1", 00:15:04.193 "name": "Null1", 00:15:04.193 "nguid": "3320E2A84E814D239780E44B2A5BD1B0", 00:15:04.193 "uuid": "3320e2a8-4e81-4d23-9780-e44b2a5bd1b0" 00:15:04.193 } 00:15:04.193 ] 00:15:04.193 }, 00:15:04.193 { 00:15:04.193 "nqn": "nqn.2016-06.io.spdk:cnode2", 00:15:04.193 "subtype": "NVMe", 00:15:04.193 "listen_addresses": [ 00:15:04.193 { 00:15:04.193 "trtype": "TCP", 00:15:04.193 "adrfam": "IPv4", 00:15:04.193 "traddr": "10.0.0.2", 00:15:04.193 "trsvcid": "4420" 00:15:04.193 } 00:15:04.193 ], 00:15:04.193 "allow_any_host": true, 00:15:04.193 "hosts": [], 00:15:04.193 "serial_number": "SPDK00000000000002", 00:15:04.193 "model_number": "SPDK bdev Controller", 00:15:04.193 "max_namespaces": 32, 00:15:04.193 "min_cntlid": 1, 00:15:04.193 "max_cntlid": 65519, 00:15:04.193 "namespaces": [ 00:15:04.193 { 00:15:04.193 "nsid": 1, 00:15:04.193 "bdev_name": "Null2", 00:15:04.193 "name": "Null2", 00:15:04.193 "nguid": "3A7EE4369E0443E3A6955B2FBB7E0023", 00:15:04.193 "uuid": "3a7ee436-9e04-43e3-a695-5b2fbb7e0023" 00:15:04.193 } 00:15:04.193 ] 00:15:04.193 }, 00:15:04.193 { 00:15:04.193 "nqn": "nqn.2016-06.io.spdk:cnode3", 00:15:04.193 "subtype": "NVMe", 00:15:04.193 "listen_addresses": [ 00:15:04.193 { 00:15:04.193 "trtype": "TCP", 00:15:04.193 "adrfam": "IPv4", 00:15:04.193 "traddr": "10.0.0.2", 00:15:04.193 "trsvcid": "4420" 00:15:04.193 } 00:15:04.193 ], 00:15:04.193 "allow_any_host": true, 00:15:04.193 "hosts": [], 00:15:04.193 "serial_number": "SPDK00000000000003", 00:15:04.193 "model_number": "SPDK bdev Controller", 00:15:04.193 "max_namespaces": 32, 00:15:04.193 "min_cntlid": 1, 00:15:04.193 "max_cntlid": 65519, 00:15:04.193 "namespaces": [ 00:15:04.193 { 00:15:04.193 "nsid": 1, 00:15:04.193 "bdev_name": "Null3", 00:15:04.193 "name": "Null3", 00:15:04.193 "nguid": "3CD7C2DABC784DA785587911BCBBCD07", 00:15:04.193 "uuid": "3cd7c2da-bc78-4da7-8558-7911bcbbcd07" 00:15:04.193 } 00:15:04.193 ] 00:15:04.193 }, 00:15:04.193 { 00:15:04.193 "nqn": "nqn.2016-06.io.spdk:cnode4", 00:15:04.193 "subtype": "NVMe", 00:15:04.193 "listen_addresses": [ 00:15:04.193 { 00:15:04.193 "trtype": "TCP", 00:15:04.193 "adrfam": "IPv4", 00:15:04.193 "traddr": "10.0.0.2", 00:15:04.193 "trsvcid": "4420" 00:15:04.193 } 00:15:04.193 ], 00:15:04.193 "allow_any_host": true, 00:15:04.193 "hosts": [], 00:15:04.193 "serial_number": "SPDK00000000000004", 00:15:04.193 "model_number": "SPDK bdev Controller", 00:15:04.193 "max_namespaces": 32, 00:15:04.193 "min_cntlid": 1, 00:15:04.193 "max_cntlid": 65519, 00:15:04.193 "namespaces": [ 00:15:04.193 { 00:15:04.193 "nsid": 1, 00:15:04.193 "bdev_name": "Null4", 00:15:04.193 "name": "Null4", 00:15:04.193 "nguid": "E5225613A07A4DB7B6DD4ABC3520ACF9", 00:15:04.193 "uuid": "e5225613-a07a-4db7-b6dd-4abc3520acf9" 00:15:04.193 } 00:15:04.193 ] 00:15:04.193 } 00:15:04.193 ] 00:15:04.193 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.193 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@42 -- # seq 1 4 00:15:04.193 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@42 -- # for i in $(seq 1 4) 00:15:04.193 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@43 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:15:04.193 17:30:45 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.193 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:04.193 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.193 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@44 -- # rpc_cmd bdev_null_delete Null1 00:15:04.193 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.193 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:04.193 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.193 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@42 -- # for i in $(seq 1 4) 00:15:04.193 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@43 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode2 00:15:04.193 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.193 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:04.193 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.193 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@44 -- # rpc_cmd bdev_null_delete Null2 00:15:04.193 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.193 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@42 -- # for i in $(seq 1 4) 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@43 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode3 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@44 -- # rpc_cmd bdev_null_delete Null3 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@42 -- # for i in $(seq 1 4) 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@43 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode4 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@44 -- # rpc_cmd bdev_null_delete Null4 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@47 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 10.0.0.2 -s 4430 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@49 -- # rpc_cmd bdev_get_bdevs 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@49 -- # jq -r '.[].name' 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@49 -- # check_bdevs= 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@50 -- # '[' -n '' ']' 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@55 -- # trap - SIGINT SIGTERM EXIT 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@57 -- # nvmftestfini 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@516 -- # nvmfcleanup 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@121 -- # sync 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@124 -- # set +e 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@125 -- # for i in {1..20} 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:15:04.452 rmmod nvme_tcp 00:15:04.452 rmmod nvme_fabrics 00:15:04.452 rmmod nvme_keyring 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@128 -- # set -e 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@129 -- # return 0 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@517 -- # '[' -n 182680 ']' 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@518 -- # killprocess 182680 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@954 -- # '[' -z 182680 ']' 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@958 -- # kill -0 182680 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@959 -- # uname 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 182680 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@972 -- # echo 'killing process with pid 182680' 00:15:04.452 killing process with pid 182680 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@973 -- # kill 182680 00:15:04.452 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@978 -- # wait 182680 00:15:04.711 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:15:04.711 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:15:04.711 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:15:04.711 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@297 -- # iptr 00:15:04.711 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@791 -- # iptables-save 00:15:04.711 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:15:04.711 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@791 -- # iptables-restore 00:15:04.711 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:15:04.711 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@302 -- # remove_spdk_ns 00:15:04.711 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:15:04.711 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:15:04.711 17:30:46 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:15:06.630 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:15:06.630 00:15:06.630 real 0m5.797s 00:15:06.630 user 0m4.942s 00:15:06.630 sys 0m1.998s 00:15:06.630 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:06.630 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:06.891 ************************************ 00:15:06.891 END TEST nvmf_target_discovery 00:15:06.891 ************************************ 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@19 -- # run_test nvmf_referrals /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/referrals.sh --transport=tcp 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:15:06.891 ************************************ 00:15:06.891 START TEST nvmf_referrals 00:15:06.891 ************************************ 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/referrals.sh --transport=tcp 00:15:06.891 * Looking for test storage... 00:15:06.891 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1711 -- # lcov --version 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@333 -- # local ver1 ver1_l 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@334 -- # local ver2 ver2_l 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@336 -- # IFS=.-: 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@336 -- # read -ra ver1 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@337 -- # IFS=.-: 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@337 -- # read -ra ver2 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@338 -- # local 'op=<' 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@340 -- # ver1_l=2 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@341 -- # ver2_l=1 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@344 -- # case "$op" in 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@345 -- # : 1 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@364 -- # (( v = 0 )) 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@365 -- # decimal 1 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@353 -- # local d=1 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@355 -- # echo 1 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@365 -- # ver1[v]=1 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@366 -- # decimal 2 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@353 -- # local d=2 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@355 -- # echo 2 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@366 -- # ver2[v]=2 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@368 -- # return 0 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:15:06.891 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:06.891 --rc genhtml_branch_coverage=1 00:15:06.891 --rc genhtml_function_coverage=1 00:15:06.891 --rc genhtml_legend=1 00:15:06.891 --rc geninfo_all_blocks=1 00:15:06.891 --rc geninfo_unexecuted_blocks=1 00:15:06.891 00:15:06.891 ' 00:15:06.891 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:15:06.891 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:06.891 --rc genhtml_branch_coverage=1 00:15:06.892 --rc genhtml_function_coverage=1 00:15:06.892 --rc genhtml_legend=1 00:15:06.892 --rc geninfo_all_blocks=1 00:15:06.892 --rc geninfo_unexecuted_blocks=1 00:15:06.892 00:15:06.892 ' 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:15:06.892 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:06.892 --rc genhtml_branch_coverage=1 00:15:06.892 --rc genhtml_function_coverage=1 00:15:06.892 --rc genhtml_legend=1 00:15:06.892 --rc geninfo_all_blocks=1 00:15:06.892 --rc geninfo_unexecuted_blocks=1 00:15:06.892 00:15:06.892 ' 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:15:06.892 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:06.892 --rc genhtml_branch_coverage=1 00:15:06.892 --rc genhtml_function_coverage=1 00:15:06.892 --rc genhtml_legend=1 00:15:06.892 --rc geninfo_all_blocks=1 00:15:06.892 --rc geninfo_unexecuted_blocks=1 00:15:06.892 00:15:06.892 ' 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@7 -- # uname -s 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@15 -- # shopt -s extglob 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- paths/export.sh@5 -- # export PATH 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@51 -- # : 0 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:15:06.892 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@55 -- # have_pci_nics=0 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@11 -- # NVMF_REFERRAL_IP_1=127.0.0.2 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@12 -- # NVMF_REFERRAL_IP_2=127.0.0.3 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@13 -- # NVMF_REFERRAL_IP_3=127.0.0.4 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@14 -- # NVMF_PORT_REFERRAL=4430 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@15 -- # DISCOVERY_NQN=nqn.2014-08.org.nvmexpress.discovery 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@16 -- # NQN=nqn.2016-06.io.spdk:cnode1 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@37 -- # nvmftestinit 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@476 -- # prepare_net_devs 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@438 -- # local -g is_hw=no 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@440 -- # remove_spdk_ns 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@309 -- # xtrace_disable 00:15:06.892 17:30:48 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@315 -- # pci_devs=() 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@315 -- # local -a pci_devs 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@316 -- # pci_net_devs=() 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@317 -- # pci_drivers=() 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@317 -- # local -A pci_drivers 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@319 -- # net_devs=() 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@319 -- # local -ga net_devs 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@320 -- # e810=() 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@320 -- # local -ga e810 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@321 -- # x722=() 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@321 -- # local -ga x722 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@322 -- # mlx=() 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@322 -- # local -ga mlx 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:15:09.434 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:15:09.434 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:09.434 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@418 -- # [[ up == up ]] 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:15:09.435 Found net devices under 0000:0a:00.0: cvl_0_0 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@418 -- # [[ up == up ]] 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:15:09.435 Found net devices under 0000:0a:00.1: cvl_0_1 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@442 -- # is_hw=yes 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:15:09.435 17:30:50 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:15:09.435 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:15:09.435 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:15:09.435 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:15:09.435 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:15:09.435 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:15:09.435 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.348 ms 00:15:09.435 00:15:09.435 --- 10.0.0.2 ping statistics --- 00:15:09.435 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:09.435 rtt min/avg/max/mdev = 0.348/0.348/0.348/0.000 ms 00:15:09.435 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:15:09.435 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:15:09.435 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.121 ms 00:15:09.435 00:15:09.435 --- 10.0.0.1 ping statistics --- 00:15:09.435 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:09.435 rtt min/avg/max/mdev = 0.121/0.121/0.121/0.000 ms 00:15:09.435 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:15:09.435 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@450 -- # return 0 00:15:09.435 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:15:09.435 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:15:09.435 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:15:09.435 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:15:09.435 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:15:09.435 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:15:09.435 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:15:09.435 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@38 -- # nvmfappstart -m 0xF 00:15:09.435 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:15:09.435 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@726 -- # xtrace_disable 00:15:09.435 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:09.435 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@509 -- # nvmfpid=184777 00:15:09.435 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:15:09.435 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@510 -- # waitforlisten 184777 00:15:09.435 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@835 -- # '[' -z 184777 ']' 00:15:09.435 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:09.435 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:09.435 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:09.435 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:09.435 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:09.435 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:09.435 [2024-12-06 17:30:51.099028] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:15:09.435 [2024-12-06 17:30:51.099099] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:09.435 [2024-12-06 17:30:51.170347] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:15:09.435 [2024-12-06 17:30:51.215978] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:15:09.435 [2024-12-06 17:30:51.216027] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:15:09.435 [2024-12-06 17:30:51.216051] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:15:09.435 [2024-12-06 17:30:51.216063] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:15:09.435 [2024-12-06 17:30:51.216073] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:15:09.435 [2024-12-06 17:30:51.217561] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:15:09.435 [2024-12-06 17:30:51.217625] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:15:09.435 [2024-12-06 17:30:51.217768] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:15:09.436 [2024-12-06 17:30:51.217772] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@868 -- # return 0 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@732 -- # xtrace_disable 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@40 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:09.695 [2024-12-06 17:30:51.362127] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@41 -- # rpc_cmd nvmf_subsystem_add_listener -t tcp -a 10.0.0.2 -s 8009 discovery 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:09.695 [2024-12-06 17:30:51.386894] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 8009 *** 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@44 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.2 -s 4430 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@45 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.3 -s 4430 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@46 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.4 -s 4430 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@48 -- # rpc_cmd nvmf_discovery_get_referrals 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@48 -- # jq length 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@48 -- # (( 3 == 3 )) 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@49 -- # get_referral_ips rpc 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ rpc == \r\p\c ]] 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # rpc_cmd nvmf_discovery_get_referrals 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # jq -r '.[].address.traddr' 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # sort 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # echo 127.0.0.2 127.0.0.3 127.0.0.4 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@49 -- # [[ 127.0.0.2 127.0.0.3 127.0.0.4 == \1\2\7\.\0\.\0\.\2\ \1\2\7\.\0\.\0\.\3\ \1\2\7\.\0\.\0\.\4 ]] 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@50 -- # get_referral_ips nvme 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:15:09.695 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -a 10.0.0.2 -s 8009 -o json 00:15:09.696 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:15:09.696 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # sort 00:15:09.955 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # echo 127.0.0.2 127.0.0.3 127.0.0.4 00:15:09.955 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@50 -- # [[ 127.0.0.2 127.0.0.3 127.0.0.4 == \1\2\7\.\0\.\0\.\2\ \1\2\7\.\0\.\0\.\3\ \1\2\7\.\0\.\0\.\4 ]] 00:15:09.955 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@52 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.2 -s 4430 00:15:09.955 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.955 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:09.955 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.955 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@53 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.3 -s 4430 00:15:09.956 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.956 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:09.956 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.956 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@54 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.4 -s 4430 00:15:09.956 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.956 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:09.956 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.956 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@56 -- # rpc_cmd nvmf_discovery_get_referrals 00:15:09.956 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@56 -- # jq length 00:15:09.956 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:09.956 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:09.956 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:09.956 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@56 -- # (( 0 == 0 )) 00:15:09.956 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@57 -- # get_referral_ips nvme 00:15:09.956 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:15:09.956 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:15:09.956 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -a 10.0.0.2 -s 8009 -o json 00:15:09.956 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:15:09.956 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # sort 00:15:10.215 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # echo 00:15:10.215 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@57 -- # [[ '' == '' ]] 00:15:10.215 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@60 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.2 -s 4430 -n discovery 00:15:10.215 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:10.215 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:10.215 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:10.215 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@62 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.2 -s 4430 -n nqn.2016-06.io.spdk:cnode1 00:15:10.215 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:10.215 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:10.215 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:10.215 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@65 -- # get_referral_ips rpc 00:15:10.215 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ rpc == \r\p\c ]] 00:15:10.215 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # rpc_cmd nvmf_discovery_get_referrals 00:15:10.215 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # jq -r '.[].address.traddr' 00:15:10.215 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:10.215 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # sort 00:15:10.215 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:10.215 17:30:51 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:10.215 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # echo 127.0.0.2 127.0.0.2 00:15:10.215 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@65 -- # [[ 127.0.0.2 127.0.0.2 == \1\2\7\.\0\.\0\.\2\ \1\2\7\.\0\.\0\.\2 ]] 00:15:10.215 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@66 -- # get_referral_ips nvme 00:15:10.215 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:15:10.215 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:15:10.215 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -a 10.0.0.2 -s 8009 -o json 00:15:10.215 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:15:10.215 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # sort 00:15:10.474 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # echo 127.0.0.2 127.0.0.2 00:15:10.474 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@66 -- # [[ 127.0.0.2 127.0.0.2 == \1\2\7\.\0\.\0\.\2\ \1\2\7\.\0\.\0\.\2 ]] 00:15:10.474 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@67 -- # get_discovery_entries 'nvme subsystem' 00:15:10.474 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@67 -- # jq -r .subnqn 00:15:10.474 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@31 -- # local 'subtype=nvme subsystem' 00:15:10.474 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@33 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -a 10.0.0.2 -s 8009 -o json 00:15:10.474 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@34 -- # jq '.records[] | select(.subtype == "nvme subsystem")' 00:15:10.733 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@67 -- # [[ nqn.2016-06.io.spdk:cnode1 == \n\q\n\.\2\0\1\6\-\0\6\.\i\o\.\s\p\d\k\:\c\n\o\d\e\1 ]] 00:15:10.733 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@68 -- # get_discovery_entries 'discovery subsystem referral' 00:15:10.733 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@68 -- # jq -r .subnqn 00:15:10.733 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@31 -- # local 'subtype=discovery subsystem referral' 00:15:10.733 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@33 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -a 10.0.0.2 -s 8009 -o json 00:15:10.733 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@34 -- # jq '.records[] | select(.subtype == "discovery subsystem referral")' 00:15:10.733 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@68 -- # [[ nqn.2014-08.org.nvmexpress.discovery == \n\q\n\.\2\0\1\4\-\0\8\.\o\r\g\.\n\v\m\e\x\p\r\e\s\s\.\d\i\s\c\o\v\e\r\y ]] 00:15:10.733 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@71 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.2 -s 4430 -n nqn.2016-06.io.spdk:cnode1 00:15:10.733 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:10.733 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:10.733 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:10.733 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@73 -- # get_referral_ips rpc 00:15:10.733 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ rpc == \r\p\c ]] 00:15:10.733 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # rpc_cmd nvmf_discovery_get_referrals 00:15:10.733 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # jq -r '.[].address.traddr' 00:15:10.733 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:10.733 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:10.733 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # sort 00:15:10.733 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:10.992 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # echo 127.0.0.2 00:15:10.992 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@73 -- # [[ 127.0.0.2 == \1\2\7\.\0\.\0\.\2 ]] 00:15:10.992 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@74 -- # get_referral_ips nvme 00:15:10.992 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:15:10.992 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:15:10.992 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -a 10.0.0.2 -s 8009 -o json 00:15:10.992 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:15:10.992 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # sort 00:15:10.992 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # echo 127.0.0.2 00:15:10.992 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@74 -- # [[ 127.0.0.2 == \1\2\7\.\0\.\0\.\2 ]] 00:15:10.992 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@75 -- # get_discovery_entries 'nvme subsystem' 00:15:10.992 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@75 -- # jq -r .subnqn 00:15:10.992 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@31 -- # local 'subtype=nvme subsystem' 00:15:10.992 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@33 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -a 10.0.0.2 -s 8009 -o json 00:15:10.992 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@34 -- # jq '.records[] | select(.subtype == "nvme subsystem")' 00:15:11.250 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@75 -- # [[ '' == '' ]] 00:15:11.250 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@76 -- # get_discovery_entries 'discovery subsystem referral' 00:15:11.250 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@76 -- # jq -r .subnqn 00:15:11.250 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@31 -- # local 'subtype=discovery subsystem referral' 00:15:11.250 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@33 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -a 10.0.0.2 -s 8009 -o json 00:15:11.250 17:30:52 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@34 -- # jq '.records[] | select(.subtype == "discovery subsystem referral")' 00:15:11.508 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@76 -- # [[ nqn.2014-08.org.nvmexpress.discovery == \n\q\n\.\2\0\1\4\-\0\8\.\o\r\g\.\n\v\m\e\x\p\r\e\s\s\.\d\i\s\c\o\v\e\r\y ]] 00:15:11.508 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@79 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.2 -s 4430 -n nqn.2014-08.org.nvmexpress.discovery 00:15:11.508 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:11.508 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:11.508 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:11.508 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@82 -- # rpc_cmd nvmf_discovery_get_referrals 00:15:11.508 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:11.508 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@82 -- # jq length 00:15:11.508 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:11.508 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:11.508 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@82 -- # (( 0 == 0 )) 00:15:11.508 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@83 -- # get_referral_ips nvme 00:15:11.508 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:15:11.508 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:15:11.509 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -a 10.0.0.2 -s 8009 -o json 00:15:11.509 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:15:11.509 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # sort 00:15:11.768 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # echo 00:15:11.768 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@83 -- # [[ '' == '' ]] 00:15:11.768 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@85 -- # trap - SIGINT SIGTERM EXIT 00:15:11.768 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@86 -- # nvmftestfini 00:15:11.768 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@516 -- # nvmfcleanup 00:15:11.768 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@121 -- # sync 00:15:11.768 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:15:11.768 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@124 -- # set +e 00:15:11.768 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@125 -- # for i in {1..20} 00:15:11.768 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:15:11.768 rmmod nvme_tcp 00:15:11.768 rmmod nvme_fabrics 00:15:11.768 rmmod nvme_keyring 00:15:11.768 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:15:11.768 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@128 -- # set -e 00:15:11.768 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@129 -- # return 0 00:15:11.768 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@517 -- # '[' -n 184777 ']' 00:15:11.768 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@518 -- # killprocess 184777 00:15:11.768 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@954 -- # '[' -z 184777 ']' 00:15:11.768 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@958 -- # kill -0 184777 00:15:11.768 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@959 -- # uname 00:15:11.768 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:15:11.768 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 184777 00:15:11.768 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:15:11.768 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:15:11.768 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@972 -- # echo 'killing process with pid 184777' 00:15:11.768 killing process with pid 184777 00:15:11.768 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@973 -- # kill 184777 00:15:11.768 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@978 -- # wait 184777 00:15:12.028 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:15:12.028 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:15:12.028 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:15:12.028 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@297 -- # iptr 00:15:12.028 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@791 -- # iptables-save 00:15:12.028 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@791 -- # iptables-restore 00:15:12.028 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:15:12.028 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:15:12.028 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@302 -- # remove_spdk_ns 00:15:12.028 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:15:12.028 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:15:12.028 17:30:53 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:15:13.937 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:15:13.937 00:15:13.937 real 0m7.245s 00:15:13.937 user 0m11.498s 00:15:13.937 sys 0m2.412s 00:15:13.937 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1130 -- # xtrace_disable 00:15:13.937 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:13.937 ************************************ 00:15:13.937 END TEST nvmf_referrals 00:15:13.937 ************************************ 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@20 -- # run_test nvmf_connect_disconnect /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_disconnect.sh --transport=tcp 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:15:14.197 ************************************ 00:15:14.197 START TEST nvmf_connect_disconnect 00:15:14.197 ************************************ 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_disconnect.sh --transport=tcp 00:15:14.197 * Looking for test storage... 00:15:14.197 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1711 -- # lcov --version 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@333 -- # local ver1 ver1_l 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@334 -- # local ver2 ver2_l 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@336 -- # IFS=.-: 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@336 -- # read -ra ver1 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@337 -- # IFS=.-: 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@337 -- # read -ra ver2 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@338 -- # local 'op=<' 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@340 -- # ver1_l=2 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@341 -- # ver2_l=1 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@344 -- # case "$op" in 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@345 -- # : 1 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@364 -- # (( v = 0 )) 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@365 -- # decimal 1 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@353 -- # local d=1 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@355 -- # echo 1 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@365 -- # ver1[v]=1 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@366 -- # decimal 2 00:15:14.197 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@353 -- # local d=2 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@355 -- # echo 2 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@366 -- # ver2[v]=2 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@368 -- # return 0 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:15:14.198 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:14.198 --rc genhtml_branch_coverage=1 00:15:14.198 --rc genhtml_function_coverage=1 00:15:14.198 --rc genhtml_legend=1 00:15:14.198 --rc geninfo_all_blocks=1 00:15:14.198 --rc geninfo_unexecuted_blocks=1 00:15:14.198 00:15:14.198 ' 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:15:14.198 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:14.198 --rc genhtml_branch_coverage=1 00:15:14.198 --rc genhtml_function_coverage=1 00:15:14.198 --rc genhtml_legend=1 00:15:14.198 --rc geninfo_all_blocks=1 00:15:14.198 --rc geninfo_unexecuted_blocks=1 00:15:14.198 00:15:14.198 ' 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:15:14.198 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:14.198 --rc genhtml_branch_coverage=1 00:15:14.198 --rc genhtml_function_coverage=1 00:15:14.198 --rc genhtml_legend=1 00:15:14.198 --rc geninfo_all_blocks=1 00:15:14.198 --rc geninfo_unexecuted_blocks=1 00:15:14.198 00:15:14.198 ' 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:15:14.198 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:14.198 --rc genhtml_branch_coverage=1 00:15:14.198 --rc genhtml_function_coverage=1 00:15:14.198 --rc genhtml_legend=1 00:15:14.198 --rc geninfo_all_blocks=1 00:15:14.198 --rc geninfo_unexecuted_blocks=1 00:15:14.198 00:15:14.198 ' 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@7 -- # uname -s 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@15 -- # shopt -s extglob 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- paths/export.sh@5 -- # export PATH 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@51 -- # : 0 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:15:14.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@55 -- # have_pci_nics=0 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@11 -- # MALLOC_BDEV_SIZE=64 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@15 -- # nvmftestinit 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@476 -- # prepare_net_devs 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@438 -- # local -g is_hw=no 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@440 -- # remove_spdk_ns 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@309 -- # xtrace_disable 00:15:14.198 17:30:55 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@315 -- # pci_devs=() 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@315 -- # local -a pci_devs 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@316 -- # pci_net_devs=() 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@317 -- # pci_drivers=() 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@317 -- # local -A pci_drivers 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@319 -- # net_devs=() 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@319 -- # local -ga net_devs 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@320 -- # e810=() 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@320 -- # local -ga e810 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@321 -- # x722=() 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@321 -- # local -ga x722 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@322 -- # mlx=() 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@322 -- # local -ga mlx 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:15:16.731 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:15:16.731 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@418 -- # [[ up == up ]] 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:15:16.731 Found net devices under 0000:0a:00.0: cvl_0_0 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@418 -- # [[ up == up ]] 00:15:16.731 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:15:16.732 Found net devices under 0000:0a:00.1: cvl_0_1 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@442 -- # is_hw=yes 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:15:16.732 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:15:16.732 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.158 ms 00:15:16.732 00:15:16.732 --- 10.0.0.2 ping statistics --- 00:15:16.732 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:16.732 rtt min/avg/max/mdev = 0.158/0.158/0.158/0.000 ms 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:15:16.732 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:15:16.732 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.098 ms 00:15:16.732 00:15:16.732 --- 10.0.0.1 ping statistics --- 00:15:16.732 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:16.732 rtt min/avg/max/mdev = 0.098/0.098/0.098/0.000 ms 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@450 -- # return 0 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@16 -- # nvmfappstart -m 0xF 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@726 -- # xtrace_disable 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@509 -- # nvmfpid=187202 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@510 -- # waitforlisten 187202 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@835 -- # '[' -z 187202 ']' 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@840 -- # local max_retries=100 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:16.732 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@844 -- # xtrace_disable 00:15:16.732 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:15:16.732 [2024-12-06 17:30:58.426135] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:15:16.732 [2024-12-06 17:30:58.426220] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:16.732 [2024-12-06 17:30:58.497466] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:15:16.732 [2024-12-06 17:30:58.541091] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:15:16.732 [2024-12-06 17:30:58.541150] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:15:16.732 [2024-12-06 17:30:58.541177] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:15:16.732 [2024-12-06 17:30:58.541188] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:15:16.732 [2024-12-06 17:30:58.541198] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:15:16.732 [2024-12-06 17:30:58.542753] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:15:16.732 [2024-12-06 17:30:58.542814] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:15:16.732 [2024-12-06 17:30:58.542877] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:15:16.732 [2024-12-06 17:30:58.542880] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@868 -- # return 0 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@732 -- # xtrace_disable 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -c 0 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:15:16.990 [2024-12-06 17:30:58.688507] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@20 -- # rpc_cmd bdev_malloc_create 64 512 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@20 -- # bdev=Malloc0 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@563 -- # xtrace_disable 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:15:16.990 [2024-12-06 17:30:58.758672] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@26 -- # '[' 1 -eq 1 ']' 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@27 -- # num_iterations=100 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@29 -- # NVME_CONNECT='nvme connect -i 8' 00:15:16.990 17:30:58 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@34 -- # set +x 00:15:19.516 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:15:22.043 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:15:23.942 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:15:26.476 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:15:28.418 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:15:30.968 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:15:33.496 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:15:35.396 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:15:37.926 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:15:40.454 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:15:42.378 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:15:44.904 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:15:47.432 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:15:49.331 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:15:51.861 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:15:54.392 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:15:56.294 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:15:58.824 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:00.723 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:03.249 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:05.774 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:08.304 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:10.218 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:12.749 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:15.274 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:17.170 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:19.693 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:21.592 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:24.131 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:26.660 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:29.189 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:31.089 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:33.615 [2024-12-06 17:32:15.222597] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x219bb20 is same with the state(6) to be set 00:16:33.615 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:36.141 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:38.042 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:40.573 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:43.102 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:45.001 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:47.533 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:50.064 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:51.965 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:54.518 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:57.043 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:58.942 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:17:01.469 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:17:03.997 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:17:05.897 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:17:08.422 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:17:10.318 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:17:12.843 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:17:15.374 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:17:17.903 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:17:19.804 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:17:22.334 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:17:24.859 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:17:26.759 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:17:29.285 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:17:31.180 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:17:33.708 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:17:36.235 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:17:38.161 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:17:40.686 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:17:42.585 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:17:45.111 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:17:47.640 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:17:49.546 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:17:52.076 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:17:54.630 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:17:56.527 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:17:59.052 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:00.949 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:03.517 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:06.040 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:07.939 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:10.468 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:12.995 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:14.888 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:17.416 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:19.944 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:21.839 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:24.362 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:26.283 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:28.812 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:31.339 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:33.240 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:35.767 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:38.301 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:40.199 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:42.728 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:44.628 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:47.157 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:49.684 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:51.591 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:54.128 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:56.657 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:58.558 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:19:01.086 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:19:02.995 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:19:05.524 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:19:08.054 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:19:08.054 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@43 -- # trap - SIGINT SIGTERM EXIT 00:19:08.054 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@45 -- # nvmftestfini 00:19:08.054 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@516 -- # nvmfcleanup 00:19:08.054 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@121 -- # sync 00:19:08.054 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:19:08.054 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@124 -- # set +e 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@125 -- # for i in {1..20} 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:19:08.055 rmmod nvme_tcp 00:19:08.055 rmmod nvme_fabrics 00:19:08.055 rmmod nvme_keyring 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@128 -- # set -e 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@129 -- # return 0 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@517 -- # '[' -n 187202 ']' 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@518 -- # killprocess 187202 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@954 -- # '[' -z 187202 ']' 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@958 -- # kill -0 187202 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@959 -- # uname 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 187202 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@972 -- # echo 'killing process with pid 187202' 00:19:08.055 killing process with pid 187202 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@973 -- # kill 187202 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@978 -- # wait 187202 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@297 -- # iptr 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@791 -- # iptables-save 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@791 -- # iptables-restore 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@302 -- # remove_spdk_ns 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:19:08.055 17:34:49 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:19:10.593 17:34:51 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:19:10.594 00:19:10.594 real 3m56.008s 00:19:10.594 user 15m0.014s 00:19:10.594 sys 0m33.724s 00:19:10.594 17:34:51 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1130 -- # xtrace_disable 00:19:10.594 17:34:51 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:19:10.594 ************************************ 00:19:10.594 END TEST nvmf_connect_disconnect 00:19:10.594 ************************************ 00:19:10.594 17:34:51 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@21 -- # run_test nvmf_multitarget /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget.sh --transport=tcp 00:19:10.594 17:34:51 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:19:10.594 17:34:51 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:19:10.594 17:34:51 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:19:10.594 ************************************ 00:19:10.594 START TEST nvmf_multitarget 00:19:10.594 ************************************ 00:19:10.594 17:34:51 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget.sh --transport=tcp 00:19:10.594 * Looking for test storage... 00:19:10.594 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:19:10.594 17:34:51 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:19:10.594 17:34:51 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1711 -- # lcov --version 00:19:10.594 17:34:51 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@333 -- # local ver1 ver1_l 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@334 -- # local ver2 ver2_l 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@336 -- # IFS=.-: 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@336 -- # read -ra ver1 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@337 -- # IFS=.-: 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@337 -- # read -ra ver2 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@338 -- # local 'op=<' 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@340 -- # ver1_l=2 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@341 -- # ver2_l=1 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@344 -- # case "$op" in 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@345 -- # : 1 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@364 -- # (( v = 0 )) 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@365 -- # decimal 1 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@353 -- # local d=1 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@355 -- # echo 1 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@365 -- # ver1[v]=1 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@366 -- # decimal 2 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@353 -- # local d=2 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@355 -- # echo 2 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@366 -- # ver2[v]=2 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@368 -- # return 0 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:19:10.594 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:10.594 --rc genhtml_branch_coverage=1 00:19:10.594 --rc genhtml_function_coverage=1 00:19:10.594 --rc genhtml_legend=1 00:19:10.594 --rc geninfo_all_blocks=1 00:19:10.594 --rc geninfo_unexecuted_blocks=1 00:19:10.594 00:19:10.594 ' 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:19:10.594 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:10.594 --rc genhtml_branch_coverage=1 00:19:10.594 --rc genhtml_function_coverage=1 00:19:10.594 --rc genhtml_legend=1 00:19:10.594 --rc geninfo_all_blocks=1 00:19:10.594 --rc geninfo_unexecuted_blocks=1 00:19:10.594 00:19:10.594 ' 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:19:10.594 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:10.594 --rc genhtml_branch_coverage=1 00:19:10.594 --rc genhtml_function_coverage=1 00:19:10.594 --rc genhtml_legend=1 00:19:10.594 --rc geninfo_all_blocks=1 00:19:10.594 --rc geninfo_unexecuted_blocks=1 00:19:10.594 00:19:10.594 ' 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:19:10.594 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:10.594 --rc genhtml_branch_coverage=1 00:19:10.594 --rc genhtml_function_coverage=1 00:19:10.594 --rc genhtml_legend=1 00:19:10.594 --rc geninfo_all_blocks=1 00:19:10.594 --rc geninfo_unexecuted_blocks=1 00:19:10.594 00:19:10.594 ' 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@7 -- # uname -s 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@15 -- # shopt -s extglob 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:10.594 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- paths/export.sh@5 -- # export PATH 00:19:10.595 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:10.595 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@51 -- # : 0 00:19:10.595 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:19:10.595 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:19:10.595 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:19:10.595 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:19:10.595 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:19:10.595 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:19:10.595 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:19:10.595 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:19:10.595 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:19:10.595 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@55 -- # have_pci_nics=0 00:19:10.595 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@13 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py 00:19:10.595 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@15 -- # nvmftestinit 00:19:10.595 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:19:10.595 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:19:10.595 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@476 -- # prepare_net_devs 00:19:10.595 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@438 -- # local -g is_hw=no 00:19:10.595 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@440 -- # remove_spdk_ns 00:19:10.595 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:19:10.595 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:19:10.595 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:19:10.595 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:19:10.595 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:19:10.595 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@309 -- # xtrace_disable 00:19:10.595 17:34:52 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@10 -- # set +x 00:19:12.497 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:19:12.497 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@315 -- # pci_devs=() 00:19:12.497 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@315 -- # local -a pci_devs 00:19:12.497 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@316 -- # pci_net_devs=() 00:19:12.497 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:19:12.497 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@317 -- # pci_drivers=() 00:19:12.497 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@317 -- # local -A pci_drivers 00:19:12.497 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@319 -- # net_devs=() 00:19:12.497 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@319 -- # local -ga net_devs 00:19:12.497 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@320 -- # e810=() 00:19:12.497 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@320 -- # local -ga e810 00:19:12.497 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@321 -- # x722=() 00:19:12.497 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@321 -- # local -ga x722 00:19:12.497 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@322 -- # mlx=() 00:19:12.497 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@322 -- # local -ga mlx 00:19:12.497 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:19:12.498 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:19:12.498 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@418 -- # [[ up == up ]] 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:19:12.498 Found net devices under 0000:0a:00.0: cvl_0_0 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@418 -- # [[ up == up ]] 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:19:12.498 Found net devices under 0000:0a:00.1: cvl_0_1 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@442 -- # is_hw=yes 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:19:12.498 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:19:12.757 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:19:12.757 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.192 ms 00:19:12.757 00:19:12.757 --- 10.0.0.2 ping statistics --- 00:19:12.757 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:19:12.757 rtt min/avg/max/mdev = 0.192/0.192/0.192/0.000 ms 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:19:12.757 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:19:12.757 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.117 ms 00:19:12.757 00:19:12.757 --- 10.0.0.1 ping statistics --- 00:19:12.757 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:19:12.757 rtt min/avg/max/mdev = 0.117/0.117/0.117/0.000 ms 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@450 -- # return 0 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@16 -- # nvmfappstart -m 0xF 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@726 -- # xtrace_disable 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@10 -- # set +x 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@509 -- # nvmfpid=218206 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@510 -- # waitforlisten 218206 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@835 -- # '[' -z 218206 ']' 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@840 -- # local max_retries=100 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:12.757 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@844 -- # xtrace_disable 00:19:12.757 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@10 -- # set +x 00:19:12.757 [2024-12-06 17:34:54.509047] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:19:12.757 [2024-12-06 17:34:54.509140] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:12.757 [2024-12-06 17:34:54.580465] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:19:13.016 [2024-12-06 17:34:54.627618] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:13.016 [2024-12-06 17:34:54.627691] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:13.016 [2024-12-06 17:34:54.627706] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:13.016 [2024-12-06 17:34:54.627718] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:13.016 [2024-12-06 17:34:54.627728] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:13.016 [2024-12-06 17:34:54.629273] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:19:13.016 [2024-12-06 17:34:54.629333] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:19:13.016 [2024-12-06 17:34:54.629398] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:19:13.016 [2024-12-06 17:34:54.629401] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:19:13.016 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:19:13.016 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@868 -- # return 0 00:19:13.016 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:19:13.016 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@732 -- # xtrace_disable 00:19:13.016 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@10 -- # set +x 00:19:13.016 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:19:13.016 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@18 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini $1; exit 1' SIGINT SIGTERM EXIT 00:19:13.016 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_get_targets 00:19:13.016 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@21 -- # jq length 00:19:13.275 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@21 -- # '[' 1 '!=' 1 ']' 00:19:13.275 17:34:54 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_create_target -n nvmf_tgt_1 -s 32 00:19:13.275 "nvmf_tgt_1" 00:19:13.275 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_create_target -n nvmf_tgt_2 -s 32 00:19:13.536 "nvmf_tgt_2" 00:19:13.536 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_get_targets 00:19:13.536 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@28 -- # jq length 00:19:13.536 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@28 -- # '[' 3 '!=' 3 ']' 00:19:13.536 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_delete_target -n nvmf_tgt_1 00:19:13.794 true 00:19:13.794 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_delete_target -n nvmf_tgt_2 00:19:13.794 true 00:19:13.794 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_get_targets 00:19:13.794 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@35 -- # jq length 00:19:13.794 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@35 -- # '[' 1 '!=' 1 ']' 00:19:13.794 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@39 -- # trap - SIGINT SIGTERM EXIT 00:19:13.794 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@41 -- # nvmftestfini 00:19:13.794 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@516 -- # nvmfcleanup 00:19:13.794 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@121 -- # sync 00:19:13.794 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:19:13.794 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@124 -- # set +e 00:19:13.794 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@125 -- # for i in {1..20} 00:19:13.794 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:19:13.794 rmmod nvme_tcp 00:19:14.052 rmmod nvme_fabrics 00:19:14.052 rmmod nvme_keyring 00:19:14.052 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:19:14.052 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@128 -- # set -e 00:19:14.052 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@129 -- # return 0 00:19:14.052 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@517 -- # '[' -n 218206 ']' 00:19:14.052 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@518 -- # killprocess 218206 00:19:14.052 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@954 -- # '[' -z 218206 ']' 00:19:14.052 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@958 -- # kill -0 218206 00:19:14.052 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@959 -- # uname 00:19:14.052 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:19:14.052 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 218206 00:19:14.052 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:19:14.052 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:19:14.052 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@972 -- # echo 'killing process with pid 218206' 00:19:14.052 killing process with pid 218206 00:19:14.052 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@973 -- # kill 218206 00:19:14.052 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@978 -- # wait 218206 00:19:14.309 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:19:14.309 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:19:14.309 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:19:14.309 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@297 -- # iptr 00:19:14.309 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@791 -- # iptables-save 00:19:14.309 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:19:14.309 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@791 -- # iptables-restore 00:19:14.309 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:19:14.309 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@302 -- # remove_spdk_ns 00:19:14.309 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:19:14.309 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:19:14.309 17:34:55 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:19:16.220 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:19:16.220 00:19:16.220 real 0m6.096s 00:19:16.220 user 0m6.936s 00:19:16.220 sys 0m2.135s 00:19:16.220 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1130 -- # xtrace_disable 00:19:16.220 17:34:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@10 -- # set +x 00:19:16.220 ************************************ 00:19:16.220 END TEST nvmf_multitarget 00:19:16.220 ************************************ 00:19:16.220 17:34:58 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@22 -- # run_test nvmf_rpc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.sh --transport=tcp 00:19:16.220 17:34:58 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:19:16.220 17:34:58 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:19:16.220 17:34:58 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:19:16.220 ************************************ 00:19:16.220 START TEST nvmf_rpc 00:19:16.220 ************************************ 00:19:16.220 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.sh --transport=tcp 00:19:16.479 * Looking for test storage... 00:19:16.479 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1711 -- # lcov --version 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@344 -- # case "$op" in 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@345 -- # : 1 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@365 -- # decimal 1 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@353 -- # local d=1 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@355 -- # echo 1 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@366 -- # decimal 2 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@353 -- # local d=2 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@355 -- # echo 2 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@368 -- # return 0 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:19:16.479 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:16.479 --rc genhtml_branch_coverage=1 00:19:16.479 --rc genhtml_function_coverage=1 00:19:16.479 --rc genhtml_legend=1 00:19:16.479 --rc geninfo_all_blocks=1 00:19:16.479 --rc geninfo_unexecuted_blocks=1 00:19:16.479 00:19:16.479 ' 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:19:16.479 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:16.479 --rc genhtml_branch_coverage=1 00:19:16.479 --rc genhtml_function_coverage=1 00:19:16.479 --rc genhtml_legend=1 00:19:16.479 --rc geninfo_all_blocks=1 00:19:16.479 --rc geninfo_unexecuted_blocks=1 00:19:16.479 00:19:16.479 ' 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:19:16.479 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:16.479 --rc genhtml_branch_coverage=1 00:19:16.479 --rc genhtml_function_coverage=1 00:19:16.479 --rc genhtml_legend=1 00:19:16.479 --rc geninfo_all_blocks=1 00:19:16.479 --rc geninfo_unexecuted_blocks=1 00:19:16.479 00:19:16.479 ' 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:19:16.479 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:16.479 --rc genhtml_branch_coverage=1 00:19:16.479 --rc genhtml_function_coverage=1 00:19:16.479 --rc genhtml_legend=1 00:19:16.479 --rc geninfo_all_blocks=1 00:19:16.479 --rc geninfo_unexecuted_blocks=1 00:19:16.479 00:19:16.479 ' 00:19:16.479 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@7 -- # uname -s 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@15 -- # shopt -s extglob 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- paths/export.sh@5 -- # export PATH 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@51 -- # : 0 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:19:16.480 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@55 -- # have_pci_nics=0 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@11 -- # loops=5 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@23 -- # nvmftestinit 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@476 -- # prepare_net_devs 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@438 -- # local -g is_hw=no 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@440 -- # remove_spdk_ns 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@309 -- # xtrace_disable 00:19:16.480 17:34:58 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@315 -- # pci_devs=() 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@315 -- # local -a pci_devs 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@316 -- # pci_net_devs=() 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@317 -- # pci_drivers=() 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@317 -- # local -A pci_drivers 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@319 -- # net_devs=() 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@319 -- # local -ga net_devs 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@320 -- # e810=() 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@320 -- # local -ga e810 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@321 -- # x722=() 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@321 -- # local -ga x722 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@322 -- # mlx=() 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@322 -- # local -ga mlx 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:19:19.020 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:19:19.020 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@418 -- # [[ up == up ]] 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:19:19.020 Found net devices under 0000:0a:00.0: cvl_0_0 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@418 -- # [[ up == up ]] 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:19:19.020 Found net devices under 0000:0a:00.1: cvl_0_1 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@442 -- # is_hw=yes 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:19:19.020 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:19:19.020 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.176 ms 00:19:19.020 00:19:19.020 --- 10.0.0.2 ping statistics --- 00:19:19.020 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:19:19.020 rtt min/avg/max/mdev = 0.176/0.176/0.176/0.000 ms 00:19:19.020 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:19:19.020 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:19:19.020 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.075 ms 00:19:19.020 00:19:19.020 --- 10.0.0.1 ping statistics --- 00:19:19.020 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:19:19.021 rtt min/avg/max/mdev = 0.075/0.075/0.075/0.000 ms 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@450 -- # return 0 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@24 -- # nvmfappstart -m 0xF 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@726 -- # xtrace_disable 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@509 -- # nvmfpid=220313 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@510 -- # waitforlisten 220313 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@835 -- # '[' -z 220313 ']' 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@840 -- # local max_retries=100 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:19.021 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@844 -- # xtrace_disable 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:19.021 [2024-12-06 17:35:00.525962] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:19:19.021 [2024-12-06 17:35:00.526060] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:19.021 [2024-12-06 17:35:00.604715] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:19:19.021 [2024-12-06 17:35:00.653966] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:19.021 [2024-12-06 17:35:00.654038] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:19.021 [2024-12-06 17:35:00.654063] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:19.021 [2024-12-06 17:35:00.654074] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:19.021 [2024-12-06 17:35:00.654085] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:19.021 [2024-12-06 17:35:00.655753] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:19:19.021 [2024-12-06 17:35:00.655819] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:19:19.021 [2024-12-06 17:35:00.655884] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:19:19.021 [2024-12-06 17:35:00.655887] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@868 -- # return 0 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@732 -- # xtrace_disable 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@26 -- # rpc_cmd nvmf_get_stats 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@26 -- # stats='{ 00:19:19.021 "tick_rate": 2700000000, 00:19:19.021 "poll_groups": [ 00:19:19.021 { 00:19:19.021 "name": "nvmf_tgt_poll_group_000", 00:19:19.021 "admin_qpairs": 0, 00:19:19.021 "io_qpairs": 0, 00:19:19.021 "current_admin_qpairs": 0, 00:19:19.021 "current_io_qpairs": 0, 00:19:19.021 "pending_bdev_io": 0, 00:19:19.021 "completed_nvme_io": 0, 00:19:19.021 "transports": [] 00:19:19.021 }, 00:19:19.021 { 00:19:19.021 "name": "nvmf_tgt_poll_group_001", 00:19:19.021 "admin_qpairs": 0, 00:19:19.021 "io_qpairs": 0, 00:19:19.021 "current_admin_qpairs": 0, 00:19:19.021 "current_io_qpairs": 0, 00:19:19.021 "pending_bdev_io": 0, 00:19:19.021 "completed_nvme_io": 0, 00:19:19.021 "transports": [] 00:19:19.021 }, 00:19:19.021 { 00:19:19.021 "name": "nvmf_tgt_poll_group_002", 00:19:19.021 "admin_qpairs": 0, 00:19:19.021 "io_qpairs": 0, 00:19:19.021 "current_admin_qpairs": 0, 00:19:19.021 "current_io_qpairs": 0, 00:19:19.021 "pending_bdev_io": 0, 00:19:19.021 "completed_nvme_io": 0, 00:19:19.021 "transports": [] 00:19:19.021 }, 00:19:19.021 { 00:19:19.021 "name": "nvmf_tgt_poll_group_003", 00:19:19.021 "admin_qpairs": 0, 00:19:19.021 "io_qpairs": 0, 00:19:19.021 "current_admin_qpairs": 0, 00:19:19.021 "current_io_qpairs": 0, 00:19:19.021 "pending_bdev_io": 0, 00:19:19.021 "completed_nvme_io": 0, 00:19:19.021 "transports": [] 00:19:19.021 } 00:19:19.021 ] 00:19:19.021 }' 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@28 -- # jcount '.poll_groups[].name' 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@14 -- # local 'filter=.poll_groups[].name' 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@15 -- # jq '.poll_groups[].name' 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@15 -- # wc -l 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@28 -- # (( 4 == 4 )) 00:19:19.021 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@29 -- # jq '.poll_groups[0].transports[0]' 00:19:19.292 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@29 -- # [[ null == null ]] 00:19:19.292 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@31 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:19:19.292 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:19.292 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:19.292 [2024-12-06 17:35:00.892800] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:19:19.292 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:19.292 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@33 -- # rpc_cmd nvmf_get_stats 00:19:19.292 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:19.292 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:19.292 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:19.292 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@33 -- # stats='{ 00:19:19.292 "tick_rate": 2700000000, 00:19:19.292 "poll_groups": [ 00:19:19.292 { 00:19:19.292 "name": "nvmf_tgt_poll_group_000", 00:19:19.292 "admin_qpairs": 0, 00:19:19.292 "io_qpairs": 0, 00:19:19.292 "current_admin_qpairs": 0, 00:19:19.292 "current_io_qpairs": 0, 00:19:19.292 "pending_bdev_io": 0, 00:19:19.292 "completed_nvme_io": 0, 00:19:19.292 "transports": [ 00:19:19.292 { 00:19:19.292 "trtype": "TCP" 00:19:19.292 } 00:19:19.292 ] 00:19:19.292 }, 00:19:19.292 { 00:19:19.292 "name": "nvmf_tgt_poll_group_001", 00:19:19.292 "admin_qpairs": 0, 00:19:19.292 "io_qpairs": 0, 00:19:19.292 "current_admin_qpairs": 0, 00:19:19.292 "current_io_qpairs": 0, 00:19:19.292 "pending_bdev_io": 0, 00:19:19.292 "completed_nvme_io": 0, 00:19:19.292 "transports": [ 00:19:19.292 { 00:19:19.292 "trtype": "TCP" 00:19:19.292 } 00:19:19.292 ] 00:19:19.292 }, 00:19:19.292 { 00:19:19.292 "name": "nvmf_tgt_poll_group_002", 00:19:19.292 "admin_qpairs": 0, 00:19:19.292 "io_qpairs": 0, 00:19:19.292 "current_admin_qpairs": 0, 00:19:19.292 "current_io_qpairs": 0, 00:19:19.292 "pending_bdev_io": 0, 00:19:19.292 "completed_nvme_io": 0, 00:19:19.292 "transports": [ 00:19:19.292 { 00:19:19.292 "trtype": "TCP" 00:19:19.292 } 00:19:19.292 ] 00:19:19.292 }, 00:19:19.292 { 00:19:19.292 "name": "nvmf_tgt_poll_group_003", 00:19:19.292 "admin_qpairs": 0, 00:19:19.292 "io_qpairs": 0, 00:19:19.292 "current_admin_qpairs": 0, 00:19:19.292 "current_io_qpairs": 0, 00:19:19.292 "pending_bdev_io": 0, 00:19:19.292 "completed_nvme_io": 0, 00:19:19.292 "transports": [ 00:19:19.292 { 00:19:19.292 "trtype": "TCP" 00:19:19.292 } 00:19:19.292 ] 00:19:19.292 } 00:19:19.292 ] 00:19:19.292 }' 00:19:19.292 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@35 -- # jsum '.poll_groups[].admin_qpairs' 00:19:19.292 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@19 -- # local 'filter=.poll_groups[].admin_qpairs' 00:19:19.292 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # jq '.poll_groups[].admin_qpairs' 00:19:19.292 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # awk '{s+=$1}END{print s}' 00:19:19.292 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@35 -- # (( 0 == 0 )) 00:19:19.292 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@36 -- # jsum '.poll_groups[].io_qpairs' 00:19:19.292 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@19 -- # local 'filter=.poll_groups[].io_qpairs' 00:19:19.292 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # jq '.poll_groups[].io_qpairs' 00:19:19.292 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # awk '{s+=$1}END{print s}' 00:19:19.292 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@36 -- # (( 0 == 0 )) 00:19:19.292 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@38 -- # '[' rdma == tcp ']' 00:19:19.292 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@46 -- # MALLOC_BDEV_SIZE=64 00:19:19.292 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@47 -- # MALLOC_BLOCK_SIZE=512 00:19:19.292 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@49 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:19:19.292 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:19.292 17:35:00 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:19.292 Malloc1 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@52 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@53 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@54 -- # rpc_cmd nvmf_subsystem_allow_any_host -d nqn.2016-06.io.spdk:cnode1 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@55 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:19.292 [2024-12-06 17:35:01.062579] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@58 -- # NOT nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -a 10.0.0.2 -s 4420 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@652 -- # local es=0 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@654 -- # valid_exec_arg nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -a 10.0.0.2 -s 4420 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@640 -- # local arg=nvme 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@644 -- # type -t nvme 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@646 -- # type -P nvme 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@646 -- # arg=/usr/sbin/nvme 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@646 -- # [[ -x /usr/sbin/nvme ]] 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@655 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -a 10.0.0.2 -s 4420 00:19:19.292 [2024-12-06 17:35:01.085241] ctrlr.c: 825:nvmf_qpair_access_allowed: *ERROR*: Subsystem 'nqn.2016-06.io.spdk:cnode1' does not allow host 'nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55' 00:19:19.292 Failed to write to /dev/nvme-fabrics: Input/output error 00:19:19.292 could not add new controller: failed to write to nvme-fabrics device 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@655 -- # es=1 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@61 -- # rpc_cmd nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:19.292 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@62 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:19:20.236 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@63 -- # waitforserial SPDKISFASTANDAWESOME 00:19:20.236 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1202 -- # local i=0 00:19:20.236 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:19:20.236 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:19:20.236 17:35:01 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1209 -- # sleep 2 00:19:22.130 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:19:22.130 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:19:22.130 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:19:22.130 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:19:22.130 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:19:22.130 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1212 -- # return 0 00:19:22.130 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@64 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:19:22.130 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:19:22.130 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@65 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:19:22.130 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1223 -- # local i=0 00:19:22.130 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:19:22.130 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:19:22.130 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:19:22.130 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:19:22.130 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1235 -- # return 0 00:19:22.131 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@68 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2016-06.io.spdk:cnode1 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:19:22.131 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:22.131 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:22.131 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:22.131 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@69 -- # NOT nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:19:22.131 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@652 -- # local es=0 00:19:22.131 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@654 -- # valid_exec_arg nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:19:22.131 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@640 -- # local arg=nvme 00:19:22.131 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:19:22.131 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@644 -- # type -t nvme 00:19:22.131 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:19:22.131 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@646 -- # type -P nvme 00:19:22.131 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:19:22.131 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@646 -- # arg=/usr/sbin/nvme 00:19:22.131 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@646 -- # [[ -x /usr/sbin/nvme ]] 00:19:22.131 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@655 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:19:22.131 [2024-12-06 17:35:03.875745] ctrlr.c: 825:nvmf_qpair_access_allowed: *ERROR*: Subsystem 'nqn.2016-06.io.spdk:cnode1' does not allow host 'nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55' 00:19:22.131 Failed to write to /dev/nvme-fabrics: Input/output error 00:19:22.131 could not add new controller: failed to write to nvme-fabrics device 00:19:22.131 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@655 -- # es=1 00:19:22.131 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:19:22.131 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:19:22.131 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:19:22.131 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@72 -- # rpc_cmd nvmf_subsystem_allow_any_host -e nqn.2016-06.io.spdk:cnode1 00:19:22.131 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:22.131 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:22.131 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:22.131 17:35:03 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@73 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:19:22.695 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@74 -- # waitforserial SPDKISFASTANDAWESOME 00:19:22.695 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1202 -- # local i=0 00:19:22.695 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:19:22.696 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:19:22.696 17:35:04 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1209 -- # sleep 2 00:19:25.216 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:19:25.216 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1212 -- # return 0 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@75 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:19:25.217 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@76 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1223 -- # local i=0 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1235 -- # return 0 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@78 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@81 -- # seq 1 5 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:25.217 [2024-12-06 17:35:06.665703] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:25.217 17:35:06 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:19:25.783 17:35:07 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:19:25.783 17:35:07 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1202 -- # local i=0 00:19:25.783 17:35:07 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:19:25.783 17:35:07 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:19:25.783 17:35:07 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1209 -- # sleep 2 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1212 -- # return 0 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:19:27.684 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1223 -- # local i=0 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1235 -- # return 0 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:27.684 [2024-12-06 17:35:09.473212] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:27.684 17:35:09 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:19:28.618 17:35:10 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:19:28.618 17:35:10 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1202 -- # local i=0 00:19:28.618 17:35:10 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:19:28.618 17:35:10 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:19:28.618 17:35:10 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1209 -- # sleep 2 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1212 -- # return 0 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:19:30.519 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1223 -- # local i=0 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1235 -- # return 0 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:30.519 [2024-12-06 17:35:12.298300] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:30.519 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:19:30.520 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:30.520 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:30.520 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:30.520 17:35:12 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:19:31.453 17:35:13 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:19:31.453 17:35:13 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1202 -- # local i=0 00:19:31.453 17:35:13 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:19:31.453 17:35:13 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:19:31.453 17:35:13 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1209 -- # sleep 2 00:19:33.355 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:19:33.355 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:19:33.355 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:19:33.355 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:19:33.355 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1212 -- # return 0 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:19:33.356 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1223 -- # local i=0 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1235 -- # return 0 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:33.356 [2024-12-06 17:35:15.108308] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:33.356 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:19:33.921 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:19:33.921 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1202 -- # local i=0 00:19:33.921 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:19:33.921 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:19:33.921 17:35:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1209 -- # sleep 2 00:19:36.452 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:19:36.452 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:19:36.452 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:19:36.452 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:19:36.452 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:19:36.452 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1212 -- # return 0 00:19:36.452 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:19:36.452 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:19:36.452 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:19:36.452 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1223 -- # local i=0 00:19:36.452 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:19:36.452 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:19:36.452 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:19:36.452 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:19:36.452 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1235 -- # return 0 00:19:36.452 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:19:36.452 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:36.452 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:36.452 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:36.452 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:19:36.452 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:36.452 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:36.452 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:36.452 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:19:36.452 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:19:36.453 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:36.453 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:36.453 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:36.453 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:19:36.453 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:36.453 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:36.453 [2024-12-06 17:35:17.839204] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:36.453 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:36.453 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:19:36.453 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:36.453 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:36.453 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:36.453 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:19:36.453 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:36.453 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:36.453 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:36.453 17:35:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:19:37.019 17:35:18 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:19:37.019 17:35:18 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1202 -- # local i=0 00:19:37.019 17:35:18 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:19:37.019 17:35:18 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:19:37.019 17:35:18 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1209 -- # sleep 2 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1212 -- # return 0 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:19:38.919 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1223 -- # local i=0 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1235 -- # return 0 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@99 -- # seq 1 5 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:38.919 [2024-12-06 17:35:20.731472] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:19:38.919 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:38.920 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:38.920 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:38.920 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:19:38.920 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:38.920 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:38.920 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:38.920 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:19:38.920 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:38.920 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:39.178 [2024-12-06 17:35:20.779533] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:39.178 [2024-12-06 17:35:20.827719] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:39.178 [2024-12-06 17:35:20.875875] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:39.178 [2024-12-06 17:35:20.924054] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@110 -- # rpc_cmd nvmf_get_stats 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:39.178 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@110 -- # stats='{ 00:19:39.178 "tick_rate": 2700000000, 00:19:39.178 "poll_groups": [ 00:19:39.178 { 00:19:39.178 "name": "nvmf_tgt_poll_group_000", 00:19:39.178 "admin_qpairs": 2, 00:19:39.178 "io_qpairs": 84, 00:19:39.178 "current_admin_qpairs": 0, 00:19:39.178 "current_io_qpairs": 0, 00:19:39.179 "pending_bdev_io": 0, 00:19:39.179 "completed_nvme_io": 182, 00:19:39.179 "transports": [ 00:19:39.179 { 00:19:39.179 "trtype": "TCP" 00:19:39.179 } 00:19:39.179 ] 00:19:39.179 }, 00:19:39.179 { 00:19:39.179 "name": "nvmf_tgt_poll_group_001", 00:19:39.179 "admin_qpairs": 2, 00:19:39.179 "io_qpairs": 84, 00:19:39.179 "current_admin_qpairs": 0, 00:19:39.179 "current_io_qpairs": 0, 00:19:39.179 "pending_bdev_io": 0, 00:19:39.179 "completed_nvme_io": 185, 00:19:39.179 "transports": [ 00:19:39.179 { 00:19:39.179 "trtype": "TCP" 00:19:39.179 } 00:19:39.179 ] 00:19:39.179 }, 00:19:39.179 { 00:19:39.179 "name": "nvmf_tgt_poll_group_002", 00:19:39.179 "admin_qpairs": 1, 00:19:39.179 "io_qpairs": 84, 00:19:39.179 "current_admin_qpairs": 0, 00:19:39.179 "current_io_qpairs": 0, 00:19:39.179 "pending_bdev_io": 0, 00:19:39.179 "completed_nvme_io": 183, 00:19:39.179 "transports": [ 00:19:39.179 { 00:19:39.179 "trtype": "TCP" 00:19:39.179 } 00:19:39.179 ] 00:19:39.179 }, 00:19:39.179 { 00:19:39.179 "name": "nvmf_tgt_poll_group_003", 00:19:39.179 "admin_qpairs": 2, 00:19:39.179 "io_qpairs": 84, 00:19:39.179 "current_admin_qpairs": 0, 00:19:39.179 "current_io_qpairs": 0, 00:19:39.179 "pending_bdev_io": 0, 00:19:39.179 "completed_nvme_io": 136, 00:19:39.179 "transports": [ 00:19:39.179 { 00:19:39.179 "trtype": "TCP" 00:19:39.179 } 00:19:39.179 ] 00:19:39.179 } 00:19:39.179 ] 00:19:39.179 }' 00:19:39.179 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@112 -- # jsum '.poll_groups[].admin_qpairs' 00:19:39.179 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@19 -- # local 'filter=.poll_groups[].admin_qpairs' 00:19:39.179 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # jq '.poll_groups[].admin_qpairs' 00:19:39.179 17:35:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # awk '{s+=$1}END{print s}' 00:19:39.179 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@112 -- # (( 7 > 0 )) 00:19:39.179 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@113 -- # jsum '.poll_groups[].io_qpairs' 00:19:39.179 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@19 -- # local 'filter=.poll_groups[].io_qpairs' 00:19:39.179 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # jq '.poll_groups[].io_qpairs' 00:19:39.179 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # awk '{s+=$1}END{print s}' 00:19:39.436 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@113 -- # (( 336 > 0 )) 00:19:39.436 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@115 -- # '[' rdma == tcp ']' 00:19:39.436 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@121 -- # trap - SIGINT SIGTERM EXIT 00:19:39.436 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@123 -- # nvmftestfini 00:19:39.436 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@516 -- # nvmfcleanup 00:19:39.436 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@121 -- # sync 00:19:39.436 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:19:39.436 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@124 -- # set +e 00:19:39.436 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@125 -- # for i in {1..20} 00:19:39.436 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:19:39.436 rmmod nvme_tcp 00:19:39.436 rmmod nvme_fabrics 00:19:39.436 rmmod nvme_keyring 00:19:39.436 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:19:39.436 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@128 -- # set -e 00:19:39.436 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@129 -- # return 0 00:19:39.436 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@517 -- # '[' -n 220313 ']' 00:19:39.436 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@518 -- # killprocess 220313 00:19:39.436 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@954 -- # '[' -z 220313 ']' 00:19:39.436 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@958 -- # kill -0 220313 00:19:39.436 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@959 -- # uname 00:19:39.436 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:19:39.436 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 220313 00:19:39.436 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:19:39.436 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:19:39.436 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 220313' 00:19:39.436 killing process with pid 220313 00:19:39.436 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@973 -- # kill 220313 00:19:39.437 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@978 -- # wait 220313 00:19:39.697 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:19:39.697 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:19:39.697 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:19:39.697 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@297 -- # iptr 00:19:39.697 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@791 -- # iptables-save 00:19:39.697 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:19:39.697 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@791 -- # iptables-restore 00:19:39.697 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:19:39.697 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@302 -- # remove_spdk_ns 00:19:39.697 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:19:39.697 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:19:39.697 17:35:21 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:19:41.607 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:19:41.607 00:19:41.607 real 0m25.356s 00:19:41.607 user 1m22.191s 00:19:41.607 sys 0m4.198s 00:19:41.607 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:19:41.607 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:19:41.607 ************************************ 00:19:41.607 END TEST nvmf_rpc 00:19:41.607 ************************************ 00:19:41.607 17:35:23 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@23 -- # run_test nvmf_invalid /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/invalid.sh --transport=tcp 00:19:41.607 17:35:23 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:19:41.607 17:35:23 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:19:41.607 17:35:23 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:19:41.607 ************************************ 00:19:41.607 START TEST nvmf_invalid 00:19:41.607 ************************************ 00:19:41.607 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/invalid.sh --transport=tcp 00:19:41.889 * Looking for test storage... 00:19:41.889 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:19:41.889 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:19:41.889 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1711 -- # lcov --version 00:19:41.889 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:19:41.889 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:19:41.889 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:19:41.889 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:19:41.889 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@336 -- # IFS=.-: 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@336 -- # read -ra ver1 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@337 -- # IFS=.-: 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@337 -- # read -ra ver2 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@338 -- # local 'op=<' 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@340 -- # ver1_l=2 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@341 -- # ver2_l=1 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@344 -- # case "$op" in 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@345 -- # : 1 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@364 -- # (( v = 0 )) 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@365 -- # decimal 1 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@353 -- # local d=1 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@355 -- # echo 1 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@365 -- # ver1[v]=1 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@366 -- # decimal 2 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@353 -- # local d=2 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@355 -- # echo 2 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@366 -- # ver2[v]=2 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@368 -- # return 0 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:19:41.890 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:41.890 --rc genhtml_branch_coverage=1 00:19:41.890 --rc genhtml_function_coverage=1 00:19:41.890 --rc genhtml_legend=1 00:19:41.890 --rc geninfo_all_blocks=1 00:19:41.890 --rc geninfo_unexecuted_blocks=1 00:19:41.890 00:19:41.890 ' 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:19:41.890 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:41.890 --rc genhtml_branch_coverage=1 00:19:41.890 --rc genhtml_function_coverage=1 00:19:41.890 --rc genhtml_legend=1 00:19:41.890 --rc geninfo_all_blocks=1 00:19:41.890 --rc geninfo_unexecuted_blocks=1 00:19:41.890 00:19:41.890 ' 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:19:41.890 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:41.890 --rc genhtml_branch_coverage=1 00:19:41.890 --rc genhtml_function_coverage=1 00:19:41.890 --rc genhtml_legend=1 00:19:41.890 --rc geninfo_all_blocks=1 00:19:41.890 --rc geninfo_unexecuted_blocks=1 00:19:41.890 00:19:41.890 ' 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:19:41.890 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:41.890 --rc genhtml_branch_coverage=1 00:19:41.890 --rc genhtml_function_coverage=1 00:19:41.890 --rc genhtml_legend=1 00:19:41.890 --rc geninfo_all_blocks=1 00:19:41.890 --rc geninfo_unexecuted_blocks=1 00:19:41.890 00:19:41.890 ' 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@7 -- # uname -s 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@15 -- # shopt -s extglob 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- paths/export.sh@5 -- # export PATH 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@51 -- # : 0 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:19:41.890 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@55 -- # have_pci_nics=0 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@11 -- # multi_target_rpc=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@12 -- # rpc=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@13 -- # nqn=nqn.2016-06.io.spdk:cnode 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@14 -- # target=foobar 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@16 -- # RANDOM=0 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@34 -- # nvmftestinit 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@476 -- # prepare_net_devs 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@438 -- # local -g is_hw=no 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@440 -- # remove_spdk_ns 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@309 -- # xtrace_disable 00:19:41.890 17:35:23 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@10 -- # set +x 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@315 -- # pci_devs=() 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@315 -- # local -a pci_devs 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@316 -- # pci_net_devs=() 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@317 -- # pci_drivers=() 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@317 -- # local -A pci_drivers 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@319 -- # net_devs=() 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@319 -- # local -ga net_devs 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@320 -- # e810=() 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@320 -- # local -ga e810 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@321 -- # x722=() 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@321 -- # local -ga x722 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@322 -- # mlx=() 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@322 -- # local -ga mlx 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:19:44.419 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:19:44.419 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@418 -- # [[ up == up ]] 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:19:44.419 Found net devices under 0000:0a:00.0: cvl_0_0 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@418 -- # [[ up == up ]] 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:19:44.419 Found net devices under 0000:0a:00.1: cvl_0_1 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@442 -- # is_hw=yes 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:19:44.419 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:19:44.419 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.141 ms 00:19:44.419 00:19:44.419 --- 10.0.0.2 ping statistics --- 00:19:44.419 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:19:44.419 rtt min/avg/max/mdev = 0.141/0.141/0.141/0.000 ms 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:19:44.419 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:19:44.419 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.091 ms 00:19:44.419 00:19:44.419 --- 10.0.0.1 ping statistics --- 00:19:44.419 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:19:44.419 rtt min/avg/max/mdev = 0.091/0.091/0.091/0.000 ms 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@450 -- # return 0 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@35 -- # nvmfappstart -m 0xF 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@726 -- # xtrace_disable 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@10 -- # set +x 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@509 -- # nvmfpid=224808 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@510 -- # waitforlisten 224808 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@835 -- # '[' -z 224808 ']' 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@840 -- # local max_retries=100 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:44.419 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@844 -- # xtrace_disable 00:19:44.419 17:35:25 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@10 -- # set +x 00:19:44.419 [2024-12-06 17:35:26.036290] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:19:44.419 [2024-12-06 17:35:26.036389] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:44.419 [2024-12-06 17:35:26.107394] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:19:44.419 [2024-12-06 17:35:26.152336] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:44.420 [2024-12-06 17:35:26.152386] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:44.420 [2024-12-06 17:35:26.152407] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:44.420 [2024-12-06 17:35:26.152432] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:44.420 [2024-12-06 17:35:26.152441] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:44.420 [2024-12-06 17:35:26.154069] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:19:44.420 [2024-12-06 17:35:26.154135] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:19:44.420 [2024-12-06 17:35:26.154211] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:19:44.420 [2024-12-06 17:35:26.154216] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:19:44.676 17:35:26 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:19:44.676 17:35:26 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@868 -- # return 0 00:19:44.676 17:35:26 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:19:44.676 17:35:26 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@732 -- # xtrace_disable 00:19:44.676 17:35:26 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@10 -- # set +x 00:19:44.676 17:35:26 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:19:44.676 17:35:26 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@37 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini $1; exit 1' SIGINT SIGTERM EXIT 00:19:44.676 17:35:26 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem -t foobar nqn.2016-06.io.spdk:cnode22075 00:19:44.934 [2024-12-06 17:35:26.544408] nvmf_rpc.c: 396:rpc_nvmf_create_subsystem: *ERROR*: Unable to find target foobar 00:19:44.934 17:35:26 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@40 -- # out='request: 00:19:44.934 { 00:19:44.934 "nqn": "nqn.2016-06.io.spdk:cnode22075", 00:19:44.934 "tgt_name": "foobar", 00:19:44.934 "method": "nvmf_create_subsystem", 00:19:44.934 "req_id": 1 00:19:44.934 } 00:19:44.934 Got JSON-RPC error response 00:19:44.934 response: 00:19:44.934 { 00:19:44.934 "code": -32603, 00:19:44.934 "message": "Unable to find target foobar" 00:19:44.934 }' 00:19:44.934 17:35:26 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@41 -- # [[ request: 00:19:44.934 { 00:19:44.934 "nqn": "nqn.2016-06.io.spdk:cnode22075", 00:19:44.934 "tgt_name": "foobar", 00:19:44.934 "method": "nvmf_create_subsystem", 00:19:44.934 "req_id": 1 00:19:44.934 } 00:19:44.934 Got JSON-RPC error response 00:19:44.934 response: 00:19:44.934 { 00:19:44.934 "code": -32603, 00:19:44.934 "message": "Unable to find target foobar" 00:19:44.934 } == *\U\n\a\b\l\e\ \t\o\ \f\i\n\d\ \t\a\r\g\e\t* ]] 00:19:44.934 17:35:26 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@45 -- # echo -e '\x1f' 00:19:44.934 17:35:26 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem -s $'SPDKISFASTANDAWESOME\037' nqn.2016-06.io.spdk:cnode30735 00:19:45.191 [2024-12-06 17:35:26.865462] nvmf_rpc.c: 413:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode30735: invalid serial number 'SPDKISFASTANDAWESOME' 00:19:45.191 17:35:26 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@45 -- # out='request: 00:19:45.191 { 00:19:45.191 "nqn": "nqn.2016-06.io.spdk:cnode30735", 00:19:45.191 "serial_number": "SPDKISFASTANDAWESOME\u001f", 00:19:45.191 "method": "nvmf_create_subsystem", 00:19:45.191 "req_id": 1 00:19:45.191 } 00:19:45.191 Got JSON-RPC error response 00:19:45.191 response: 00:19:45.191 { 00:19:45.191 "code": -32602, 00:19:45.191 "message": "Invalid SN SPDKISFASTANDAWESOME\u001f" 00:19:45.191 }' 00:19:45.191 17:35:26 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@46 -- # [[ request: 00:19:45.191 { 00:19:45.191 "nqn": "nqn.2016-06.io.spdk:cnode30735", 00:19:45.191 "serial_number": "SPDKISFASTANDAWESOME\u001f", 00:19:45.191 "method": "nvmf_create_subsystem", 00:19:45.191 "req_id": 1 00:19:45.191 } 00:19:45.191 Got JSON-RPC error response 00:19:45.191 response: 00:19:45.191 { 00:19:45.191 "code": -32602, 00:19:45.191 "message": "Invalid SN SPDKISFASTANDAWESOME\u001f" 00:19:45.191 } == *\I\n\v\a\l\i\d\ \S\N* ]] 00:19:45.191 17:35:26 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@50 -- # echo -e '\x1f' 00:19:45.191 17:35:26 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem -d $'SPDK_Controller\037' nqn.2016-06.io.spdk:cnode11154 00:19:45.448 [2024-12-06 17:35:27.154414] nvmf_rpc.c: 422:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode11154: invalid model number 'SPDK_Controller' 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@50 -- # out='request: 00:19:45.448 { 00:19:45.448 "nqn": "nqn.2016-06.io.spdk:cnode11154", 00:19:45.448 "model_number": "SPDK_Controller\u001f", 00:19:45.448 "method": "nvmf_create_subsystem", 00:19:45.448 "req_id": 1 00:19:45.448 } 00:19:45.448 Got JSON-RPC error response 00:19:45.448 response: 00:19:45.448 { 00:19:45.448 "code": -32602, 00:19:45.448 "message": "Invalid MN SPDK_Controller\u001f" 00:19:45.448 }' 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@51 -- # [[ request: 00:19:45.448 { 00:19:45.448 "nqn": "nqn.2016-06.io.spdk:cnode11154", 00:19:45.448 "model_number": "SPDK_Controller\u001f", 00:19:45.448 "method": "nvmf_create_subsystem", 00:19:45.448 "req_id": 1 00:19:45.448 } 00:19:45.448 Got JSON-RPC error response 00:19:45.448 response: 00:19:45.448 { 00:19:45.448 "code": -32602, 00:19:45.448 "message": "Invalid MN SPDK_Controller\u001f" 00:19:45.448 } == *\I\n\v\a\l\i\d\ \M\N* ]] 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@54 -- # gen_random_s 21 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@19 -- # local length=21 ll 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@21 -- # chars=('32' '33' '34' '35' '36' '37' '38' '39' '40' '41' '42' '43' '44' '45' '46' '47' '48' '49' '50' '51' '52' '53' '54' '55' '56' '57' '58' '59' '60' '61' '62' '63' '64' '65' '66' '67' '68' '69' '70' '71' '72' '73' '74' '75' '76' '77' '78' '79' '80' '81' '82' '83' '84' '85' '86' '87' '88' '89' '90' '91' '92' '93' '94' '95' '96' '97' '98' '99' '100' '101' '102' '103' '104' '105' '106' '107' '108' '109' '110' '111' '112' '113' '114' '115' '116' '117' '118' '119' '120' '121' '122' '123' '124' '125' '126' '127') 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@21 -- # local chars 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@22 -- # local string 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll = 0 )) 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 94 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x5e' 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='^' 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 79 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x4f' 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=O 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 96 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x60' 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='`' 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 114 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x72' 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=r 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 97 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x61' 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=a 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 111 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x6f' 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=o 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 94 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x5e' 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='^' 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.448 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 98 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x62' 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=b 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 33 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x21' 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='!' 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 118 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x76' 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=v 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 37 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x25' 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=% 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 104 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x68' 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=h 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 88 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x58' 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=X 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 118 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x76' 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=v 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 63 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x3f' 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='?' 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 65 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x41' 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=A 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 98 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x62' 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=b 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 38 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x26' 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='&' 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 101 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x65' 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=e 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 54 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x36' 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=6 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 87 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x57' 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=W 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@28 -- # [[ ^ == \- ]] 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@31 -- # echo '^O`rao^b!v%hXv?Ab&e6W' 00:19:45.449 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem -s '^O`rao^b!v%hXv?Ab&e6W' nqn.2016-06.io.spdk:cnode6579 00:19:45.706 [2024-12-06 17:35:27.499594] nvmf_rpc.c: 413:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode6579: invalid serial number '^O`rao^b!v%hXv?Ab&e6W' 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@54 -- # out='request: 00:19:45.706 { 00:19:45.706 "nqn": "nqn.2016-06.io.spdk:cnode6579", 00:19:45.706 "serial_number": "^O`rao^b!v%hXv?Ab&e6W", 00:19:45.706 "method": "nvmf_create_subsystem", 00:19:45.706 "req_id": 1 00:19:45.706 } 00:19:45.706 Got JSON-RPC error response 00:19:45.706 response: 00:19:45.706 { 00:19:45.706 "code": -32602, 00:19:45.706 "message": "Invalid SN ^O`rao^b!v%hXv?Ab&e6W" 00:19:45.706 }' 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@55 -- # [[ request: 00:19:45.706 { 00:19:45.706 "nqn": "nqn.2016-06.io.spdk:cnode6579", 00:19:45.706 "serial_number": "^O`rao^b!v%hXv?Ab&e6W", 00:19:45.706 "method": "nvmf_create_subsystem", 00:19:45.706 "req_id": 1 00:19:45.706 } 00:19:45.706 Got JSON-RPC error response 00:19:45.706 response: 00:19:45.706 { 00:19:45.706 "code": -32602, 00:19:45.706 "message": "Invalid SN ^O`rao^b!v%hXv?Ab&e6W" 00:19:45.706 } == *\I\n\v\a\l\i\d\ \S\N* ]] 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@58 -- # gen_random_s 41 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@19 -- # local length=41 ll 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@21 -- # chars=('32' '33' '34' '35' '36' '37' '38' '39' '40' '41' '42' '43' '44' '45' '46' '47' '48' '49' '50' '51' '52' '53' '54' '55' '56' '57' '58' '59' '60' '61' '62' '63' '64' '65' '66' '67' '68' '69' '70' '71' '72' '73' '74' '75' '76' '77' '78' '79' '80' '81' '82' '83' '84' '85' '86' '87' '88' '89' '90' '91' '92' '93' '94' '95' '96' '97' '98' '99' '100' '101' '102' '103' '104' '105' '106' '107' '108' '109' '110' '111' '112' '113' '114' '115' '116' '117' '118' '119' '120' '121' '122' '123' '124' '125' '126' '127') 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@21 -- # local chars 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@22 -- # local string 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll = 0 )) 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 69 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x45' 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=E 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 35 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x23' 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='#' 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 57 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x39' 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=9 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 58 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x3a' 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=: 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.706 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 102 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x66' 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=f 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 55 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x37' 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=7 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 44 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x2c' 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=, 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 37 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x25' 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=% 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 43 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x2b' 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=+ 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 69 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x45' 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=E 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 44 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x2c' 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=, 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 73 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x49' 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=I 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 110 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x6e' 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=n 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 113 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x71' 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=q 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 51 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x33' 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=3 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 73 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x49' 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=I 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 40 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x28' 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='(' 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 114 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x72' 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=r 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 108 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x6c' 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=l 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 51 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x33' 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=3 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 121 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x79' 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=y 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 107 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x6b' 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=k 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 43 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x2b' 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=+ 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 54 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x36' 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=6 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 63 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x3f' 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='?' 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.964 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 89 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x59' 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=Y 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 77 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x4d' 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=M 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 121 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x79' 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=y 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 46 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x2e' 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=. 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 113 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x71' 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=q 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 92 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x5c' 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='\' 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 65 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x41' 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=A 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 102 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x66' 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=f 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 120 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x78' 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=x 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 77 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x4d' 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=M 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 39 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x27' 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=\' 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 37 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x25' 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=% 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 50 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x32' 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=2 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 120 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x78' 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=x 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 99 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x63' 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=c 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 70 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x46' 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=F 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@28 -- # [[ E == \- ]] 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@31 -- # echo 'E#9:f7,%+E,Inq3I(rl3yk+6?YMy.q\AfxM'\''%2xcF' 00:19:45.965 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem -d 'E#9:f7,%+E,Inq3I(rl3yk+6?YMy.q\AfxM'\''%2xcF' nqn.2016-06.io.spdk:cnode1229 00:19:46.223 [2024-12-06 17:35:27.920997] nvmf_rpc.c: 422:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode1229: invalid model number 'E#9:f7,%+E,Inq3I(rl3yk+6?YMy.q\AfxM'%2xcF' 00:19:46.223 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@58 -- # out='request: 00:19:46.223 { 00:19:46.223 "nqn": "nqn.2016-06.io.spdk:cnode1229", 00:19:46.223 "model_number": "E#9:f7,%+E,Inq3I(rl3yk+6?YMy.q\\AfxM'\''%2xcF", 00:19:46.223 "method": "nvmf_create_subsystem", 00:19:46.223 "req_id": 1 00:19:46.223 } 00:19:46.223 Got JSON-RPC error response 00:19:46.223 response: 00:19:46.223 { 00:19:46.223 "code": -32602, 00:19:46.223 "message": "Invalid MN E#9:f7,%+E,Inq3I(rl3yk+6?YMy.q\\AfxM'\''%2xcF" 00:19:46.223 }' 00:19:46.223 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@59 -- # [[ request: 00:19:46.223 { 00:19:46.223 "nqn": "nqn.2016-06.io.spdk:cnode1229", 00:19:46.223 "model_number": "E#9:f7,%+E,Inq3I(rl3yk+6?YMy.q\\AfxM'%2xcF", 00:19:46.223 "method": "nvmf_create_subsystem", 00:19:46.223 "req_id": 1 00:19:46.223 } 00:19:46.223 Got JSON-RPC error response 00:19:46.223 response: 00:19:46.223 { 00:19:46.223 "code": -32602, 00:19:46.223 "message": "Invalid MN E#9:f7,%+E,Inq3I(rl3yk+6?YMy.q\\AfxM'%2xcF" 00:19:46.223 } == *\I\n\v\a\l\i\d\ \M\N* ]] 00:19:46.223 17:35:27 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@62 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport --trtype tcp 00:19:46.481 [2024-12-06 17:35:28.186006] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:19:46.481 17:35:28 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode -s SPDK001 -a 00:19:46.747 17:35:28 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@64 -- # [[ tcp == \T\C\P ]] 00:19:46.747 17:35:28 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@67 -- # echo '' 00:19:46.747 17:35:28 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@67 -- # head -n 1 00:19:46.747 17:35:28 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@67 -- # IP= 00:19:46.747 17:35:28 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@69 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode -t tcp -a '' -s 4421 00:19:47.020 [2024-12-06 17:35:28.759874] nvmf_rpc.c: 783:nvmf_rpc_listen_paused: *ERROR*: Unable to remove listener, rc -2 00:19:47.020 17:35:28 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@69 -- # out='request: 00:19:47.020 { 00:19:47.020 "nqn": "nqn.2016-06.io.spdk:cnode", 00:19:47.020 "listen_address": { 00:19:47.020 "trtype": "tcp", 00:19:47.020 "traddr": "", 00:19:47.020 "trsvcid": "4421" 00:19:47.020 }, 00:19:47.020 "method": "nvmf_subsystem_remove_listener", 00:19:47.020 "req_id": 1 00:19:47.020 } 00:19:47.020 Got JSON-RPC error response 00:19:47.020 response: 00:19:47.020 { 00:19:47.020 "code": -32602, 00:19:47.020 "message": "Invalid parameters" 00:19:47.020 }' 00:19:47.020 17:35:28 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@70 -- # [[ request: 00:19:47.020 { 00:19:47.020 "nqn": "nqn.2016-06.io.spdk:cnode", 00:19:47.020 "listen_address": { 00:19:47.020 "trtype": "tcp", 00:19:47.020 "traddr": "", 00:19:47.020 "trsvcid": "4421" 00:19:47.020 }, 00:19:47.020 "method": "nvmf_subsystem_remove_listener", 00:19:47.020 "req_id": 1 00:19:47.020 } 00:19:47.020 Got JSON-RPC error response 00:19:47.020 response: 00:19:47.021 { 00:19:47.021 "code": -32602, 00:19:47.021 "message": "Invalid parameters" 00:19:47.021 } != *\U\n\a\b\l\e\ \t\o\ \s\t\o\p\ \l\i\s\t\e\n\e\r\.* ]] 00:19:47.021 17:35:28 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode5025 -i 0 00:19:47.296 [2024-12-06 17:35:29.028769] nvmf_rpc.c: 434:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode5025: invalid cntlid range [0-65519] 00:19:47.296 17:35:29 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@73 -- # out='request: 00:19:47.296 { 00:19:47.296 "nqn": "nqn.2016-06.io.spdk:cnode5025", 00:19:47.296 "min_cntlid": 0, 00:19:47.296 "method": "nvmf_create_subsystem", 00:19:47.296 "req_id": 1 00:19:47.296 } 00:19:47.296 Got JSON-RPC error response 00:19:47.296 response: 00:19:47.296 { 00:19:47.296 "code": -32602, 00:19:47.296 "message": "Invalid cntlid range [0-65519]" 00:19:47.296 }' 00:19:47.296 17:35:29 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@74 -- # [[ request: 00:19:47.296 { 00:19:47.296 "nqn": "nqn.2016-06.io.spdk:cnode5025", 00:19:47.296 "min_cntlid": 0, 00:19:47.296 "method": "nvmf_create_subsystem", 00:19:47.296 "req_id": 1 00:19:47.296 } 00:19:47.296 Got JSON-RPC error response 00:19:47.296 response: 00:19:47.296 { 00:19:47.296 "code": -32602, 00:19:47.296 "message": "Invalid cntlid range [0-65519]" 00:19:47.296 } == *\I\n\v\a\l\i\d\ \c\n\t\l\i\d\ \r\a\n\g\e* ]] 00:19:47.296 17:35:29 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@75 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode15344 -i 65520 00:19:47.571 [2024-12-06 17:35:29.301688] nvmf_rpc.c: 434:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode15344: invalid cntlid range [65520-65519] 00:19:47.571 17:35:29 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@75 -- # out='request: 00:19:47.571 { 00:19:47.571 "nqn": "nqn.2016-06.io.spdk:cnode15344", 00:19:47.571 "min_cntlid": 65520, 00:19:47.571 "method": "nvmf_create_subsystem", 00:19:47.572 "req_id": 1 00:19:47.572 } 00:19:47.572 Got JSON-RPC error response 00:19:47.572 response: 00:19:47.572 { 00:19:47.572 "code": -32602, 00:19:47.572 "message": "Invalid cntlid range [65520-65519]" 00:19:47.572 }' 00:19:47.572 17:35:29 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@76 -- # [[ request: 00:19:47.572 { 00:19:47.572 "nqn": "nqn.2016-06.io.spdk:cnode15344", 00:19:47.572 "min_cntlid": 65520, 00:19:47.572 "method": "nvmf_create_subsystem", 00:19:47.572 "req_id": 1 00:19:47.572 } 00:19:47.572 Got JSON-RPC error response 00:19:47.572 response: 00:19:47.572 { 00:19:47.572 "code": -32602, 00:19:47.572 "message": "Invalid cntlid range [65520-65519]" 00:19:47.572 } == *\I\n\v\a\l\i\d\ \c\n\t\l\i\d\ \r\a\n\g\e* ]] 00:19:47.572 17:35:29 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@77 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode27629 -I 0 00:19:47.843 [2024-12-06 17:35:29.590658] nvmf_rpc.c: 434:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode27629: invalid cntlid range [1-0] 00:19:47.843 17:35:29 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@77 -- # out='request: 00:19:47.843 { 00:19:47.843 "nqn": "nqn.2016-06.io.spdk:cnode27629", 00:19:47.843 "max_cntlid": 0, 00:19:47.843 "method": "nvmf_create_subsystem", 00:19:47.843 "req_id": 1 00:19:47.843 } 00:19:47.843 Got JSON-RPC error response 00:19:47.843 response: 00:19:47.843 { 00:19:47.843 "code": -32602, 00:19:47.843 "message": "Invalid cntlid range [1-0]" 00:19:47.843 }' 00:19:47.843 17:35:29 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@78 -- # [[ request: 00:19:47.843 { 00:19:47.843 "nqn": "nqn.2016-06.io.spdk:cnode27629", 00:19:47.843 "max_cntlid": 0, 00:19:47.843 "method": "nvmf_create_subsystem", 00:19:47.843 "req_id": 1 00:19:47.843 } 00:19:47.843 Got JSON-RPC error response 00:19:47.843 response: 00:19:47.843 { 00:19:47.843 "code": -32602, 00:19:47.843 "message": "Invalid cntlid range [1-0]" 00:19:47.843 } == *\I\n\v\a\l\i\d\ \c\n\t\l\i\d\ \r\a\n\g\e* ]] 00:19:47.843 17:35:29 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode901 -I 65520 00:19:48.117 [2024-12-06 17:35:29.855520] nvmf_rpc.c: 434:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode901: invalid cntlid range [1-65520] 00:19:48.117 17:35:29 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@79 -- # out='request: 00:19:48.117 { 00:19:48.117 "nqn": "nqn.2016-06.io.spdk:cnode901", 00:19:48.117 "max_cntlid": 65520, 00:19:48.117 "method": "nvmf_create_subsystem", 00:19:48.117 "req_id": 1 00:19:48.117 } 00:19:48.117 Got JSON-RPC error response 00:19:48.117 response: 00:19:48.117 { 00:19:48.117 "code": -32602, 00:19:48.117 "message": "Invalid cntlid range [1-65520]" 00:19:48.117 }' 00:19:48.117 17:35:29 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@80 -- # [[ request: 00:19:48.117 { 00:19:48.117 "nqn": "nqn.2016-06.io.spdk:cnode901", 00:19:48.117 "max_cntlid": 65520, 00:19:48.117 "method": "nvmf_create_subsystem", 00:19:48.117 "req_id": 1 00:19:48.117 } 00:19:48.117 Got JSON-RPC error response 00:19:48.117 response: 00:19:48.117 { 00:19:48.117 "code": -32602, 00:19:48.117 "message": "Invalid cntlid range [1-65520]" 00:19:48.117 } == *\I\n\v\a\l\i\d\ \c\n\t\l\i\d\ \r\a\n\g\e* ]] 00:19:48.117 17:35:29 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode32046 -i 6 -I 5 00:19:48.398 [2024-12-06 17:35:30.136512] nvmf_rpc.c: 434:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode32046: invalid cntlid range [6-5] 00:19:48.398 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@83 -- # out='request: 00:19:48.398 { 00:19:48.398 "nqn": "nqn.2016-06.io.spdk:cnode32046", 00:19:48.398 "min_cntlid": 6, 00:19:48.398 "max_cntlid": 5, 00:19:48.398 "method": "nvmf_create_subsystem", 00:19:48.398 "req_id": 1 00:19:48.398 } 00:19:48.398 Got JSON-RPC error response 00:19:48.398 response: 00:19:48.398 { 00:19:48.398 "code": -32602, 00:19:48.398 "message": "Invalid cntlid range [6-5]" 00:19:48.398 }' 00:19:48.398 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@84 -- # [[ request: 00:19:48.398 { 00:19:48.398 "nqn": "nqn.2016-06.io.spdk:cnode32046", 00:19:48.398 "min_cntlid": 6, 00:19:48.398 "max_cntlid": 5, 00:19:48.398 "method": "nvmf_create_subsystem", 00:19:48.398 "req_id": 1 00:19:48.398 } 00:19:48.399 Got JSON-RPC error response 00:19:48.399 response: 00:19:48.399 { 00:19:48.399 "code": -32602, 00:19:48.399 "message": "Invalid cntlid range [6-5]" 00:19:48.399 } == *\I\n\v\a\l\i\d\ \c\n\t\l\i\d\ \r\a\n\g\e* ]] 00:19:48.399 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@87 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_delete_target --name foobar 00:19:48.670 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@87 -- # out='request: 00:19:48.670 { 00:19:48.670 "name": "foobar", 00:19:48.670 "method": "nvmf_delete_target", 00:19:48.670 "req_id": 1 00:19:48.670 } 00:19:48.670 Got JSON-RPC error response 00:19:48.670 response: 00:19:48.670 { 00:19:48.670 "code": -32602, 00:19:48.670 "message": "The specified target doesn'\''t exist, cannot delete it." 00:19:48.670 }' 00:19:48.670 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@88 -- # [[ request: 00:19:48.670 { 00:19:48.670 "name": "foobar", 00:19:48.670 "method": "nvmf_delete_target", 00:19:48.670 "req_id": 1 00:19:48.670 } 00:19:48.670 Got JSON-RPC error response 00:19:48.670 response: 00:19:48.670 { 00:19:48.670 "code": -32602, 00:19:48.670 "message": "The specified target doesn't exist, cannot delete it." 00:19:48.670 } == *\T\h\e\ \s\p\e\c\i\f\i\e\d\ \t\a\r\g\e\t\ \d\o\e\s\n\'\t\ \e\x\i\s\t\,\ \c\a\n\n\o\t\ \d\e\l\e\t\e\ \i\t\.* ]] 00:19:48.670 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@90 -- # trap - SIGINT SIGTERM EXIT 00:19:48.670 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@91 -- # nvmftestfini 00:19:48.670 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@516 -- # nvmfcleanup 00:19:48.670 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@121 -- # sync 00:19:48.670 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:19:48.670 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@124 -- # set +e 00:19:48.670 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@125 -- # for i in {1..20} 00:19:48.670 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:19:48.670 rmmod nvme_tcp 00:19:48.670 rmmod nvme_fabrics 00:19:48.670 rmmod nvme_keyring 00:19:48.670 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:19:48.670 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@128 -- # set -e 00:19:48.670 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@129 -- # return 0 00:19:48.670 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@517 -- # '[' -n 224808 ']' 00:19:48.670 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@518 -- # killprocess 224808 00:19:48.670 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@954 -- # '[' -z 224808 ']' 00:19:48.670 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@958 -- # kill -0 224808 00:19:48.670 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@959 -- # uname 00:19:48.670 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:19:48.670 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 224808 00:19:48.670 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:19:48.670 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:19:48.670 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@972 -- # echo 'killing process with pid 224808' 00:19:48.670 killing process with pid 224808 00:19:48.670 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@973 -- # kill 224808 00:19:48.670 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@978 -- # wait 224808 00:19:48.950 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:19:48.950 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:19:48.950 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:19:48.950 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@297 -- # iptr 00:19:48.950 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@791 -- # iptables-save 00:19:48.950 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:19:48.950 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@791 -- # iptables-restore 00:19:48.950 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:19:48.950 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@302 -- # remove_spdk_ns 00:19:48.950 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:19:48.950 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:19:48.950 17:35:30 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:19:50.928 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:19:50.928 00:19:50.928 real 0m9.169s 00:19:50.928 user 0m21.784s 00:19:50.928 sys 0m2.608s 00:19:50.928 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1130 -- # xtrace_disable 00:19:50.928 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@10 -- # set +x 00:19:50.928 ************************************ 00:19:50.928 END TEST nvmf_invalid 00:19:50.928 ************************************ 00:19:50.928 17:35:32 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@24 -- # run_test nvmf_connect_stress /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_stress.sh --transport=tcp 00:19:50.928 17:35:32 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:19:50.928 17:35:32 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:19:50.928 17:35:32 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:19:50.928 ************************************ 00:19:50.928 START TEST nvmf_connect_stress 00:19:50.928 ************************************ 00:19:50.928 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_stress.sh --transport=tcp 00:19:50.928 * Looking for test storage... 00:19:50.928 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:19:50.928 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:19:50.928 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1711 -- # lcov --version 00:19:50.928 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@333 -- # local ver1 ver1_l 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@334 -- # local ver2 ver2_l 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@336 -- # IFS=.-: 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@336 -- # read -ra ver1 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@337 -- # IFS=.-: 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@337 -- # read -ra ver2 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@338 -- # local 'op=<' 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@340 -- # ver1_l=2 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@341 -- # ver2_l=1 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@344 -- # case "$op" in 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@345 -- # : 1 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@364 -- # (( v = 0 )) 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@365 -- # decimal 1 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@353 -- # local d=1 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@355 -- # echo 1 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@365 -- # ver1[v]=1 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@366 -- # decimal 2 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@353 -- # local d=2 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@355 -- # echo 2 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@366 -- # ver2[v]=2 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@368 -- # return 0 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:19:51.197 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:51.197 --rc genhtml_branch_coverage=1 00:19:51.197 --rc genhtml_function_coverage=1 00:19:51.197 --rc genhtml_legend=1 00:19:51.197 --rc geninfo_all_blocks=1 00:19:51.197 --rc geninfo_unexecuted_blocks=1 00:19:51.197 00:19:51.197 ' 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:19:51.197 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:51.197 --rc genhtml_branch_coverage=1 00:19:51.197 --rc genhtml_function_coverage=1 00:19:51.197 --rc genhtml_legend=1 00:19:51.197 --rc geninfo_all_blocks=1 00:19:51.197 --rc geninfo_unexecuted_blocks=1 00:19:51.197 00:19:51.197 ' 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:19:51.197 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:51.197 --rc genhtml_branch_coverage=1 00:19:51.197 --rc genhtml_function_coverage=1 00:19:51.197 --rc genhtml_legend=1 00:19:51.197 --rc geninfo_all_blocks=1 00:19:51.197 --rc geninfo_unexecuted_blocks=1 00:19:51.197 00:19:51.197 ' 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:19:51.197 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:51.197 --rc genhtml_branch_coverage=1 00:19:51.197 --rc genhtml_function_coverage=1 00:19:51.197 --rc genhtml_legend=1 00:19:51.197 --rc geninfo_all_blocks=1 00:19:51.197 --rc geninfo_unexecuted_blocks=1 00:19:51.197 00:19:51.197 ' 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@7 -- # uname -s 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:19:51.197 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@15 -- # shopt -s extglob 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- paths/export.sh@5 -- # export PATH 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@51 -- # : 0 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:19:51.198 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@55 -- # have_pci_nics=0 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@12 -- # nvmftestinit 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@476 -- # prepare_net_devs 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@438 -- # local -g is_hw=no 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@440 -- # remove_spdk_ns 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@309 -- # xtrace_disable 00:19:51.198 17:35:32 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@315 -- # pci_devs=() 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@315 -- # local -a pci_devs 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@316 -- # pci_net_devs=() 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@317 -- # pci_drivers=() 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@317 -- # local -A pci_drivers 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@319 -- # net_devs=() 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@319 -- # local -ga net_devs 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@320 -- # e810=() 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@320 -- # local -ga e810 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@321 -- # x722=() 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@321 -- # local -ga x722 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@322 -- # mlx=() 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@322 -- # local -ga mlx 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:19:53.251 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:19:53.251 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@418 -- # [[ up == up ]] 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:19:53.251 Found net devices under 0000:0a:00.0: cvl_0_0 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@418 -- # [[ up == up ]] 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:19:53.251 Found net devices under 0000:0a:00.1: cvl_0_1 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@442 -- # is_hw=yes 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:19:53.251 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:19:53.252 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:19:53.252 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:19:53.252 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:19:53.252 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:19:53.252 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:19:53.252 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:19:53.252 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:19:53.252 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:19:53.252 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:19:53.252 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:19:53.252 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:19:53.252 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:19:53.252 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:19:53.252 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:19:53.252 17:35:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:19:53.252 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:19:53.252 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:19:53.252 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:19:53.252 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:19:53.252 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.297 ms 00:19:53.252 00:19:53.252 --- 10.0.0.2 ping statistics --- 00:19:53.252 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:19:53.252 rtt min/avg/max/mdev = 0.297/0.297/0.297/0.000 ms 00:19:53.252 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:19:53.252 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:19:53.252 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.163 ms 00:19:53.252 00:19:53.252 --- 10.0.0.1 ping statistics --- 00:19:53.252 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:19:53.252 rtt min/avg/max/mdev = 0.163/0.163/0.163/0.000 ms 00:19:53.252 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:19:53.252 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@450 -- # return 0 00:19:53.252 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:19:53.252 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:19:53.252 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:19:53.252 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:19:53.252 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:19:53.252 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:19:53.252 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:19:53.252 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@13 -- # nvmfappstart -m 0xE 00:19:53.252 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:19:53.252 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@726 -- # xtrace_disable 00:19:53.252 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:19:53.252 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@509 -- # nvmfpid=227475 00:19:53.252 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:19:53.252 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@510 -- # waitforlisten 227475 00:19:53.252 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@835 -- # '[' -z 227475 ']' 00:19:53.252 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:53.252 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@840 -- # local max_retries=100 00:19:53.252 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:53.252 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:53.252 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@844 -- # xtrace_disable 00:19:53.252 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:19:53.531 [2024-12-06 17:35:35.103390] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:19:53.531 [2024-12-06 17:35:35.103480] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:53.531 [2024-12-06 17:35:35.177947] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:19:53.531 [2024-12-06 17:35:35.225936] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:53.531 [2024-12-06 17:35:35.226008] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:53.531 [2024-12-06 17:35:35.226036] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:53.531 [2024-12-06 17:35:35.226047] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:53.531 [2024-12-06 17:35:35.226057] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:53.531 [2024-12-06 17:35:35.227431] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:19:53.531 [2024-12-06 17:35:35.227497] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:19:53.531 [2024-12-06 17:35:35.227494] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:19:53.531 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:19:53.531 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@868 -- # return 0 00:19:53.531 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:19:53.531 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@732 -- # xtrace_disable 00:19:53.531 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:19:53.816 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:19:53.816 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@15 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:19:53.816 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:53.816 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:19:53.816 [2024-12-06 17:35:35.375604] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:19:53.816 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:53.816 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@16 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:19:53.816 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:53.816 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:19:53.816 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:53.816 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@17 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:19:53.816 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:53.816 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:19:53.816 [2024-12-06 17:35:35.392933] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:53.816 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:53.816 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@18 -- # rpc_cmd bdev_null_create NULL1 1000 512 00:19:53.816 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:53.816 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:19:53.816 NULL1 00:19:53.816 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@21 -- # PERF_PID=227623 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@23 -- # rpcs=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.txt 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@25 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.txt 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@20 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/connect_stress/connect_stress -c 0x1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' -t 10 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # seq 1 20 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:53.817 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:19:54.091 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:54.091 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:19:54.091 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:19:54.091 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:54.091 17:35:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:19:54.364 17:35:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:54.364 17:35:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:19:54.364 17:35:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:19:54.364 17:35:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:54.364 17:35:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:19:54.633 17:35:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:54.633 17:35:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:19:54.633 17:35:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:19:54.633 17:35:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:54.633 17:35:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:19:55.228 17:35:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:55.228 17:35:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:19:55.228 17:35:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:19:55.228 17:35:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:55.228 17:35:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:19:55.512 17:35:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:55.512 17:35:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:19:55.512 17:35:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:19:55.512 17:35:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:55.512 17:35:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:19:55.794 17:35:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:55.794 17:35:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:19:55.794 17:35:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:19:55.794 17:35:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:55.794 17:35:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:19:56.071 17:35:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:56.071 17:35:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:19:56.071 17:35:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:19:56.071 17:35:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:56.071 17:35:37 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:19:56.341 17:35:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:56.341 17:35:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:19:56.341 17:35:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:19:56.341 17:35:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:56.341 17:35:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:19:56.613 17:35:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:56.613 17:35:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:19:56.613 17:35:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:19:56.613 17:35:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:56.613 17:35:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:19:56.909 17:35:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:56.909 17:35:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:19:56.909 17:35:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:19:56.909 17:35:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:56.909 17:35:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:19:57.172 17:35:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:57.173 17:35:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:19:57.173 17:35:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:19:57.173 17:35:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:57.173 17:35:38 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:19:57.739 17:35:39 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:57.739 17:35:39 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:19:57.739 17:35:39 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:19:57.739 17:35:39 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:57.739 17:35:39 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:19:57.997 17:35:39 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:57.997 17:35:39 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:19:57.997 17:35:39 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:19:57.997 17:35:39 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:57.997 17:35:39 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:19:58.255 17:35:39 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:58.255 17:35:39 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:19:58.255 17:35:39 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:19:58.255 17:35:39 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:58.255 17:35:39 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:19:58.514 17:35:40 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:58.514 17:35:40 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:19:58.514 17:35:40 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:19:58.514 17:35:40 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:58.514 17:35:40 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:19:58.772 17:35:40 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:58.772 17:35:40 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:19:58.772 17:35:40 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:19:58.772 17:35:40 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:58.772 17:35:40 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:19:59.338 17:35:40 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:59.338 17:35:40 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:19:59.338 17:35:40 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:19:59.338 17:35:40 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:59.338 17:35:40 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:19:59.596 17:35:41 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:59.596 17:35:41 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:19:59.596 17:35:41 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:19:59.596 17:35:41 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:59.596 17:35:41 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:19:59.854 17:35:41 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:19:59.854 17:35:41 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:19:59.854 17:35:41 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:19:59.854 17:35:41 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:19:59.854 17:35:41 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:20:00.111 17:35:41 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:20:00.111 17:35:41 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:20:00.111 17:35:41 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:20:00.111 17:35:41 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:20:00.111 17:35:41 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:20:00.369 17:35:42 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:20:00.369 17:35:42 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:20:00.369 17:35:42 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:20:00.369 17:35:42 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:20:00.369 17:35:42 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:20:00.933 17:35:42 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:20:00.933 17:35:42 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:20:00.933 17:35:42 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:20:00.933 17:35:42 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:20:00.933 17:35:42 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:20:01.190 17:35:42 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:20:01.190 17:35:42 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:20:01.190 17:35:42 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:20:01.190 17:35:42 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:20:01.190 17:35:42 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:20:01.446 17:35:43 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:20:01.446 17:35:43 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:20:01.446 17:35:43 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:20:01.446 17:35:43 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:20:01.446 17:35:43 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:20:01.703 17:35:43 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:20:01.703 17:35:43 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:20:01.703 17:35:43 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:20:01.703 17:35:43 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:20:01.703 17:35:43 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:20:02.266 17:35:43 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:20:02.266 17:35:43 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:20:02.266 17:35:43 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:20:02.266 17:35:43 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:20:02.266 17:35:43 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:20:02.524 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:20:02.524 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:20:02.524 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:20:02.524 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:20:02.524 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:20:02.781 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:20:02.781 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:20:02.781 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:20:02.781 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:20:02.781 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:20:03.038 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:20:03.038 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:20:03.038 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:20:03.038 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:20:03.038 17:35:44 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:20:03.295 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:20:03.295 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:20:03.295 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:20:03.295 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:20:03.295 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:20:03.860 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:20:03.860 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:20:03.860 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:20:03.860 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@563 -- # xtrace_disable 00:20:03.860 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:20:03.860 Testing NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:20:04.117 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:20:04.117 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 227623 00:20:04.117 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_stress.sh: line 34: kill: (227623) - No such process 00:20:04.117 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@38 -- # wait 227623 00:20:04.117 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@39 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.txt 00:20:04.117 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@41 -- # trap - SIGINT SIGTERM EXIT 00:20:04.117 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@43 -- # nvmftestfini 00:20:04.117 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@516 -- # nvmfcleanup 00:20:04.117 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@121 -- # sync 00:20:04.117 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:20:04.117 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@124 -- # set +e 00:20:04.117 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@125 -- # for i in {1..20} 00:20:04.117 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:20:04.117 rmmod nvme_tcp 00:20:04.117 rmmod nvme_fabrics 00:20:04.117 rmmod nvme_keyring 00:20:04.117 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:20:04.117 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@128 -- # set -e 00:20:04.117 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@129 -- # return 0 00:20:04.117 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@517 -- # '[' -n 227475 ']' 00:20:04.118 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@518 -- # killprocess 227475 00:20:04.118 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@954 -- # '[' -z 227475 ']' 00:20:04.118 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@958 -- # kill -0 227475 00:20:04.118 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@959 -- # uname 00:20:04.118 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:20:04.118 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 227475 00:20:04.118 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:20:04.118 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:20:04.118 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@972 -- # echo 'killing process with pid 227475' 00:20:04.118 killing process with pid 227475 00:20:04.118 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@973 -- # kill 227475 00:20:04.118 17:35:45 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@978 -- # wait 227475 00:20:04.378 17:35:46 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:20:04.378 17:35:46 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:20:04.378 17:35:46 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:20:04.378 17:35:46 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@297 -- # iptr 00:20:04.378 17:35:46 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@791 -- # iptables-save 00:20:04.378 17:35:46 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:20:04.378 17:35:46 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@791 -- # iptables-restore 00:20:04.378 17:35:46 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:20:04.378 17:35:46 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@302 -- # remove_spdk_ns 00:20:04.378 17:35:46 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:20:04.378 17:35:46 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:20:04.378 17:35:46 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:20:06.289 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:20:06.289 00:20:06.289 real 0m15.435s 00:20:06.289 user 0m39.975s 00:20:06.289 sys 0m4.689s 00:20:06.289 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1130 -- # xtrace_disable 00:20:06.289 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:20:06.289 ************************************ 00:20:06.289 END TEST nvmf_connect_stress 00:20:06.289 ************************************ 00:20:06.289 17:35:48 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@25 -- # run_test nvmf_fused_ordering /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fused_ordering.sh --transport=tcp 00:20:06.289 17:35:48 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:20:06.289 17:35:48 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:20:06.289 17:35:48 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:20:06.549 ************************************ 00:20:06.549 START TEST nvmf_fused_ordering 00:20:06.549 ************************************ 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fused_ordering.sh --transport=tcp 00:20:06.549 * Looking for test storage... 00:20:06.549 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1711 -- # lcov --version 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@333 -- # local ver1 ver1_l 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@334 -- # local ver2 ver2_l 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@336 -- # IFS=.-: 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@336 -- # read -ra ver1 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@337 -- # IFS=.-: 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@337 -- # read -ra ver2 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@338 -- # local 'op=<' 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@340 -- # ver1_l=2 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@341 -- # ver2_l=1 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@344 -- # case "$op" in 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@345 -- # : 1 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@364 -- # (( v = 0 )) 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@365 -- # decimal 1 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@353 -- # local d=1 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@355 -- # echo 1 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@365 -- # ver1[v]=1 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@366 -- # decimal 2 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@353 -- # local d=2 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@355 -- # echo 2 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@366 -- # ver2[v]=2 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@368 -- # return 0 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:20:06.549 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:06.549 --rc genhtml_branch_coverage=1 00:20:06.549 --rc genhtml_function_coverage=1 00:20:06.549 --rc genhtml_legend=1 00:20:06.549 --rc geninfo_all_blocks=1 00:20:06.549 --rc geninfo_unexecuted_blocks=1 00:20:06.549 00:20:06.549 ' 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:20:06.549 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:06.549 --rc genhtml_branch_coverage=1 00:20:06.549 --rc genhtml_function_coverage=1 00:20:06.549 --rc genhtml_legend=1 00:20:06.549 --rc geninfo_all_blocks=1 00:20:06.549 --rc geninfo_unexecuted_blocks=1 00:20:06.549 00:20:06.549 ' 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:20:06.549 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:06.549 --rc genhtml_branch_coverage=1 00:20:06.549 --rc genhtml_function_coverage=1 00:20:06.549 --rc genhtml_legend=1 00:20:06.549 --rc geninfo_all_blocks=1 00:20:06.549 --rc geninfo_unexecuted_blocks=1 00:20:06.549 00:20:06.549 ' 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:20:06.549 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:06.549 --rc genhtml_branch_coverage=1 00:20:06.549 --rc genhtml_function_coverage=1 00:20:06.549 --rc genhtml_legend=1 00:20:06.549 --rc geninfo_all_blocks=1 00:20:06.549 --rc geninfo_unexecuted_blocks=1 00:20:06.549 00:20:06.549 ' 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@7 -- # uname -s 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@15 -- # shopt -s extglob 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:20:06.549 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:06.550 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:06.550 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:06.550 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- paths/export.sh@5 -- # export PATH 00:20:06.550 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:06.550 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@51 -- # : 0 00:20:06.550 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:20:06.550 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:20:06.550 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:20:06.550 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:20:06.550 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:20:06.550 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:20:06.550 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:20:06.550 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:20:06.550 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:20:06.550 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@55 -- # have_pci_nics=0 00:20:06.550 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@12 -- # nvmftestinit 00:20:06.550 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:20:06.550 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:20:06.550 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@476 -- # prepare_net_devs 00:20:06.550 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@438 -- # local -g is_hw=no 00:20:06.550 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@440 -- # remove_spdk_ns 00:20:06.550 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:20:06.550 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:20:06.550 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:20:06.550 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:20:06.550 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:20:06.550 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@309 -- # xtrace_disable 00:20:06.550 17:35:48 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:20:09.086 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:20:09.086 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@315 -- # pci_devs=() 00:20:09.086 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@315 -- # local -a pci_devs 00:20:09.086 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@316 -- # pci_net_devs=() 00:20:09.086 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:20:09.086 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@317 -- # pci_drivers=() 00:20:09.086 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@317 -- # local -A pci_drivers 00:20:09.086 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@319 -- # net_devs=() 00:20:09.086 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@319 -- # local -ga net_devs 00:20:09.086 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@320 -- # e810=() 00:20:09.086 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@320 -- # local -ga e810 00:20:09.086 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@321 -- # x722=() 00:20:09.086 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@321 -- # local -ga x722 00:20:09.086 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@322 -- # mlx=() 00:20:09.086 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@322 -- # local -ga mlx 00:20:09.086 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:20:09.086 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:20:09.086 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:20:09.086 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:20:09.086 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:20:09.086 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:20:09.086 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:20:09.087 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:20:09.087 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@418 -- # [[ up == up ]] 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:20:09.087 Found net devices under 0000:0a:00.0: cvl_0_0 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@418 -- # [[ up == up ]] 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:20:09.087 Found net devices under 0000:0a:00.1: cvl_0_1 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@442 -- # is_hw=yes 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:20:09.087 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:20:09.087 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.255 ms 00:20:09.087 00:20:09.087 --- 10.0.0.2 ping statistics --- 00:20:09.087 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:20:09.087 rtt min/avg/max/mdev = 0.255/0.255/0.255/0.000 ms 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:20:09.087 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:20:09.087 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.131 ms 00:20:09.087 00:20:09.087 --- 10.0.0.1 ping statistics --- 00:20:09.087 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:20:09.087 rtt min/avg/max/mdev = 0.131/0.131/0.131/0.000 ms 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@450 -- # return 0 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@13 -- # nvmfappstart -m 0x2 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@726 -- # xtrace_disable 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@509 -- # nvmfpid=230790 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@510 -- # waitforlisten 230790 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@835 -- # '[' -z 230790 ']' 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:09.087 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@840 -- # local max_retries=100 00:20:09.088 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:09.088 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:09.088 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@844 -- # xtrace_disable 00:20:09.088 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:20:09.088 [2024-12-06 17:35:50.680745] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:20:09.088 [2024-12-06 17:35:50.680848] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:20:09.088 [2024-12-06 17:35:50.754835] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:09.088 [2024-12-06 17:35:50.798661] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:20:09.088 [2024-12-06 17:35:50.798728] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:20:09.088 [2024-12-06 17:35:50.798752] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:20:09.088 [2024-12-06 17:35:50.798763] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:20:09.088 [2024-12-06 17:35:50.798772] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:20:09.088 [2024-12-06 17:35:50.799309] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:20:09.088 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:20:09.088 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@868 -- # return 0 00:20:09.088 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:20:09.088 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@732 -- # xtrace_disable 00:20:09.088 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:20:09.346 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:20:09.346 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@15 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:20:09.346 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@563 -- # xtrace_disable 00:20:09.346 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:20:09.346 [2024-12-06 17:35:50.948122] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:20:09.346 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:20:09.346 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@16 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:20:09.346 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@563 -- # xtrace_disable 00:20:09.346 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:20:09.346 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:20:09.346 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@17 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:20:09.346 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@563 -- # xtrace_disable 00:20:09.346 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:20:09.346 [2024-12-06 17:35:50.964312] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:20:09.346 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:20:09.346 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@18 -- # rpc_cmd bdev_null_create NULL1 1000 512 00:20:09.346 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@563 -- # xtrace_disable 00:20:09.347 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:20:09.347 NULL1 00:20:09.347 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:20:09.347 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@19 -- # rpc_cmd bdev_wait_for_examine 00:20:09.347 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@563 -- # xtrace_disable 00:20:09.347 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:20:09.347 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:20:09.347 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@20 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 NULL1 00:20:09.347 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@563 -- # xtrace_disable 00:20:09.347 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:20:09.347 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:20:09.347 17:35:50 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@22 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/fused_ordering/fused_ordering -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:20:09.347 [2024-12-06 17:35:51.008083] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:20:09.347 [2024-12-06 17:35:51.008116] [ DPDK EAL parameters: fused_ordering --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid230826 ] 00:20:09.913 Attached to nqn.2016-06.io.spdk:cnode1 00:20:09.913 Namespace ID: 1 size: 1GB 00:20:09.913 fused_ordering(0) 00:20:09.913 fused_ordering(1) 00:20:09.913 fused_ordering(2) 00:20:09.913 fused_ordering(3) 00:20:09.913 fused_ordering(4) 00:20:09.913 fused_ordering(5) 00:20:09.913 fused_ordering(6) 00:20:09.913 fused_ordering(7) 00:20:09.913 fused_ordering(8) 00:20:09.913 fused_ordering(9) 00:20:09.913 fused_ordering(10) 00:20:09.913 fused_ordering(11) 00:20:09.913 fused_ordering(12) 00:20:09.913 fused_ordering(13) 00:20:09.913 fused_ordering(14) 00:20:09.913 fused_ordering(15) 00:20:09.913 fused_ordering(16) 00:20:09.913 fused_ordering(17) 00:20:09.913 fused_ordering(18) 00:20:09.913 fused_ordering(19) 00:20:09.913 fused_ordering(20) 00:20:09.913 fused_ordering(21) 00:20:09.913 fused_ordering(22) 00:20:09.913 fused_ordering(23) 00:20:09.913 fused_ordering(24) 00:20:09.913 fused_ordering(25) 00:20:09.913 fused_ordering(26) 00:20:09.913 fused_ordering(27) 00:20:09.913 fused_ordering(28) 00:20:09.913 fused_ordering(29) 00:20:09.913 fused_ordering(30) 00:20:09.913 fused_ordering(31) 00:20:09.913 fused_ordering(32) 00:20:09.913 fused_ordering(33) 00:20:09.913 fused_ordering(34) 00:20:09.913 fused_ordering(35) 00:20:09.913 fused_ordering(36) 00:20:09.913 fused_ordering(37) 00:20:09.913 fused_ordering(38) 00:20:09.913 fused_ordering(39) 00:20:09.913 fused_ordering(40) 00:20:09.913 fused_ordering(41) 00:20:09.913 fused_ordering(42) 00:20:09.913 fused_ordering(43) 00:20:09.913 fused_ordering(44) 00:20:09.913 fused_ordering(45) 00:20:09.913 fused_ordering(46) 00:20:09.913 fused_ordering(47) 00:20:09.913 fused_ordering(48) 00:20:09.913 fused_ordering(49) 00:20:09.913 fused_ordering(50) 00:20:09.913 fused_ordering(51) 00:20:09.913 fused_ordering(52) 00:20:09.913 fused_ordering(53) 00:20:09.913 fused_ordering(54) 00:20:09.913 fused_ordering(55) 00:20:09.913 fused_ordering(56) 00:20:09.913 fused_ordering(57) 00:20:09.913 fused_ordering(58) 00:20:09.913 fused_ordering(59) 00:20:09.913 fused_ordering(60) 00:20:09.913 fused_ordering(61) 00:20:09.913 fused_ordering(62) 00:20:09.913 fused_ordering(63) 00:20:09.913 fused_ordering(64) 00:20:09.913 fused_ordering(65) 00:20:09.913 fused_ordering(66) 00:20:09.913 fused_ordering(67) 00:20:09.913 fused_ordering(68) 00:20:09.913 fused_ordering(69) 00:20:09.913 fused_ordering(70) 00:20:09.913 fused_ordering(71) 00:20:09.913 fused_ordering(72) 00:20:09.913 fused_ordering(73) 00:20:09.913 fused_ordering(74) 00:20:09.913 fused_ordering(75) 00:20:09.913 fused_ordering(76) 00:20:09.913 fused_ordering(77) 00:20:09.913 fused_ordering(78) 00:20:09.913 fused_ordering(79) 00:20:09.913 fused_ordering(80) 00:20:09.913 fused_ordering(81) 00:20:09.913 fused_ordering(82) 00:20:09.913 fused_ordering(83) 00:20:09.913 fused_ordering(84) 00:20:09.913 fused_ordering(85) 00:20:09.913 fused_ordering(86) 00:20:09.913 fused_ordering(87) 00:20:09.913 fused_ordering(88) 00:20:09.913 fused_ordering(89) 00:20:09.913 fused_ordering(90) 00:20:09.913 fused_ordering(91) 00:20:09.913 fused_ordering(92) 00:20:09.913 fused_ordering(93) 00:20:09.913 fused_ordering(94) 00:20:09.913 fused_ordering(95) 00:20:09.913 fused_ordering(96) 00:20:09.913 fused_ordering(97) 00:20:09.913 fused_ordering(98) 00:20:09.913 fused_ordering(99) 00:20:09.913 fused_ordering(100) 00:20:09.913 fused_ordering(101) 00:20:09.913 fused_ordering(102) 00:20:09.913 fused_ordering(103) 00:20:09.913 fused_ordering(104) 00:20:09.913 fused_ordering(105) 00:20:09.913 fused_ordering(106) 00:20:09.913 fused_ordering(107) 00:20:09.913 fused_ordering(108) 00:20:09.913 fused_ordering(109) 00:20:09.913 fused_ordering(110) 00:20:09.913 fused_ordering(111) 00:20:09.913 fused_ordering(112) 00:20:09.913 fused_ordering(113) 00:20:09.913 fused_ordering(114) 00:20:09.913 fused_ordering(115) 00:20:09.913 fused_ordering(116) 00:20:09.913 fused_ordering(117) 00:20:09.913 fused_ordering(118) 00:20:09.913 fused_ordering(119) 00:20:09.913 fused_ordering(120) 00:20:09.913 fused_ordering(121) 00:20:09.913 fused_ordering(122) 00:20:09.913 fused_ordering(123) 00:20:09.913 fused_ordering(124) 00:20:09.913 fused_ordering(125) 00:20:09.913 fused_ordering(126) 00:20:09.913 fused_ordering(127) 00:20:09.913 fused_ordering(128) 00:20:09.913 fused_ordering(129) 00:20:09.913 fused_ordering(130) 00:20:09.913 fused_ordering(131) 00:20:09.913 fused_ordering(132) 00:20:09.913 fused_ordering(133) 00:20:09.913 fused_ordering(134) 00:20:09.913 fused_ordering(135) 00:20:09.913 fused_ordering(136) 00:20:09.913 fused_ordering(137) 00:20:09.913 fused_ordering(138) 00:20:09.913 fused_ordering(139) 00:20:09.913 fused_ordering(140) 00:20:09.913 fused_ordering(141) 00:20:09.913 fused_ordering(142) 00:20:09.913 fused_ordering(143) 00:20:09.913 fused_ordering(144) 00:20:09.913 fused_ordering(145) 00:20:09.913 fused_ordering(146) 00:20:09.913 fused_ordering(147) 00:20:09.913 fused_ordering(148) 00:20:09.913 fused_ordering(149) 00:20:09.913 fused_ordering(150) 00:20:09.913 fused_ordering(151) 00:20:09.913 fused_ordering(152) 00:20:09.913 fused_ordering(153) 00:20:09.913 fused_ordering(154) 00:20:09.913 fused_ordering(155) 00:20:09.913 fused_ordering(156) 00:20:09.913 fused_ordering(157) 00:20:09.913 fused_ordering(158) 00:20:09.913 fused_ordering(159) 00:20:09.913 fused_ordering(160) 00:20:09.914 fused_ordering(161) 00:20:09.914 fused_ordering(162) 00:20:09.914 fused_ordering(163) 00:20:09.914 fused_ordering(164) 00:20:09.914 fused_ordering(165) 00:20:09.914 fused_ordering(166) 00:20:09.914 fused_ordering(167) 00:20:09.914 fused_ordering(168) 00:20:09.914 fused_ordering(169) 00:20:09.914 fused_ordering(170) 00:20:09.914 fused_ordering(171) 00:20:09.914 fused_ordering(172) 00:20:09.914 fused_ordering(173) 00:20:09.914 fused_ordering(174) 00:20:09.914 fused_ordering(175) 00:20:09.914 fused_ordering(176) 00:20:09.914 fused_ordering(177) 00:20:09.914 fused_ordering(178) 00:20:09.914 fused_ordering(179) 00:20:09.914 fused_ordering(180) 00:20:09.914 fused_ordering(181) 00:20:09.914 fused_ordering(182) 00:20:09.914 fused_ordering(183) 00:20:09.914 fused_ordering(184) 00:20:09.914 fused_ordering(185) 00:20:09.914 fused_ordering(186) 00:20:09.914 fused_ordering(187) 00:20:09.914 fused_ordering(188) 00:20:09.914 fused_ordering(189) 00:20:09.914 fused_ordering(190) 00:20:09.914 fused_ordering(191) 00:20:09.914 fused_ordering(192) 00:20:09.914 fused_ordering(193) 00:20:09.914 fused_ordering(194) 00:20:09.914 fused_ordering(195) 00:20:09.914 fused_ordering(196) 00:20:09.914 fused_ordering(197) 00:20:09.914 fused_ordering(198) 00:20:09.914 fused_ordering(199) 00:20:09.914 fused_ordering(200) 00:20:09.914 fused_ordering(201) 00:20:09.914 fused_ordering(202) 00:20:09.914 fused_ordering(203) 00:20:09.914 fused_ordering(204) 00:20:09.914 fused_ordering(205) 00:20:10.172 fused_ordering(206) 00:20:10.172 fused_ordering(207) 00:20:10.172 fused_ordering(208) 00:20:10.172 fused_ordering(209) 00:20:10.172 fused_ordering(210) 00:20:10.172 fused_ordering(211) 00:20:10.172 fused_ordering(212) 00:20:10.172 fused_ordering(213) 00:20:10.172 fused_ordering(214) 00:20:10.172 fused_ordering(215) 00:20:10.172 fused_ordering(216) 00:20:10.172 fused_ordering(217) 00:20:10.172 fused_ordering(218) 00:20:10.172 fused_ordering(219) 00:20:10.172 fused_ordering(220) 00:20:10.172 fused_ordering(221) 00:20:10.172 fused_ordering(222) 00:20:10.172 fused_ordering(223) 00:20:10.172 fused_ordering(224) 00:20:10.172 fused_ordering(225) 00:20:10.172 fused_ordering(226) 00:20:10.172 fused_ordering(227) 00:20:10.172 fused_ordering(228) 00:20:10.172 fused_ordering(229) 00:20:10.172 fused_ordering(230) 00:20:10.172 fused_ordering(231) 00:20:10.172 fused_ordering(232) 00:20:10.172 fused_ordering(233) 00:20:10.172 fused_ordering(234) 00:20:10.172 fused_ordering(235) 00:20:10.172 fused_ordering(236) 00:20:10.172 fused_ordering(237) 00:20:10.172 fused_ordering(238) 00:20:10.172 fused_ordering(239) 00:20:10.172 fused_ordering(240) 00:20:10.172 fused_ordering(241) 00:20:10.172 fused_ordering(242) 00:20:10.172 fused_ordering(243) 00:20:10.172 fused_ordering(244) 00:20:10.172 fused_ordering(245) 00:20:10.172 fused_ordering(246) 00:20:10.172 fused_ordering(247) 00:20:10.172 fused_ordering(248) 00:20:10.172 fused_ordering(249) 00:20:10.172 fused_ordering(250) 00:20:10.172 fused_ordering(251) 00:20:10.172 fused_ordering(252) 00:20:10.172 fused_ordering(253) 00:20:10.172 fused_ordering(254) 00:20:10.172 fused_ordering(255) 00:20:10.172 fused_ordering(256) 00:20:10.172 fused_ordering(257) 00:20:10.172 fused_ordering(258) 00:20:10.172 fused_ordering(259) 00:20:10.172 fused_ordering(260) 00:20:10.172 fused_ordering(261) 00:20:10.172 fused_ordering(262) 00:20:10.172 fused_ordering(263) 00:20:10.172 fused_ordering(264) 00:20:10.172 fused_ordering(265) 00:20:10.172 fused_ordering(266) 00:20:10.172 fused_ordering(267) 00:20:10.172 fused_ordering(268) 00:20:10.172 fused_ordering(269) 00:20:10.172 fused_ordering(270) 00:20:10.172 fused_ordering(271) 00:20:10.172 fused_ordering(272) 00:20:10.172 fused_ordering(273) 00:20:10.172 fused_ordering(274) 00:20:10.172 fused_ordering(275) 00:20:10.172 fused_ordering(276) 00:20:10.172 fused_ordering(277) 00:20:10.172 fused_ordering(278) 00:20:10.172 fused_ordering(279) 00:20:10.172 fused_ordering(280) 00:20:10.172 fused_ordering(281) 00:20:10.172 fused_ordering(282) 00:20:10.172 fused_ordering(283) 00:20:10.172 fused_ordering(284) 00:20:10.172 fused_ordering(285) 00:20:10.172 fused_ordering(286) 00:20:10.172 fused_ordering(287) 00:20:10.172 fused_ordering(288) 00:20:10.172 fused_ordering(289) 00:20:10.172 fused_ordering(290) 00:20:10.172 fused_ordering(291) 00:20:10.172 fused_ordering(292) 00:20:10.172 fused_ordering(293) 00:20:10.172 fused_ordering(294) 00:20:10.172 fused_ordering(295) 00:20:10.172 fused_ordering(296) 00:20:10.172 fused_ordering(297) 00:20:10.172 fused_ordering(298) 00:20:10.172 fused_ordering(299) 00:20:10.172 fused_ordering(300) 00:20:10.172 fused_ordering(301) 00:20:10.172 fused_ordering(302) 00:20:10.172 fused_ordering(303) 00:20:10.172 fused_ordering(304) 00:20:10.172 fused_ordering(305) 00:20:10.172 fused_ordering(306) 00:20:10.172 fused_ordering(307) 00:20:10.172 fused_ordering(308) 00:20:10.172 fused_ordering(309) 00:20:10.172 fused_ordering(310) 00:20:10.172 fused_ordering(311) 00:20:10.172 fused_ordering(312) 00:20:10.172 fused_ordering(313) 00:20:10.172 fused_ordering(314) 00:20:10.172 fused_ordering(315) 00:20:10.172 fused_ordering(316) 00:20:10.172 fused_ordering(317) 00:20:10.172 fused_ordering(318) 00:20:10.172 fused_ordering(319) 00:20:10.172 fused_ordering(320) 00:20:10.172 fused_ordering(321) 00:20:10.172 fused_ordering(322) 00:20:10.172 fused_ordering(323) 00:20:10.172 fused_ordering(324) 00:20:10.172 fused_ordering(325) 00:20:10.172 fused_ordering(326) 00:20:10.172 fused_ordering(327) 00:20:10.172 fused_ordering(328) 00:20:10.172 fused_ordering(329) 00:20:10.172 fused_ordering(330) 00:20:10.172 fused_ordering(331) 00:20:10.172 fused_ordering(332) 00:20:10.172 fused_ordering(333) 00:20:10.172 fused_ordering(334) 00:20:10.172 fused_ordering(335) 00:20:10.172 fused_ordering(336) 00:20:10.172 fused_ordering(337) 00:20:10.172 fused_ordering(338) 00:20:10.172 fused_ordering(339) 00:20:10.172 fused_ordering(340) 00:20:10.172 fused_ordering(341) 00:20:10.172 fused_ordering(342) 00:20:10.172 fused_ordering(343) 00:20:10.172 fused_ordering(344) 00:20:10.172 fused_ordering(345) 00:20:10.172 fused_ordering(346) 00:20:10.172 fused_ordering(347) 00:20:10.172 fused_ordering(348) 00:20:10.172 fused_ordering(349) 00:20:10.172 fused_ordering(350) 00:20:10.172 fused_ordering(351) 00:20:10.172 fused_ordering(352) 00:20:10.172 fused_ordering(353) 00:20:10.172 fused_ordering(354) 00:20:10.172 fused_ordering(355) 00:20:10.172 fused_ordering(356) 00:20:10.172 fused_ordering(357) 00:20:10.172 fused_ordering(358) 00:20:10.172 fused_ordering(359) 00:20:10.172 fused_ordering(360) 00:20:10.172 fused_ordering(361) 00:20:10.172 fused_ordering(362) 00:20:10.172 fused_ordering(363) 00:20:10.172 fused_ordering(364) 00:20:10.172 fused_ordering(365) 00:20:10.172 fused_ordering(366) 00:20:10.172 fused_ordering(367) 00:20:10.172 fused_ordering(368) 00:20:10.172 fused_ordering(369) 00:20:10.172 fused_ordering(370) 00:20:10.172 fused_ordering(371) 00:20:10.172 fused_ordering(372) 00:20:10.172 fused_ordering(373) 00:20:10.172 fused_ordering(374) 00:20:10.172 fused_ordering(375) 00:20:10.172 fused_ordering(376) 00:20:10.172 fused_ordering(377) 00:20:10.172 fused_ordering(378) 00:20:10.172 fused_ordering(379) 00:20:10.172 fused_ordering(380) 00:20:10.172 fused_ordering(381) 00:20:10.172 fused_ordering(382) 00:20:10.172 fused_ordering(383) 00:20:10.172 fused_ordering(384) 00:20:10.172 fused_ordering(385) 00:20:10.172 fused_ordering(386) 00:20:10.173 fused_ordering(387) 00:20:10.173 fused_ordering(388) 00:20:10.173 fused_ordering(389) 00:20:10.173 fused_ordering(390) 00:20:10.173 fused_ordering(391) 00:20:10.173 fused_ordering(392) 00:20:10.173 fused_ordering(393) 00:20:10.173 fused_ordering(394) 00:20:10.173 fused_ordering(395) 00:20:10.173 fused_ordering(396) 00:20:10.173 fused_ordering(397) 00:20:10.173 fused_ordering(398) 00:20:10.173 fused_ordering(399) 00:20:10.173 fused_ordering(400) 00:20:10.173 fused_ordering(401) 00:20:10.173 fused_ordering(402) 00:20:10.173 fused_ordering(403) 00:20:10.173 fused_ordering(404) 00:20:10.173 fused_ordering(405) 00:20:10.173 fused_ordering(406) 00:20:10.173 fused_ordering(407) 00:20:10.173 fused_ordering(408) 00:20:10.173 fused_ordering(409) 00:20:10.173 fused_ordering(410) 00:20:10.431 fused_ordering(411) 00:20:10.431 fused_ordering(412) 00:20:10.431 fused_ordering(413) 00:20:10.431 fused_ordering(414) 00:20:10.431 fused_ordering(415) 00:20:10.431 fused_ordering(416) 00:20:10.431 fused_ordering(417) 00:20:10.431 fused_ordering(418) 00:20:10.431 fused_ordering(419) 00:20:10.431 fused_ordering(420) 00:20:10.431 fused_ordering(421) 00:20:10.431 fused_ordering(422) 00:20:10.431 fused_ordering(423) 00:20:10.431 fused_ordering(424) 00:20:10.431 fused_ordering(425) 00:20:10.431 fused_ordering(426) 00:20:10.431 fused_ordering(427) 00:20:10.431 fused_ordering(428) 00:20:10.431 fused_ordering(429) 00:20:10.431 fused_ordering(430) 00:20:10.431 fused_ordering(431) 00:20:10.431 fused_ordering(432) 00:20:10.431 fused_ordering(433) 00:20:10.431 fused_ordering(434) 00:20:10.431 fused_ordering(435) 00:20:10.431 fused_ordering(436) 00:20:10.431 fused_ordering(437) 00:20:10.431 fused_ordering(438) 00:20:10.431 fused_ordering(439) 00:20:10.431 fused_ordering(440) 00:20:10.431 fused_ordering(441) 00:20:10.431 fused_ordering(442) 00:20:10.431 fused_ordering(443) 00:20:10.431 fused_ordering(444) 00:20:10.431 fused_ordering(445) 00:20:10.431 fused_ordering(446) 00:20:10.431 fused_ordering(447) 00:20:10.431 fused_ordering(448) 00:20:10.431 fused_ordering(449) 00:20:10.431 fused_ordering(450) 00:20:10.431 fused_ordering(451) 00:20:10.431 fused_ordering(452) 00:20:10.431 fused_ordering(453) 00:20:10.431 fused_ordering(454) 00:20:10.431 fused_ordering(455) 00:20:10.431 fused_ordering(456) 00:20:10.431 fused_ordering(457) 00:20:10.431 fused_ordering(458) 00:20:10.431 fused_ordering(459) 00:20:10.431 fused_ordering(460) 00:20:10.431 fused_ordering(461) 00:20:10.431 fused_ordering(462) 00:20:10.431 fused_ordering(463) 00:20:10.431 fused_ordering(464) 00:20:10.431 fused_ordering(465) 00:20:10.431 fused_ordering(466) 00:20:10.431 fused_ordering(467) 00:20:10.431 fused_ordering(468) 00:20:10.431 fused_ordering(469) 00:20:10.431 fused_ordering(470) 00:20:10.431 fused_ordering(471) 00:20:10.431 fused_ordering(472) 00:20:10.431 fused_ordering(473) 00:20:10.431 fused_ordering(474) 00:20:10.431 fused_ordering(475) 00:20:10.431 fused_ordering(476) 00:20:10.431 fused_ordering(477) 00:20:10.431 fused_ordering(478) 00:20:10.431 fused_ordering(479) 00:20:10.431 fused_ordering(480) 00:20:10.431 fused_ordering(481) 00:20:10.431 fused_ordering(482) 00:20:10.431 fused_ordering(483) 00:20:10.431 fused_ordering(484) 00:20:10.431 fused_ordering(485) 00:20:10.431 fused_ordering(486) 00:20:10.431 fused_ordering(487) 00:20:10.431 fused_ordering(488) 00:20:10.431 fused_ordering(489) 00:20:10.431 fused_ordering(490) 00:20:10.431 fused_ordering(491) 00:20:10.431 fused_ordering(492) 00:20:10.431 fused_ordering(493) 00:20:10.431 fused_ordering(494) 00:20:10.431 fused_ordering(495) 00:20:10.431 fused_ordering(496) 00:20:10.431 fused_ordering(497) 00:20:10.431 fused_ordering(498) 00:20:10.431 fused_ordering(499) 00:20:10.431 fused_ordering(500) 00:20:10.431 fused_ordering(501) 00:20:10.431 fused_ordering(502) 00:20:10.431 fused_ordering(503) 00:20:10.431 fused_ordering(504) 00:20:10.431 fused_ordering(505) 00:20:10.431 fused_ordering(506) 00:20:10.431 fused_ordering(507) 00:20:10.431 fused_ordering(508) 00:20:10.431 fused_ordering(509) 00:20:10.431 fused_ordering(510) 00:20:10.431 fused_ordering(511) 00:20:10.431 fused_ordering(512) 00:20:10.431 fused_ordering(513) 00:20:10.431 fused_ordering(514) 00:20:10.431 fused_ordering(515) 00:20:10.431 fused_ordering(516) 00:20:10.431 fused_ordering(517) 00:20:10.431 fused_ordering(518) 00:20:10.431 fused_ordering(519) 00:20:10.431 fused_ordering(520) 00:20:10.431 fused_ordering(521) 00:20:10.431 fused_ordering(522) 00:20:10.431 fused_ordering(523) 00:20:10.431 fused_ordering(524) 00:20:10.431 fused_ordering(525) 00:20:10.431 fused_ordering(526) 00:20:10.431 fused_ordering(527) 00:20:10.431 fused_ordering(528) 00:20:10.431 fused_ordering(529) 00:20:10.431 fused_ordering(530) 00:20:10.431 fused_ordering(531) 00:20:10.431 fused_ordering(532) 00:20:10.431 fused_ordering(533) 00:20:10.431 fused_ordering(534) 00:20:10.431 fused_ordering(535) 00:20:10.431 fused_ordering(536) 00:20:10.431 fused_ordering(537) 00:20:10.431 fused_ordering(538) 00:20:10.431 fused_ordering(539) 00:20:10.431 fused_ordering(540) 00:20:10.431 fused_ordering(541) 00:20:10.431 fused_ordering(542) 00:20:10.431 fused_ordering(543) 00:20:10.431 fused_ordering(544) 00:20:10.431 fused_ordering(545) 00:20:10.431 fused_ordering(546) 00:20:10.431 fused_ordering(547) 00:20:10.431 fused_ordering(548) 00:20:10.431 fused_ordering(549) 00:20:10.431 fused_ordering(550) 00:20:10.431 fused_ordering(551) 00:20:10.431 fused_ordering(552) 00:20:10.431 fused_ordering(553) 00:20:10.431 fused_ordering(554) 00:20:10.431 fused_ordering(555) 00:20:10.431 fused_ordering(556) 00:20:10.431 fused_ordering(557) 00:20:10.431 fused_ordering(558) 00:20:10.431 fused_ordering(559) 00:20:10.431 fused_ordering(560) 00:20:10.431 fused_ordering(561) 00:20:10.431 fused_ordering(562) 00:20:10.431 fused_ordering(563) 00:20:10.431 fused_ordering(564) 00:20:10.431 fused_ordering(565) 00:20:10.431 fused_ordering(566) 00:20:10.431 fused_ordering(567) 00:20:10.431 fused_ordering(568) 00:20:10.431 fused_ordering(569) 00:20:10.431 fused_ordering(570) 00:20:10.431 fused_ordering(571) 00:20:10.431 fused_ordering(572) 00:20:10.431 fused_ordering(573) 00:20:10.431 fused_ordering(574) 00:20:10.431 fused_ordering(575) 00:20:10.431 fused_ordering(576) 00:20:10.431 fused_ordering(577) 00:20:10.431 fused_ordering(578) 00:20:10.431 fused_ordering(579) 00:20:10.431 fused_ordering(580) 00:20:10.432 fused_ordering(581) 00:20:10.432 fused_ordering(582) 00:20:10.432 fused_ordering(583) 00:20:10.432 fused_ordering(584) 00:20:10.432 fused_ordering(585) 00:20:10.432 fused_ordering(586) 00:20:10.432 fused_ordering(587) 00:20:10.432 fused_ordering(588) 00:20:10.432 fused_ordering(589) 00:20:10.432 fused_ordering(590) 00:20:10.432 fused_ordering(591) 00:20:10.432 fused_ordering(592) 00:20:10.432 fused_ordering(593) 00:20:10.432 fused_ordering(594) 00:20:10.432 fused_ordering(595) 00:20:10.432 fused_ordering(596) 00:20:10.432 fused_ordering(597) 00:20:10.432 fused_ordering(598) 00:20:10.432 fused_ordering(599) 00:20:10.432 fused_ordering(600) 00:20:10.432 fused_ordering(601) 00:20:10.432 fused_ordering(602) 00:20:10.432 fused_ordering(603) 00:20:10.432 fused_ordering(604) 00:20:10.432 fused_ordering(605) 00:20:10.432 fused_ordering(606) 00:20:10.432 fused_ordering(607) 00:20:10.432 fused_ordering(608) 00:20:10.432 fused_ordering(609) 00:20:10.432 fused_ordering(610) 00:20:10.432 fused_ordering(611) 00:20:10.432 fused_ordering(612) 00:20:10.432 fused_ordering(613) 00:20:10.432 fused_ordering(614) 00:20:10.432 fused_ordering(615) 00:20:10.998 fused_ordering(616) 00:20:10.998 fused_ordering(617) 00:20:10.998 fused_ordering(618) 00:20:10.998 fused_ordering(619) 00:20:10.998 fused_ordering(620) 00:20:10.998 fused_ordering(621) 00:20:10.998 fused_ordering(622) 00:20:10.998 fused_ordering(623) 00:20:10.998 fused_ordering(624) 00:20:10.998 fused_ordering(625) 00:20:10.998 fused_ordering(626) 00:20:10.998 fused_ordering(627) 00:20:10.998 fused_ordering(628) 00:20:10.998 fused_ordering(629) 00:20:10.998 fused_ordering(630) 00:20:10.998 fused_ordering(631) 00:20:10.998 fused_ordering(632) 00:20:10.998 fused_ordering(633) 00:20:10.998 fused_ordering(634) 00:20:10.998 fused_ordering(635) 00:20:10.998 fused_ordering(636) 00:20:10.998 fused_ordering(637) 00:20:10.998 fused_ordering(638) 00:20:10.998 fused_ordering(639) 00:20:10.998 fused_ordering(640) 00:20:10.998 fused_ordering(641) 00:20:10.998 fused_ordering(642) 00:20:10.998 fused_ordering(643) 00:20:10.998 fused_ordering(644) 00:20:10.998 fused_ordering(645) 00:20:10.998 fused_ordering(646) 00:20:10.998 fused_ordering(647) 00:20:10.998 fused_ordering(648) 00:20:10.998 fused_ordering(649) 00:20:10.998 fused_ordering(650) 00:20:10.998 fused_ordering(651) 00:20:10.998 fused_ordering(652) 00:20:10.998 fused_ordering(653) 00:20:10.998 fused_ordering(654) 00:20:10.998 fused_ordering(655) 00:20:10.998 fused_ordering(656) 00:20:10.998 fused_ordering(657) 00:20:10.998 fused_ordering(658) 00:20:10.998 fused_ordering(659) 00:20:10.998 fused_ordering(660) 00:20:10.998 fused_ordering(661) 00:20:10.998 fused_ordering(662) 00:20:10.998 fused_ordering(663) 00:20:10.998 fused_ordering(664) 00:20:10.998 fused_ordering(665) 00:20:10.998 fused_ordering(666) 00:20:10.998 fused_ordering(667) 00:20:10.998 fused_ordering(668) 00:20:10.998 fused_ordering(669) 00:20:10.998 fused_ordering(670) 00:20:10.998 fused_ordering(671) 00:20:10.998 fused_ordering(672) 00:20:10.998 fused_ordering(673) 00:20:10.998 fused_ordering(674) 00:20:10.998 fused_ordering(675) 00:20:10.998 fused_ordering(676) 00:20:10.998 fused_ordering(677) 00:20:10.998 fused_ordering(678) 00:20:10.998 fused_ordering(679) 00:20:10.998 fused_ordering(680) 00:20:10.998 fused_ordering(681) 00:20:10.998 fused_ordering(682) 00:20:10.998 fused_ordering(683) 00:20:10.998 fused_ordering(684) 00:20:10.998 fused_ordering(685) 00:20:10.998 fused_ordering(686) 00:20:10.998 fused_ordering(687) 00:20:10.998 fused_ordering(688) 00:20:10.998 fused_ordering(689) 00:20:10.998 fused_ordering(690) 00:20:10.998 fused_ordering(691) 00:20:10.998 fused_ordering(692) 00:20:10.998 fused_ordering(693) 00:20:10.998 fused_ordering(694) 00:20:10.998 fused_ordering(695) 00:20:10.998 fused_ordering(696) 00:20:10.998 fused_ordering(697) 00:20:10.998 fused_ordering(698) 00:20:10.998 fused_ordering(699) 00:20:10.998 fused_ordering(700) 00:20:10.998 fused_ordering(701) 00:20:10.998 fused_ordering(702) 00:20:10.998 fused_ordering(703) 00:20:10.998 fused_ordering(704) 00:20:10.998 fused_ordering(705) 00:20:10.998 fused_ordering(706) 00:20:10.998 fused_ordering(707) 00:20:10.998 fused_ordering(708) 00:20:10.998 fused_ordering(709) 00:20:10.998 fused_ordering(710) 00:20:10.998 fused_ordering(711) 00:20:10.998 fused_ordering(712) 00:20:10.998 fused_ordering(713) 00:20:10.998 fused_ordering(714) 00:20:10.998 fused_ordering(715) 00:20:10.998 fused_ordering(716) 00:20:10.998 fused_ordering(717) 00:20:10.998 fused_ordering(718) 00:20:10.998 fused_ordering(719) 00:20:10.998 fused_ordering(720) 00:20:10.998 fused_ordering(721) 00:20:10.998 fused_ordering(722) 00:20:10.998 fused_ordering(723) 00:20:10.998 fused_ordering(724) 00:20:10.998 fused_ordering(725) 00:20:10.998 fused_ordering(726) 00:20:10.998 fused_ordering(727) 00:20:10.998 fused_ordering(728) 00:20:10.998 fused_ordering(729) 00:20:10.998 fused_ordering(730) 00:20:10.998 fused_ordering(731) 00:20:10.998 fused_ordering(732) 00:20:10.998 fused_ordering(733) 00:20:10.998 fused_ordering(734) 00:20:10.998 fused_ordering(735) 00:20:10.998 fused_ordering(736) 00:20:10.998 fused_ordering(737) 00:20:10.998 fused_ordering(738) 00:20:10.998 fused_ordering(739) 00:20:10.998 fused_ordering(740) 00:20:10.998 fused_ordering(741) 00:20:10.998 fused_ordering(742) 00:20:10.998 fused_ordering(743) 00:20:10.998 fused_ordering(744) 00:20:10.998 fused_ordering(745) 00:20:10.998 fused_ordering(746) 00:20:10.998 fused_ordering(747) 00:20:10.998 fused_ordering(748) 00:20:10.998 fused_ordering(749) 00:20:10.998 fused_ordering(750) 00:20:10.998 fused_ordering(751) 00:20:10.998 fused_ordering(752) 00:20:10.998 fused_ordering(753) 00:20:10.998 fused_ordering(754) 00:20:10.998 fused_ordering(755) 00:20:10.998 fused_ordering(756) 00:20:10.998 fused_ordering(757) 00:20:10.998 fused_ordering(758) 00:20:10.998 fused_ordering(759) 00:20:10.998 fused_ordering(760) 00:20:10.998 fused_ordering(761) 00:20:10.998 fused_ordering(762) 00:20:10.998 fused_ordering(763) 00:20:10.998 fused_ordering(764) 00:20:10.998 fused_ordering(765) 00:20:10.998 fused_ordering(766) 00:20:10.998 fused_ordering(767) 00:20:10.998 fused_ordering(768) 00:20:10.998 fused_ordering(769) 00:20:10.998 fused_ordering(770) 00:20:10.998 fused_ordering(771) 00:20:10.998 fused_ordering(772) 00:20:10.998 fused_ordering(773) 00:20:10.998 fused_ordering(774) 00:20:10.998 fused_ordering(775) 00:20:10.998 fused_ordering(776) 00:20:10.998 fused_ordering(777) 00:20:10.998 fused_ordering(778) 00:20:10.998 fused_ordering(779) 00:20:10.998 fused_ordering(780) 00:20:10.998 fused_ordering(781) 00:20:10.998 fused_ordering(782) 00:20:10.998 fused_ordering(783) 00:20:10.998 fused_ordering(784) 00:20:10.998 fused_ordering(785) 00:20:10.998 fused_ordering(786) 00:20:10.998 fused_ordering(787) 00:20:10.998 fused_ordering(788) 00:20:10.998 fused_ordering(789) 00:20:10.998 fused_ordering(790) 00:20:10.999 fused_ordering(791) 00:20:10.999 fused_ordering(792) 00:20:10.999 fused_ordering(793) 00:20:10.999 fused_ordering(794) 00:20:10.999 fused_ordering(795) 00:20:10.999 fused_ordering(796) 00:20:10.999 fused_ordering(797) 00:20:10.999 fused_ordering(798) 00:20:10.999 fused_ordering(799) 00:20:10.999 fused_ordering(800) 00:20:10.999 fused_ordering(801) 00:20:10.999 fused_ordering(802) 00:20:10.999 fused_ordering(803) 00:20:10.999 fused_ordering(804) 00:20:10.999 fused_ordering(805) 00:20:10.999 fused_ordering(806) 00:20:10.999 fused_ordering(807) 00:20:10.999 fused_ordering(808) 00:20:10.999 fused_ordering(809) 00:20:10.999 fused_ordering(810) 00:20:10.999 fused_ordering(811) 00:20:10.999 fused_ordering(812) 00:20:10.999 fused_ordering(813) 00:20:10.999 fused_ordering(814) 00:20:10.999 fused_ordering(815) 00:20:10.999 fused_ordering(816) 00:20:10.999 fused_ordering(817) 00:20:10.999 fused_ordering(818) 00:20:10.999 fused_ordering(819) 00:20:10.999 fused_ordering(820) 00:20:11.565 fused_ordering(821) 00:20:11.565 fused_ordering(822) 00:20:11.565 fused_ordering(823) 00:20:11.565 fused_ordering(824) 00:20:11.565 fused_ordering(825) 00:20:11.565 fused_ordering(826) 00:20:11.565 fused_ordering(827) 00:20:11.565 fused_ordering(828) 00:20:11.565 fused_ordering(829) 00:20:11.565 fused_ordering(830) 00:20:11.565 fused_ordering(831) 00:20:11.565 fused_ordering(832) 00:20:11.565 fused_ordering(833) 00:20:11.565 fused_ordering(834) 00:20:11.565 fused_ordering(835) 00:20:11.565 fused_ordering(836) 00:20:11.565 fused_ordering(837) 00:20:11.565 fused_ordering(838) 00:20:11.565 fused_ordering(839) 00:20:11.565 fused_ordering(840) 00:20:11.565 fused_ordering(841) 00:20:11.565 fused_ordering(842) 00:20:11.565 fused_ordering(843) 00:20:11.565 fused_ordering(844) 00:20:11.565 fused_ordering(845) 00:20:11.565 fused_ordering(846) 00:20:11.565 fused_ordering(847) 00:20:11.565 fused_ordering(848) 00:20:11.565 fused_ordering(849) 00:20:11.565 fused_ordering(850) 00:20:11.565 fused_ordering(851) 00:20:11.565 fused_ordering(852) 00:20:11.565 fused_ordering(853) 00:20:11.565 fused_ordering(854) 00:20:11.565 fused_ordering(855) 00:20:11.565 fused_ordering(856) 00:20:11.565 fused_ordering(857) 00:20:11.565 fused_ordering(858) 00:20:11.565 fused_ordering(859) 00:20:11.565 fused_ordering(860) 00:20:11.565 fused_ordering(861) 00:20:11.565 fused_ordering(862) 00:20:11.565 fused_ordering(863) 00:20:11.565 fused_ordering(864) 00:20:11.565 fused_ordering(865) 00:20:11.565 fused_ordering(866) 00:20:11.565 fused_ordering(867) 00:20:11.565 fused_ordering(868) 00:20:11.565 fused_ordering(869) 00:20:11.565 fused_ordering(870) 00:20:11.565 fused_ordering(871) 00:20:11.565 fused_ordering(872) 00:20:11.565 fused_ordering(873) 00:20:11.565 fused_ordering(874) 00:20:11.565 fused_ordering(875) 00:20:11.565 fused_ordering(876) 00:20:11.565 fused_ordering(877) 00:20:11.565 fused_ordering(878) 00:20:11.565 fused_ordering(879) 00:20:11.565 fused_ordering(880) 00:20:11.565 fused_ordering(881) 00:20:11.565 fused_ordering(882) 00:20:11.565 fused_ordering(883) 00:20:11.565 fused_ordering(884) 00:20:11.565 fused_ordering(885) 00:20:11.565 fused_ordering(886) 00:20:11.565 fused_ordering(887) 00:20:11.565 fused_ordering(888) 00:20:11.565 fused_ordering(889) 00:20:11.565 fused_ordering(890) 00:20:11.565 fused_ordering(891) 00:20:11.565 fused_ordering(892) 00:20:11.565 fused_ordering(893) 00:20:11.565 fused_ordering(894) 00:20:11.565 fused_ordering(895) 00:20:11.565 fused_ordering(896) 00:20:11.565 fused_ordering(897) 00:20:11.565 fused_ordering(898) 00:20:11.565 fused_ordering(899) 00:20:11.565 fused_ordering(900) 00:20:11.565 fused_ordering(901) 00:20:11.565 fused_ordering(902) 00:20:11.565 fused_ordering(903) 00:20:11.565 fused_ordering(904) 00:20:11.565 fused_ordering(905) 00:20:11.565 fused_ordering(906) 00:20:11.565 fused_ordering(907) 00:20:11.565 fused_ordering(908) 00:20:11.565 fused_ordering(909) 00:20:11.565 fused_ordering(910) 00:20:11.565 fused_ordering(911) 00:20:11.565 fused_ordering(912) 00:20:11.565 fused_ordering(913) 00:20:11.565 fused_ordering(914) 00:20:11.565 fused_ordering(915) 00:20:11.565 fused_ordering(916) 00:20:11.565 fused_ordering(917) 00:20:11.565 fused_ordering(918) 00:20:11.565 fused_ordering(919) 00:20:11.565 fused_ordering(920) 00:20:11.565 fused_ordering(921) 00:20:11.565 fused_ordering(922) 00:20:11.565 fused_ordering(923) 00:20:11.565 fused_ordering(924) 00:20:11.565 fused_ordering(925) 00:20:11.565 fused_ordering(926) 00:20:11.565 fused_ordering(927) 00:20:11.565 fused_ordering(928) 00:20:11.565 fused_ordering(929) 00:20:11.565 fused_ordering(930) 00:20:11.565 fused_ordering(931) 00:20:11.565 fused_ordering(932) 00:20:11.565 fused_ordering(933) 00:20:11.565 fused_ordering(934) 00:20:11.565 fused_ordering(935) 00:20:11.565 fused_ordering(936) 00:20:11.565 fused_ordering(937) 00:20:11.565 fused_ordering(938) 00:20:11.565 fused_ordering(939) 00:20:11.565 fused_ordering(940) 00:20:11.565 fused_ordering(941) 00:20:11.565 fused_ordering(942) 00:20:11.565 fused_ordering(943) 00:20:11.565 fused_ordering(944) 00:20:11.565 fused_ordering(945) 00:20:11.565 fused_ordering(946) 00:20:11.565 fused_ordering(947) 00:20:11.565 fused_ordering(948) 00:20:11.565 fused_ordering(949) 00:20:11.565 fused_ordering(950) 00:20:11.565 fused_ordering(951) 00:20:11.565 fused_ordering(952) 00:20:11.565 fused_ordering(953) 00:20:11.565 fused_ordering(954) 00:20:11.565 fused_ordering(955) 00:20:11.565 fused_ordering(956) 00:20:11.565 fused_ordering(957) 00:20:11.565 fused_ordering(958) 00:20:11.565 fused_ordering(959) 00:20:11.565 fused_ordering(960) 00:20:11.565 fused_ordering(961) 00:20:11.565 fused_ordering(962) 00:20:11.565 fused_ordering(963) 00:20:11.565 fused_ordering(964) 00:20:11.565 fused_ordering(965) 00:20:11.565 fused_ordering(966) 00:20:11.565 fused_ordering(967) 00:20:11.565 fused_ordering(968) 00:20:11.565 fused_ordering(969) 00:20:11.565 fused_ordering(970) 00:20:11.565 fused_ordering(971) 00:20:11.565 fused_ordering(972) 00:20:11.565 fused_ordering(973) 00:20:11.565 fused_ordering(974) 00:20:11.565 fused_ordering(975) 00:20:11.565 fused_ordering(976) 00:20:11.565 fused_ordering(977) 00:20:11.565 fused_ordering(978) 00:20:11.565 fused_ordering(979) 00:20:11.565 fused_ordering(980) 00:20:11.565 fused_ordering(981) 00:20:11.565 fused_ordering(982) 00:20:11.565 fused_ordering(983) 00:20:11.565 fused_ordering(984) 00:20:11.565 fused_ordering(985) 00:20:11.565 fused_ordering(986) 00:20:11.565 fused_ordering(987) 00:20:11.565 fused_ordering(988) 00:20:11.565 fused_ordering(989) 00:20:11.565 fused_ordering(990) 00:20:11.565 fused_ordering(991) 00:20:11.565 fused_ordering(992) 00:20:11.565 fused_ordering(993) 00:20:11.565 fused_ordering(994) 00:20:11.565 fused_ordering(995) 00:20:11.565 fused_ordering(996) 00:20:11.565 fused_ordering(997) 00:20:11.565 fused_ordering(998) 00:20:11.565 fused_ordering(999) 00:20:11.565 fused_ordering(1000) 00:20:11.565 fused_ordering(1001) 00:20:11.565 fused_ordering(1002) 00:20:11.566 fused_ordering(1003) 00:20:11.566 fused_ordering(1004) 00:20:11.566 fused_ordering(1005) 00:20:11.566 fused_ordering(1006) 00:20:11.566 fused_ordering(1007) 00:20:11.566 fused_ordering(1008) 00:20:11.566 fused_ordering(1009) 00:20:11.566 fused_ordering(1010) 00:20:11.566 fused_ordering(1011) 00:20:11.566 fused_ordering(1012) 00:20:11.566 fused_ordering(1013) 00:20:11.566 fused_ordering(1014) 00:20:11.566 fused_ordering(1015) 00:20:11.566 fused_ordering(1016) 00:20:11.566 fused_ordering(1017) 00:20:11.566 fused_ordering(1018) 00:20:11.566 fused_ordering(1019) 00:20:11.566 fused_ordering(1020) 00:20:11.566 fused_ordering(1021) 00:20:11.566 fused_ordering(1022) 00:20:11.566 fused_ordering(1023) 00:20:11.566 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@23 -- # trap - SIGINT SIGTERM EXIT 00:20:11.566 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@25 -- # nvmftestfini 00:20:11.566 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@516 -- # nvmfcleanup 00:20:11.566 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@121 -- # sync 00:20:11.566 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:20:11.566 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@124 -- # set +e 00:20:11.566 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@125 -- # for i in {1..20} 00:20:11.566 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:20:11.566 rmmod nvme_tcp 00:20:11.566 rmmod nvme_fabrics 00:20:11.566 rmmod nvme_keyring 00:20:11.566 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:20:11.566 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@128 -- # set -e 00:20:11.566 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@129 -- # return 0 00:20:11.566 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@517 -- # '[' -n 230790 ']' 00:20:11.566 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@518 -- # killprocess 230790 00:20:11.566 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@954 -- # '[' -z 230790 ']' 00:20:11.566 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@958 -- # kill -0 230790 00:20:11.566 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@959 -- # uname 00:20:11.566 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:20:11.566 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 230790 00:20:11.566 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:20:11.566 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:20:11.566 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@972 -- # echo 'killing process with pid 230790' 00:20:11.566 killing process with pid 230790 00:20:11.566 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@973 -- # kill 230790 00:20:11.566 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@978 -- # wait 230790 00:20:11.824 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:20:11.824 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:20:11.824 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:20:11.824 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@297 -- # iptr 00:20:11.824 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@791 -- # iptables-save 00:20:11.824 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:20:11.824 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@791 -- # iptables-restore 00:20:11.824 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:20:11.824 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@302 -- # remove_spdk_ns 00:20:11.825 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:20:11.825 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:20:11.825 17:35:53 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:20:14.361 00:20:14.361 real 0m7.437s 00:20:14.361 user 0m5.108s 00:20:14.361 sys 0m2.943s 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1130 -- # xtrace_disable 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:20:14.361 ************************************ 00:20:14.361 END TEST nvmf_fused_ordering 00:20:14.361 ************************************ 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@26 -- # run_test nvmf_ns_masking test/nvmf/target/ns_masking.sh --transport=tcp 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:20:14.361 ************************************ 00:20:14.361 START TEST nvmf_ns_masking 00:20:14.361 ************************************ 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1129 -- # test/nvmf/target/ns_masking.sh --transport=tcp 00:20:14.361 * Looking for test storage... 00:20:14.361 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1711 -- # lcov --version 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@333 -- # local ver1 ver1_l 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@334 -- # local ver2 ver2_l 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@336 -- # IFS=.-: 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@336 -- # read -ra ver1 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@337 -- # IFS=.-: 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@337 -- # read -ra ver2 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@338 -- # local 'op=<' 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@340 -- # ver1_l=2 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@341 -- # ver2_l=1 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@344 -- # case "$op" in 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@345 -- # : 1 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@364 -- # (( v = 0 )) 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@365 -- # decimal 1 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@353 -- # local d=1 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@355 -- # echo 1 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@365 -- # ver1[v]=1 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@366 -- # decimal 2 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@353 -- # local d=2 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@355 -- # echo 2 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@366 -- # ver2[v]=2 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@368 -- # return 0 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:20:14.361 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:14.361 --rc genhtml_branch_coverage=1 00:20:14.361 --rc genhtml_function_coverage=1 00:20:14.361 --rc genhtml_legend=1 00:20:14.361 --rc geninfo_all_blocks=1 00:20:14.361 --rc geninfo_unexecuted_blocks=1 00:20:14.361 00:20:14.361 ' 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:20:14.361 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:14.361 --rc genhtml_branch_coverage=1 00:20:14.361 --rc genhtml_function_coverage=1 00:20:14.361 --rc genhtml_legend=1 00:20:14.361 --rc geninfo_all_blocks=1 00:20:14.361 --rc geninfo_unexecuted_blocks=1 00:20:14.361 00:20:14.361 ' 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:20:14.361 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:14.361 --rc genhtml_branch_coverage=1 00:20:14.361 --rc genhtml_function_coverage=1 00:20:14.361 --rc genhtml_legend=1 00:20:14.361 --rc geninfo_all_blocks=1 00:20:14.361 --rc geninfo_unexecuted_blocks=1 00:20:14.361 00:20:14.361 ' 00:20:14.361 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:20:14.361 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:14.362 --rc genhtml_branch_coverage=1 00:20:14.362 --rc genhtml_function_coverage=1 00:20:14.362 --rc genhtml_legend=1 00:20:14.362 --rc geninfo_all_blocks=1 00:20:14.362 --rc geninfo_unexecuted_blocks=1 00:20:14.362 00:20:14.362 ' 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@8 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@7 -- # uname -s 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@15 -- # shopt -s extglob 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- paths/export.sh@5 -- # export PATH 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@51 -- # : 0 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:20:14.362 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@55 -- # have_pci_nics=0 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@10 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@11 -- # hostsock=/var/tmp/host.sock 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@12 -- # loops=5 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@13 -- # uuidgen 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@13 -- # ns1uuid=08a4d33a-f0ed-4528-baec-ffb64069a77c 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@14 -- # uuidgen 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@14 -- # ns2uuid=957cadae-2228-4120-bdf5-2f5e983b7b3c 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@16 -- # SUBSYSNQN=nqn.2016-06.io.spdk:cnode1 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@17 -- # HOSTNQN1=nqn.2016-06.io.spdk:host1 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@18 -- # HOSTNQN2=nqn.2016-06.io.spdk:host2 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@19 -- # uuidgen 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@19 -- # HOSTID=70b65a8f-dc2d-4e3c-a332-3d0ff2d08a3a 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@50 -- # nvmftestinit 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@476 -- # prepare_net_devs 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@438 -- # local -g is_hw=no 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@440 -- # remove_spdk_ns 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@309 -- # xtrace_disable 00:20:14.362 17:35:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:20:16.268 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:20:16.268 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@315 -- # pci_devs=() 00:20:16.268 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@315 -- # local -a pci_devs 00:20:16.268 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@316 -- # pci_net_devs=() 00:20:16.268 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:20:16.268 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@317 -- # pci_drivers=() 00:20:16.268 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@317 -- # local -A pci_drivers 00:20:16.268 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@319 -- # net_devs=() 00:20:16.268 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@319 -- # local -ga net_devs 00:20:16.268 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@320 -- # e810=() 00:20:16.268 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@320 -- # local -ga e810 00:20:16.268 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@321 -- # x722=() 00:20:16.268 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@321 -- # local -ga x722 00:20:16.268 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@322 -- # mlx=() 00:20:16.268 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@322 -- # local -ga mlx 00:20:16.268 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:20:16.268 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:20:16.268 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:20:16.269 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:20:16.269 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@418 -- # [[ up == up ]] 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:20:16.269 Found net devices under 0000:0a:00.0: cvl_0_0 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@418 -- # [[ up == up ]] 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:20:16.269 Found net devices under 0000:0a:00.1: cvl_0_1 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@442 -- # is_hw=yes 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:20:16.269 17:35:57 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:20:16.269 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:20:16.269 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:20:16.269 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:20:16.269 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:20:16.269 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:20:16.269 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:20:16.269 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:20:16.269 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:20:16.269 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:20:16.269 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.331 ms 00:20:16.269 00:20:16.269 --- 10.0.0.2 ping statistics --- 00:20:16.269 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:20:16.269 rtt min/avg/max/mdev = 0.331/0.331/0.331/0.000 ms 00:20:16.269 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:20:16.269 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:20:16.269 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.139 ms 00:20:16.269 00:20:16.269 --- 10.0.0.1 ping statistics --- 00:20:16.269 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:20:16.269 rtt min/avg/max/mdev = 0.139/0.139/0.139/0.000 ms 00:20:16.269 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:20:16.269 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@450 -- # return 0 00:20:16.269 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:20:16.269 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:20:16.269 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:20:16.269 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:20:16.269 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:20:16.269 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:20:16.269 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:20:16.528 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@51 -- # nvmfappstart 00:20:16.528 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:20:16.528 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@726 -- # xtrace_disable 00:20:16.528 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:20:16.528 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@509 -- # nvmfpid=233138 00:20:16.528 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@510 -- # waitforlisten 233138 00:20:16.528 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF 00:20:16.528 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@835 -- # '[' -z 233138 ']' 00:20:16.528 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:16.528 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@840 -- # local max_retries=100 00:20:16.528 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:16.528 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:16.528 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@844 -- # xtrace_disable 00:20:16.528 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:20:16.528 [2024-12-06 17:35:58.173078] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:20:16.528 [2024-12-06 17:35:58.173163] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:20:16.528 [2024-12-06 17:35:58.245324] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:16.528 [2024-12-06 17:35:58.293048] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:20:16.528 [2024-12-06 17:35:58.293111] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:20:16.528 [2024-12-06 17:35:58.293124] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:20:16.528 [2024-12-06 17:35:58.293134] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:20:16.528 [2024-12-06 17:35:58.293143] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:20:16.528 [2024-12-06 17:35:58.293744] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:20:16.787 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:20:16.787 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@868 -- # return 0 00:20:16.787 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:20:16.787 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@732 -- # xtrace_disable 00:20:16.787 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:20:16.787 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:20:16.787 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:20:17.045 [2024-12-06 17:35:58.743008] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:20:17.045 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@55 -- # MALLOC_BDEV_SIZE=64 00:20:17.045 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@56 -- # MALLOC_BLOCK_SIZE=512 00:20:17.045 17:35:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc1 00:20:17.304 Malloc1 00:20:17.304 17:35:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc2 00:20:17.562 Malloc2 00:20:17.562 17:35:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@62 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:20:17.820 17:35:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 1 00:20:18.386 17:35:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:20:18.643 [2024-12-06 17:36:00.257224] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:20:18.643 17:36:00 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@67 -- # connect 00:20:18.643 17:36:00 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@22 -- # nvme connect -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 -I 70b65a8f-dc2d-4e3c-a332-3d0ff2d08a3a -a 10.0.0.2 -s 4420 -i 4 00:20:18.900 17:36:00 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@24 -- # waitforserial SPDKISFASTANDAWESOME 00:20:18.900 17:36:00 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1202 -- # local i=0 00:20:18.900 17:36:00 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:20:18.900 17:36:00 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:20:18.900 17:36:00 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1209 -- # sleep 2 00:20:20.798 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:20:20.798 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:20:20.798 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:20:20.798 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:20:20.798 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:20:20.798 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1212 -- # return 0 00:20:20.798 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # nvme list-subsys -o json 00:20:20.798 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # jq -r '.[].Subsystems[] | select(.NQN=="nqn.2016-06.io.spdk:cnode1") | .Paths[0].Name' 00:20:20.798 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # ctrl_id=nvme0 00:20:20.798 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@27 -- # [[ -z nvme0 ]] 00:20:20.798 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@68 -- # ns_is_visible 0x1 00:20:20.798 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:20:20.798 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:20:20.798 [ 0]:0x1 00:20:20.798 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:20:20.798 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:20:21.056 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=459b07a705eb4b35a27dabff97462bcc 00:20:21.056 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 459b07a705eb4b35a27dabff97462bcc != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:20:21.056 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc2 -n 2 00:20:21.314 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@72 -- # ns_is_visible 0x1 00:20:21.314 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:20:21.314 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:20:21.314 [ 0]:0x1 00:20:21.314 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:20:21.314 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:20:21.314 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=459b07a705eb4b35a27dabff97462bcc 00:20:21.314 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 459b07a705eb4b35a27dabff97462bcc != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:20:21.314 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@73 -- # ns_is_visible 0x2 00:20:21.314 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:20:21.314 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:20:21.314 [ 1]:0x2 00:20:21.314 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:20:21.314 17:36:02 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:20:21.314 17:36:03 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=b881f9b1cabf486c8f254f23f54c5f63 00:20:21.314 17:36:03 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ b881f9b1cabf486c8f254f23f54c5f63 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:20:21.314 17:36:03 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@75 -- # disconnect 00:20:21.314 17:36:03 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@38 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:20:21.314 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:20:21.314 17:36:03 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:20:21.880 17:36:03 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@80 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 1 --no-auto-visible 00:20:22.138 17:36:03 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@83 -- # connect 1 00:20:22.138 17:36:03 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@22 -- # nvme connect -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 -I 70b65a8f-dc2d-4e3c-a332-3d0ff2d08a3a -a 10.0.0.2 -s 4420 -i 4 00:20:22.138 17:36:03 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@24 -- # waitforserial SPDKISFASTANDAWESOME 1 00:20:22.138 17:36:03 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1202 -- # local i=0 00:20:22.138 17:36:03 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:20:22.138 17:36:03 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1204 -- # [[ -n 1 ]] 00:20:22.138 17:36:03 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1205 -- # nvme_device_counter=1 00:20:22.138 17:36:03 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1209 -- # sleep 2 00:20:24.036 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:20:24.036 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:20:24.036 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:20:24.036 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:20:24.036 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:20:24.036 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1212 -- # return 0 00:20:24.036 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # nvme list-subsys -o json 00:20:24.036 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # jq -r '.[].Subsystems[] | select(.NQN=="nqn.2016-06.io.spdk:cnode1") | .Paths[0].Name' 00:20:24.294 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # ctrl_id=nvme0 00:20:24.294 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@27 -- # [[ -z nvme0 ]] 00:20:24.294 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@84 -- # NOT ns_is_visible 0x1 00:20:24.294 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@652 -- # local es=0 00:20:24.294 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@654 -- # valid_exec_arg ns_is_visible 0x1 00:20:24.294 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@640 -- # local arg=ns_is_visible 00:20:24.294 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:20:24.294 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # type -t ns_is_visible 00:20:24.294 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:20:24.294 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@655 -- # ns_is_visible 0x1 00:20:24.294 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:20:24.294 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:20:24.294 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:20:24.294 17:36:05 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:20:24.294 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=00000000000000000000000000000000 00:20:24.294 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 00000000000000000000000000000000 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:20:24.294 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@655 -- # es=1 00:20:24.294 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:20:24.294 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:20:24.294 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:20:24.294 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@85 -- # ns_is_visible 0x2 00:20:24.294 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:20:24.294 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:20:24.294 [ 0]:0x2 00:20:24.294 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:20:24.294 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:20:24.294 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=b881f9b1cabf486c8f254f23f54c5f63 00:20:24.294 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ b881f9b1cabf486c8f254f23f54c5f63 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:20:24.294 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_add_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:20:24.860 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@89 -- # ns_is_visible 0x1 00:20:24.860 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:20:24.860 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:20:24.860 [ 0]:0x1 00:20:24.860 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:20:24.860 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:20:24.860 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=459b07a705eb4b35a27dabff97462bcc 00:20:24.860 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 459b07a705eb4b35a27dabff97462bcc != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:20:24.860 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@90 -- # ns_is_visible 0x2 00:20:24.860 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:20:24.860 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:20:24.860 [ 1]:0x2 00:20:24.860 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:20:24.860 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:20:24.860 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=b881f9b1cabf486c8f254f23f54c5f63 00:20:24.860 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ b881f9b1cabf486c8f254f23f54c5f63 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:20:24.860 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@93 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:20:25.119 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@94 -- # NOT ns_is_visible 0x1 00:20:25.119 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@652 -- # local es=0 00:20:25.119 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@654 -- # valid_exec_arg ns_is_visible 0x1 00:20:25.119 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@640 -- # local arg=ns_is_visible 00:20:25.119 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:20:25.119 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # type -t ns_is_visible 00:20:25.119 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:20:25.119 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@655 -- # ns_is_visible 0x1 00:20:25.119 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:20:25.119 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:20:25.119 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:20:25.119 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:20:25.119 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=00000000000000000000000000000000 00:20:25.119 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 00000000000000000000000000000000 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:20:25.119 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@655 -- # es=1 00:20:25.119 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:20:25.119 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:20:25.119 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:20:25.119 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@95 -- # ns_is_visible 0x2 00:20:25.119 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:20:25.119 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:20:25.119 [ 0]:0x2 00:20:25.119 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:20:25.119 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:20:25.119 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=b881f9b1cabf486c8f254f23f54c5f63 00:20:25.119 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ b881f9b1cabf486c8f254f23f54c5f63 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:20:25.119 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@97 -- # disconnect 00:20:25.119 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@38 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:20:25.119 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:20:25.119 17:36:06 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_add_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:20:25.685 17:36:07 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@101 -- # connect 2 00:20:25.685 17:36:07 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@22 -- # nvme connect -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 -I 70b65a8f-dc2d-4e3c-a332-3d0ff2d08a3a -a 10.0.0.2 -s 4420 -i 4 00:20:25.685 17:36:07 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@24 -- # waitforserial SPDKISFASTANDAWESOME 2 00:20:25.685 17:36:07 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1202 -- # local i=0 00:20:25.685 17:36:07 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:20:25.685 17:36:07 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1204 -- # [[ -n 2 ]] 00:20:25.685 17:36:07 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1205 -- # nvme_device_counter=2 00:20:25.685 17:36:07 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1209 -- # sleep 2 00:20:27.585 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:20:27.585 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:20:27.585 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:20:27.585 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1211 -- # nvme_devices=2 00:20:27.585 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:20:27.585 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1212 -- # return 0 00:20:27.585 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # nvme list-subsys -o json 00:20:27.585 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # jq -r '.[].Subsystems[] | select(.NQN=="nqn.2016-06.io.spdk:cnode1") | .Paths[0].Name' 00:20:27.585 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # ctrl_id=nvme0 00:20:27.585 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@27 -- # [[ -z nvme0 ]] 00:20:27.585 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@102 -- # ns_is_visible 0x1 00:20:27.585 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:20:27.585 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:20:27.843 [ 0]:0x1 00:20:27.843 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:20:27.843 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:20:27.843 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=459b07a705eb4b35a27dabff97462bcc 00:20:27.843 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 459b07a705eb4b35a27dabff97462bcc != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:20:27.843 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@103 -- # ns_is_visible 0x2 00:20:27.843 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:20:27.843 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:20:27.843 [ 1]:0x2 00:20:27.843 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:20:27.843 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:20:27.843 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=b881f9b1cabf486c8f254f23f54c5f63 00:20:27.843 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ b881f9b1cabf486c8f254f23f54c5f63 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:20:27.843 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@106 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@107 -- # NOT ns_is_visible 0x1 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@652 -- # local es=0 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@654 -- # valid_exec_arg ns_is_visible 0x1 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@640 -- # local arg=ns_is_visible 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # type -t ns_is_visible 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@655 -- # ns_is_visible 0x1 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=00000000000000000000000000000000 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 00000000000000000000000000000000 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@655 -- # es=1 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@108 -- # ns_is_visible 0x2 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:20:28.101 [ 0]:0x2 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=b881f9b1cabf486c8f254f23f54c5f63 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ b881f9b1cabf486c8f254f23f54c5f63 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@111 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 2 nqn.2016-06.io.spdk:host1 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@652 -- # local es=0 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@654 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 2 nqn.2016-06.io.spdk:host1 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@640 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@646 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@646 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@646 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:20:28.101 17:36:09 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@655 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 2 nqn.2016-06.io.spdk:host1 00:20:28.360 [2024-12-06 17:36:10.170920] nvmf_rpc.c:1873:nvmf_rpc_ns_visible_paused: *ERROR*: Unable to add/remove nqn.2016-06.io.spdk:host1 to namespace ID 2 00:20:28.360 request: 00:20:28.360 { 00:20:28.360 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:20:28.360 "nsid": 2, 00:20:28.360 "host": "nqn.2016-06.io.spdk:host1", 00:20:28.360 "method": "nvmf_ns_remove_host", 00:20:28.360 "req_id": 1 00:20:28.360 } 00:20:28.360 Got JSON-RPC error response 00:20:28.360 response: 00:20:28.360 { 00:20:28.360 "code": -32602, 00:20:28.360 "message": "Invalid parameters" 00:20:28.360 } 00:20:28.360 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@655 -- # es=1 00:20:28.360 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:20:28.360 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:20:28.360 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:20:28.360 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@112 -- # NOT ns_is_visible 0x1 00:20:28.360 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@652 -- # local es=0 00:20:28.360 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@654 -- # valid_exec_arg ns_is_visible 0x1 00:20:28.360 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@640 -- # local arg=ns_is_visible 00:20:28.360 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:20:28.360 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # type -t ns_is_visible 00:20:28.360 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:20:28.360 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@655 -- # ns_is_visible 0x1 00:20:28.360 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:20:28.360 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:20:28.619 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:20:28.619 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:20:28.619 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=00000000000000000000000000000000 00:20:28.619 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 00000000000000000000000000000000 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:20:28.619 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@655 -- # es=1 00:20:28.619 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:20:28.619 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:20:28.619 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:20:28.619 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@113 -- # ns_is_visible 0x2 00:20:28.619 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:20:28.619 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:20:28.619 [ 0]:0x2 00:20:28.619 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:20:28.619 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:20:28.619 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=b881f9b1cabf486c8f254f23f54c5f63 00:20:28.619 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ b881f9b1cabf486c8f254f23f54c5f63 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:20:28.619 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@114 -- # disconnect 00:20:28.619 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@38 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:20:28.619 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:20:28.619 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@118 -- # hostpid=235258 00:20:28.619 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@117 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -r /var/tmp/host.sock -m 2 00:20:28.619 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@119 -- # trap 'killprocess $hostpid; nvmftestfini' SIGINT SIGTERM EXIT 00:20:28.619 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@121 -- # waitforlisten 235258 /var/tmp/host.sock 00:20:28.619 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@835 -- # '[' -z 235258 ']' 00:20:28.619 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/host.sock 00:20:28.619 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@840 -- # local max_retries=100 00:20:28.619 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/host.sock...' 00:20:28.619 Waiting for process to start up and listen on UNIX domain socket /var/tmp/host.sock... 00:20:28.619 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@844 -- # xtrace_disable 00:20:28.619 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:20:28.619 [2024-12-06 17:36:10.373579] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:20:28.619 [2024-12-06 17:36:10.373684] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid235258 ] 00:20:28.619 [2024-12-06 17:36:10.443695] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:28.877 [2024-12-06 17:36:10.490017] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:20:29.134 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:20:29.134 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@868 -- # return 0 00:20:29.134 17:36:10 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@122 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:20:29.392 17:36:11 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:20:29.649 17:36:11 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@124 -- # uuid2nguid 08a4d33a-f0ed-4528-baec-ffb64069a77c 00:20:29.649 17:36:11 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@787 -- # tr -d - 00:20:29.649 17:36:11 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@124 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 1 -g 08A4D33AF0ED4528BAECFFB64069A77C -i 00:20:29.907 17:36:11 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@125 -- # uuid2nguid 957cadae-2228-4120-bdf5-2f5e983b7b3c 00:20:29.907 17:36:11 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@787 -- # tr -d - 00:20:29.907 17:36:11 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc2 -n 2 -g 957CADAE22284120BDF52F5E983B7B3C -i 00:20:30.164 17:36:11 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@126 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_add_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:20:30.422 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@127 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_add_host nqn.2016-06.io.spdk:cnode1 2 nqn.2016-06.io.spdk:host2 00:20:30.680 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@129 -- # hostrpc bdev_nvme_attach_controller -t tcp -a 10.0.0.2 -f ipv4 -s 4420 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 -b nvme0 00:20:30.680 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -a 10.0.0.2 -f ipv4 -s 4420 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 -b nvme0 00:20:31.246 nvme0n1 00:20:31.246 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@131 -- # hostrpc bdev_nvme_attach_controller -t tcp -a 10.0.0.2 -f ipv4 -s 4420 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host2 -b nvme1 00:20:31.246 17:36:12 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -a 10.0.0.2 -f ipv4 -s 4420 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host2 -b nvme1 00:20:31.505 nvme1n2 00:20:31.505 17:36:13 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@134 -- # hostrpc bdev_get_bdevs 00:20:31.505 17:36:13 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@134 -- # jq -r '.[].name' 00:20:31.505 17:36:13 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_get_bdevs 00:20:31.505 17:36:13 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@134 -- # sort 00:20:31.505 17:36:13 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@134 -- # xargs 00:20:31.764 17:36:13 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@134 -- # [[ nvme0n1 nvme1n2 == \n\v\m\e\0\n\1\ \n\v\m\e\1\n\2 ]] 00:20:31.764 17:36:13 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@135 -- # hostrpc bdev_get_bdevs -b nvme0n1 00:20:31.764 17:36:13 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@135 -- # jq -r '.[].uuid' 00:20:31.764 17:36:13 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_get_bdevs -b nvme0n1 00:20:32.330 17:36:13 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@135 -- # [[ 08a4d33a-f0ed-4528-baec-ffb64069a77c == \0\8\a\4\d\3\3\a\-\f\0\e\d\-\4\5\2\8\-\b\a\e\c\-\f\f\b\6\4\0\6\9\a\7\7\c ]] 00:20:32.330 17:36:13 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@136 -- # hostrpc bdev_get_bdevs -b nvme1n2 00:20:32.330 17:36:13 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@136 -- # jq -r '.[].uuid' 00:20:32.330 17:36:13 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_get_bdevs -b nvme1n2 00:20:32.330 17:36:14 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@136 -- # [[ 957cadae-2228-4120-bdf5-2f5e983b7b3c == \9\5\7\c\a\d\a\e\-\2\2\2\8\-\4\1\2\0\-\b\d\f\5\-\2\f\5\e\9\8\3\b\7\b\3\c ]] 00:20:32.330 17:36:14 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@137 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:20:32.588 17:36:14 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@138 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:20:32.862 17:36:14 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@141 -- # uuid2nguid 08a4d33a-f0ed-4528-baec-ffb64069a77c 00:20:32.862 17:36:14 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@787 -- # tr -d - 00:20:32.862 17:36:14 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@141 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 invalid -n 1 -g 08A4D33AF0ED4528BAECFFB64069A77C 00:20:32.862 17:36:14 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@652 -- # local es=0 00:20:32.862 17:36:14 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@654 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 invalid -n 1 -g 08A4D33AF0ED4528BAECFFB64069A77C 00:20:32.862 17:36:14 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@640 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:20:32.862 17:36:14 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:20:32.862 17:36:14 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:20:32.862 17:36:14 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:20:32.862 17:36:14 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@646 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:20:32.862 17:36:14 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:20:32.862 17:36:14 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@646 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:20:32.862 17:36:14 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@646 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:20:32.862 17:36:14 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@655 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 invalid -n 1 -g 08A4D33AF0ED4528BAECFFB64069A77C 00:20:33.119 [2024-12-06 17:36:14.948803] bdev.c:8674:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: invalid 00:20:33.119 [2024-12-06 17:36:14.948844] subsystem.c:2160:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode1: bdev invalid cannot be opened, error=-19 00:20:33.119 [2024-12-06 17:36:14.948871] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:20:33.119 request: 00:20:33.119 { 00:20:33.119 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:20:33.119 "namespace": { 00:20:33.119 "bdev_name": "invalid", 00:20:33.119 "nsid": 1, 00:20:33.119 "nguid": "08A4D33AF0ED4528BAECFFB64069A77C", 00:20:33.119 "no_auto_visible": false, 00:20:33.119 "hide_metadata": false 00:20:33.119 }, 00:20:33.119 "method": "nvmf_subsystem_add_ns", 00:20:33.119 "req_id": 1 00:20:33.119 } 00:20:33.119 Got JSON-RPC error response 00:20:33.119 response: 00:20:33.119 { 00:20:33.119 "code": -32602, 00:20:33.119 "message": "Invalid parameters" 00:20:33.119 } 00:20:33.377 17:36:14 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@655 -- # es=1 00:20:33.377 17:36:14 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:20:33.377 17:36:14 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:20:33.377 17:36:14 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:20:33.377 17:36:14 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@142 -- # uuid2nguid 08a4d33a-f0ed-4528-baec-ffb64069a77c 00:20:33.377 17:36:14 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@787 -- # tr -d - 00:20:33.377 17:36:14 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@142 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 1 -g 08A4D33AF0ED4528BAECFFB64069A77C -i 00:20:33.635 17:36:15 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@143 -- # sleep 2s 00:20:35.536 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@144 -- # hostrpc bdev_get_bdevs 00:20:35.536 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@144 -- # jq length 00:20:35.536 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_get_bdevs 00:20:35.795 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@144 -- # (( 0 == 0 )) 00:20:35.795 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@146 -- # killprocess 235258 00:20:35.795 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@954 -- # '[' -z 235258 ']' 00:20:35.795 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@958 -- # kill -0 235258 00:20:35.795 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@959 -- # uname 00:20:35.795 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:20:35.795 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 235258 00:20:35.795 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:20:35.795 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:20:35.795 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@972 -- # echo 'killing process with pid 235258' 00:20:35.795 killing process with pid 235258 00:20:35.795 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@973 -- # kill 235258 00:20:35.795 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@978 -- # wait 235258 00:20:36.361 17:36:17 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@147 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:20:36.361 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@149 -- # trap - SIGINT SIGTERM EXIT 00:20:36.361 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@150 -- # nvmftestfini 00:20:36.361 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@516 -- # nvmfcleanup 00:20:36.361 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@121 -- # sync 00:20:36.361 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:20:36.361 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@124 -- # set +e 00:20:36.361 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@125 -- # for i in {1..20} 00:20:36.361 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:20:36.361 rmmod nvme_tcp 00:20:36.670 rmmod nvme_fabrics 00:20:36.670 rmmod nvme_keyring 00:20:36.670 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:20:36.670 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@128 -- # set -e 00:20:36.670 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@129 -- # return 0 00:20:36.670 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@517 -- # '[' -n 233138 ']' 00:20:36.670 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@518 -- # killprocess 233138 00:20:36.670 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@954 -- # '[' -z 233138 ']' 00:20:36.670 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@958 -- # kill -0 233138 00:20:36.670 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@959 -- # uname 00:20:36.670 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:20:36.670 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 233138 00:20:36.670 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:20:36.670 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:20:36.670 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@972 -- # echo 'killing process with pid 233138' 00:20:36.670 killing process with pid 233138 00:20:36.670 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@973 -- # kill 233138 00:20:36.670 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@978 -- # wait 233138 00:20:36.947 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:20:36.947 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:20:36.947 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:20:36.947 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@297 -- # iptr 00:20:36.947 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@791 -- # iptables-save 00:20:36.947 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:20:36.947 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@791 -- # iptables-restore 00:20:36.947 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:20:36.947 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@302 -- # remove_spdk_ns 00:20:36.947 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:20:36.947 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:20:36.947 17:36:18 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:20:38.878 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:20:38.878 00:20:38.878 real 0m24.975s 00:20:38.878 user 0m36.355s 00:20:38.878 sys 0m4.671s 00:20:38.878 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1130 -- # xtrace_disable 00:20:38.878 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:20:38.878 ************************************ 00:20:38.878 END TEST nvmf_ns_masking 00:20:38.878 ************************************ 00:20:38.878 17:36:20 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@27 -- # [[ 1 -eq 1 ]] 00:20:38.878 17:36:20 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@28 -- # run_test nvmf_nvme_cli /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvme_cli.sh --transport=tcp 00:20:38.878 17:36:20 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:20:38.878 17:36:20 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:20:38.878 17:36:20 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:20:38.878 ************************************ 00:20:38.878 START TEST nvmf_nvme_cli 00:20:38.878 ************************************ 00:20:38.878 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvme_cli.sh --transport=tcp 00:20:38.878 * Looking for test storage... 00:20:38.878 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:20:38.878 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:20:38.878 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1711 -- # lcov --version 00:20:38.878 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@333 -- # local ver1 ver1_l 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@334 -- # local ver2 ver2_l 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@336 -- # IFS=.-: 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@336 -- # read -ra ver1 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@337 -- # IFS=.-: 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@337 -- # read -ra ver2 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@338 -- # local 'op=<' 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@340 -- # ver1_l=2 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@341 -- # ver2_l=1 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@344 -- # case "$op" in 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@345 -- # : 1 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@364 -- # (( v = 0 )) 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@365 -- # decimal 1 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@353 -- # local d=1 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@355 -- # echo 1 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@365 -- # ver1[v]=1 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@366 -- # decimal 2 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@353 -- # local d=2 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@355 -- # echo 2 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@366 -- # ver2[v]=2 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@368 -- # return 0 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:20:39.139 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:39.139 --rc genhtml_branch_coverage=1 00:20:39.139 --rc genhtml_function_coverage=1 00:20:39.139 --rc genhtml_legend=1 00:20:39.139 --rc geninfo_all_blocks=1 00:20:39.139 --rc geninfo_unexecuted_blocks=1 00:20:39.139 00:20:39.139 ' 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:20:39.139 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:39.139 --rc genhtml_branch_coverage=1 00:20:39.139 --rc genhtml_function_coverage=1 00:20:39.139 --rc genhtml_legend=1 00:20:39.139 --rc geninfo_all_blocks=1 00:20:39.139 --rc geninfo_unexecuted_blocks=1 00:20:39.139 00:20:39.139 ' 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:20:39.139 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:39.139 --rc genhtml_branch_coverage=1 00:20:39.139 --rc genhtml_function_coverage=1 00:20:39.139 --rc genhtml_legend=1 00:20:39.139 --rc geninfo_all_blocks=1 00:20:39.139 --rc geninfo_unexecuted_blocks=1 00:20:39.139 00:20:39.139 ' 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:20:39.139 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:39.139 --rc genhtml_branch_coverage=1 00:20:39.139 --rc genhtml_function_coverage=1 00:20:39.139 --rc genhtml_legend=1 00:20:39.139 --rc geninfo_all_blocks=1 00:20:39.139 --rc geninfo_unexecuted_blocks=1 00:20:39.139 00:20:39.139 ' 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@7 -- # uname -s 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@15 -- # shopt -s extglob 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- paths/export.sh@5 -- # export PATH 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@51 -- # : 0 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:20:39.139 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@55 -- # have_pci_nics=0 00:20:39.139 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@11 -- # MALLOC_BDEV_SIZE=64 00:20:39.140 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:20:39.140 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@14 -- # devs=() 00:20:39.140 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@16 -- # nvmftestinit 00:20:39.140 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:20:39.140 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:20:39.140 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@476 -- # prepare_net_devs 00:20:39.140 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@438 -- # local -g is_hw=no 00:20:39.140 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@440 -- # remove_spdk_ns 00:20:39.140 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:20:39.140 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:20:39.140 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:20:39.140 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:20:39.140 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:20:39.140 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@309 -- # xtrace_disable 00:20:39.140 17:36:20 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@315 -- # pci_devs=() 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@315 -- # local -a pci_devs 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@316 -- # pci_net_devs=() 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@317 -- # pci_drivers=() 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@317 -- # local -A pci_drivers 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@319 -- # net_devs=() 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@319 -- # local -ga net_devs 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@320 -- # e810=() 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@320 -- # local -ga e810 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@321 -- # x722=() 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@321 -- # local -ga x722 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@322 -- # mlx=() 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@322 -- # local -ga mlx 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:20:41.676 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:20:41.676 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@418 -- # [[ up == up ]] 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:20:41.676 Found net devices under 0000:0a:00.0: cvl_0_0 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@418 -- # [[ up == up ]] 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:20:41.676 Found net devices under 0000:0a:00.1: cvl_0_1 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@442 -- # is_hw=yes 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:20:41.676 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:20:41.676 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.331 ms 00:20:41.676 00:20:41.676 --- 10.0.0.2 ping statistics --- 00:20:41.676 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:20:41.676 rtt min/avg/max/mdev = 0.331/0.331/0.331/0.000 ms 00:20:41.676 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:20:41.676 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:20:41.676 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.123 ms 00:20:41.676 00:20:41.676 --- 10.0.0.1 ping statistics --- 00:20:41.676 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:20:41.676 rtt min/avg/max/mdev = 0.123/0.123/0.123/0.000 ms 00:20:41.677 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:20:41.677 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@450 -- # return 0 00:20:41.677 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:20:41.677 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:20:41.677 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:20:41.677 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:20:41.677 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:20:41.677 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:20:41.677 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:20:41.677 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@17 -- # nvmfappstart -m 0xF 00:20:41.677 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:20:41.677 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@726 -- # xtrace_disable 00:20:41.677 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:20:41.677 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@509 -- # nvmfpid=238287 00:20:41.677 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:20:41.677 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@510 -- # waitforlisten 238287 00:20:41.677 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@835 -- # '[' -z 238287 ']' 00:20:41.677 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:41.677 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@840 -- # local max_retries=100 00:20:41.677 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:41.677 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:41.677 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@844 -- # xtrace_disable 00:20:41.677 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:20:41.677 [2024-12-06 17:36:23.288143] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:20:41.677 [2024-12-06 17:36:23.288228] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:20:41.677 [2024-12-06 17:36:23.362036] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:20:41.677 [2024-12-06 17:36:23.410984] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:20:41.677 [2024-12-06 17:36:23.411050] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:20:41.677 [2024-12-06 17:36:23.411074] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:20:41.677 [2024-12-06 17:36:23.411085] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:20:41.677 [2024-12-06 17:36:23.411094] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:20:41.677 [2024-12-06 17:36:23.412534] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:20:41.677 [2024-12-06 17:36:23.412598] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:20:41.677 [2024-12-06 17:36:23.412681] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:20:41.677 [2024-12-06 17:36:23.412686] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:20:41.934 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:20:41.934 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@868 -- # return 0 00:20:41.934 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:20:41.934 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@732 -- # xtrace_disable 00:20:41.934 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:20:41.934 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@19 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@563 -- # xtrace_disable 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:20:41.935 [2024-12-06 17:36:23.547161] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@21 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@563 -- # xtrace_disable 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:20:41.935 Malloc0 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@22 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@563 -- # xtrace_disable 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:20:41.935 Malloc1 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME -d SPDK_Controller1 -i 291 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@563 -- # xtrace_disable 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@25 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@563 -- # xtrace_disable 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@26 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@563 -- # xtrace_disable 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@563 -- # xtrace_disable 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:20:41.935 [2024-12-06 17:36:23.637069] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@28 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@563 -- # xtrace_disable 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:20:41.935 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@30 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -a 10.0.0.2 -s 4420 00:20:42.193 00:20:42.193 Discovery Log Number of Records 2, Generation counter 2 00:20:42.193 =====Discovery Log Entry 0====== 00:20:42.193 trtype: tcp 00:20:42.193 adrfam: ipv4 00:20:42.193 subtype: current discovery subsystem 00:20:42.193 treq: not required 00:20:42.193 portid: 0 00:20:42.193 trsvcid: 4420 00:20:42.193 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:20:42.193 traddr: 10.0.0.2 00:20:42.193 eflags: explicit discovery connections, duplicate discovery information 00:20:42.193 sectype: none 00:20:42.193 =====Discovery Log Entry 1====== 00:20:42.193 trtype: tcp 00:20:42.193 adrfam: ipv4 00:20:42.193 subtype: nvme subsystem 00:20:42.193 treq: not required 00:20:42.193 portid: 0 00:20:42.193 trsvcid: 4420 00:20:42.193 subnqn: nqn.2016-06.io.spdk:cnode1 00:20:42.193 traddr: 10.0.0.2 00:20:42.193 eflags: none 00:20:42.193 sectype: none 00:20:42.193 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@31 -- # devs=($(get_nvme_devs)) 00:20:42.193 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@31 -- # get_nvme_devs 00:20:42.193 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # local dev _ 00:20:42.193 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # read -r dev _ 00:20:42.193 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@549 -- # nvme list 00:20:42.193 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@553 -- # [[ Node == /dev/nvme* ]] 00:20:42.193 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # read -r dev _ 00:20:42.193 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@553 -- # [[ --------------------- == /dev/nvme* ]] 00:20:42.193 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # read -r dev _ 00:20:42.193 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@31 -- # nvme_num_before_connection=0 00:20:42.193 17:36:23 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@32 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:20:42.759 17:36:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@34 -- # waitforserial SPDKISFASTANDAWESOME 2 00:20:42.759 17:36:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1202 -- # local i=0 00:20:42.759 17:36:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:20:42.759 17:36:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1204 -- # [[ -n 2 ]] 00:20:42.759 17:36:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1205 -- # nvme_device_counter=2 00:20:42.759 17:36:24 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1209 -- # sleep 2 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1211 -- # nvme_devices=2 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1212 -- # return 0 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@35 -- # get_nvme_devs 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # local dev _ 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # read -r dev _ 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@549 -- # nvme list 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@553 -- # [[ Node == /dev/nvme* ]] 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # read -r dev _ 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@553 -- # [[ --------------------- == /dev/nvme* ]] 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # read -r dev _ 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@553 -- # [[ /dev/nvme0n1 == /dev/nvme* ]] 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@554 -- # echo /dev/nvme0n1 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # read -r dev _ 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@553 -- # [[ /dev/nvme0n2 == /dev/nvme* ]] 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@554 -- # echo /dev/nvme0n2 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # read -r dev _ 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@35 -- # [[ -z /dev/nvme0n1 00:20:44.662 /dev/nvme0n2 ]] 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@59 -- # devs=($(get_nvme_devs)) 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@59 -- # get_nvme_devs 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # local dev _ 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # read -r dev _ 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@549 -- # nvme list 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@553 -- # [[ Node == /dev/nvme* ]] 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # read -r dev _ 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@553 -- # [[ --------------------- == /dev/nvme* ]] 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # read -r dev _ 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@553 -- # [[ /dev/nvme0n1 == /dev/nvme* ]] 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@554 -- # echo /dev/nvme0n1 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # read -r dev _ 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@553 -- # [[ /dev/nvme0n2 == /dev/nvme* ]] 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@554 -- # echo /dev/nvme0n2 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # read -r dev _ 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@59 -- # nvme_num=2 00:20:44.662 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@60 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:20:44.919 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@61 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1223 -- # local i=0 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1235 -- # return 0 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@62 -- # (( nvme_num <= nvme_num_before_connection )) 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@67 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@563 -- # xtrace_disable 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@68 -- # trap - SIGINT SIGTERM EXIT 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@70 -- # nvmftestfini 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@516 -- # nvmfcleanup 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@121 -- # sync 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@124 -- # set +e 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@125 -- # for i in {1..20} 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:20:44.919 rmmod nvme_tcp 00:20:44.919 rmmod nvme_fabrics 00:20:44.919 rmmod nvme_keyring 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@128 -- # set -e 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@129 -- # return 0 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@517 -- # '[' -n 238287 ']' 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@518 -- # killprocess 238287 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@954 -- # '[' -z 238287 ']' 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@958 -- # kill -0 238287 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@959 -- # uname 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 238287 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@972 -- # echo 'killing process with pid 238287' 00:20:44.919 killing process with pid 238287 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@973 -- # kill 238287 00:20:44.919 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@978 -- # wait 238287 00:20:45.177 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:20:45.177 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:20:45.177 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:20:45.177 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@297 -- # iptr 00:20:45.177 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@791 -- # iptables-save 00:20:45.177 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:20:45.177 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@791 -- # iptables-restore 00:20:45.177 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:20:45.177 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@302 -- # remove_spdk_ns 00:20:45.177 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:20:45.177 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:20:45.177 17:36:26 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:20:47.714 17:36:28 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:20:47.714 00:20:47.714 real 0m8.271s 00:20:47.714 user 0m14.585s 00:20:47.714 sys 0m2.321s 00:20:47.714 17:36:28 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1130 -- # xtrace_disable 00:20:47.714 17:36:28 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:20:47.714 ************************************ 00:20:47.714 END TEST nvmf_nvme_cli 00:20:47.714 ************************************ 00:20:47.714 17:36:28 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@30 -- # [[ 1 -eq 1 ]] 00:20:47.714 17:36:28 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@31 -- # run_test nvmf_vfio_user /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_vfio_user.sh --transport=tcp 00:20:47.714 17:36:28 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:20:47.714 17:36:28 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:20:47.714 17:36:28 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:20:47.714 ************************************ 00:20:47.714 START TEST nvmf_vfio_user 00:20:47.714 ************************************ 00:20:47.714 17:36:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_vfio_user.sh --transport=tcp 00:20:47.714 * Looking for test storage... 00:20:47.714 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1711 -- # lcov --version 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@333 -- # local ver1 ver1_l 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@334 -- # local ver2 ver2_l 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@336 -- # IFS=.-: 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@336 -- # read -ra ver1 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@337 -- # IFS=.-: 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@337 -- # read -ra ver2 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@338 -- # local 'op=<' 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@340 -- # ver1_l=2 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@341 -- # ver2_l=1 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@344 -- # case "$op" in 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@345 -- # : 1 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@364 -- # (( v = 0 )) 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@365 -- # decimal 1 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@353 -- # local d=1 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@355 -- # echo 1 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@365 -- # ver1[v]=1 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@366 -- # decimal 2 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@353 -- # local d=2 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@355 -- # echo 2 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@366 -- # ver2[v]=2 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@368 -- # return 0 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:20:47.714 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:47.714 --rc genhtml_branch_coverage=1 00:20:47.714 --rc genhtml_function_coverage=1 00:20:47.714 --rc genhtml_legend=1 00:20:47.714 --rc geninfo_all_blocks=1 00:20:47.714 --rc geninfo_unexecuted_blocks=1 00:20:47.714 00:20:47.714 ' 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:20:47.714 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:47.714 --rc genhtml_branch_coverage=1 00:20:47.714 --rc genhtml_function_coverage=1 00:20:47.714 --rc genhtml_legend=1 00:20:47.714 --rc geninfo_all_blocks=1 00:20:47.714 --rc geninfo_unexecuted_blocks=1 00:20:47.714 00:20:47.714 ' 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:20:47.714 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:47.714 --rc genhtml_branch_coverage=1 00:20:47.714 --rc genhtml_function_coverage=1 00:20:47.714 --rc genhtml_legend=1 00:20:47.714 --rc geninfo_all_blocks=1 00:20:47.714 --rc geninfo_unexecuted_blocks=1 00:20:47.714 00:20:47.714 ' 00:20:47.714 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:20:47.714 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:47.715 --rc genhtml_branch_coverage=1 00:20:47.715 --rc genhtml_function_coverage=1 00:20:47.715 --rc genhtml_legend=1 00:20:47.715 --rc geninfo_all_blocks=1 00:20:47.715 --rc geninfo_unexecuted_blocks=1 00:20:47.715 00:20:47.715 ' 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@7 -- # uname -s 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@15 -- # shopt -s extglob 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- paths/export.sh@5 -- # export PATH 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@51 -- # : 0 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:20:47.715 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@55 -- # have_pci_nics=0 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@12 -- # MALLOC_BDEV_SIZE=64 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@14 -- # NUM_DEVICES=2 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@16 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@18 -- # export TEST_TRANSPORT=VFIOUSER 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@18 -- # TEST_TRANSPORT=VFIOUSER 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@47 -- # rm -rf /var/run/vfio-user 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@103 -- # setup_nvmf_vfio_user '' '' 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@51 -- # local nvmf_app_args= 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@52 -- # local transport_args= 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@55 -- # nvmfpid=239114 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m '[0,1,2,3]' 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@57 -- # echo 'Process pid: 239114' 00:20:47.715 Process pid: 239114 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@59 -- # trap 'killprocess $nvmfpid; exit 1' SIGINT SIGTERM EXIT 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@60 -- # waitforlisten 239114 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@835 -- # '[' -z 239114 ']' 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@840 -- # local max_retries=100 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:47.715 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@844 -- # xtrace_disable 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@10 -- # set +x 00:20:47.715 [2024-12-06 17:36:29.211320] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:20:47.715 [2024-12-06 17:36:29.211399] [ DPDK EAL parameters: nvmf -l 0,1,2,3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:20:47.715 [2024-12-06 17:36:29.278046] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:20:47.715 [2024-12-06 17:36:29.326282] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:20:47.715 [2024-12-06 17:36:29.326344] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:20:47.715 [2024-12-06 17:36:29.326357] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:20:47.715 [2024-12-06 17:36:29.326368] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:20:47.715 [2024-12-06 17:36:29.326378] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:20:47.715 [2024-12-06 17:36:29.327947] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:20:47.715 [2024-12-06 17:36:29.328007] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:20:47.715 [2024-12-06 17:36:29.328071] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:20:47.715 [2024-12-06 17:36:29.328074] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@868 -- # return 0 00:20:47.715 17:36:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@62 -- # sleep 1 00:20:48.648 17:36:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t VFIOUSER 00:20:49.213 17:36:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@66 -- # mkdir -p /var/run/vfio-user 00:20:49.213 17:36:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # seq 1 2 00:20:49.213 17:36:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # for i in $(seq 1 $NUM_DEVICES) 00:20:49.213 17:36:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@69 -- # mkdir -p /var/run/vfio-user/domain/vfio-user1/1 00:20:49.213 17:36:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc1 00:20:49.471 Malloc1 00:20:49.471 17:36:31 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2019-07.io.spdk:cnode1 -a -s SPDK1 00:20:49.729 17:36:31 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode1 Malloc1 00:20:49.986 17:36:31 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2019-07.io.spdk:cnode1 -t VFIOUSER -a /var/run/vfio-user/domain/vfio-user1/1 -s 0 00:20:50.245 17:36:31 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # for i in $(seq 1 $NUM_DEVICES) 00:20:50.245 17:36:31 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@69 -- # mkdir -p /var/run/vfio-user/domain/vfio-user2/2 00:20:50.245 17:36:31 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc2 00:20:50.503 Malloc2 00:20:50.503 17:36:32 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2019-07.io.spdk:cnode2 -a -s SPDK2 00:20:50.761 17:36:32 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode2 Malloc2 00:20:51.028 17:36:32 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2019-07.io.spdk:cnode2 -t VFIOUSER -a /var/run/vfio-user/domain/vfio-user2/2 -s 0 00:20:51.285 17:36:33 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@104 -- # run_nvmf_vfio_user 00:20:51.285 17:36:33 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@80 -- # seq 1 2 00:20:51.286 17:36:33 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@80 -- # for i in $(seq 1 $NUM_DEVICES) 00:20:51.286 17:36:33 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@81 -- # test_traddr=/var/run/vfio-user/domain/vfio-user1/1 00:20:51.286 17:36:33 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@82 -- # test_subnqn=nqn.2019-07.io.spdk:cnode1 00:20:51.286 17:36:33 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -g -L nvme -L nvme_vfio -L vfio_pci 00:20:51.286 [2024-12-06 17:36:33.034240] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:20:51.286 [2024-12-06 17:36:33.034281] [ DPDK EAL parameters: identify --no-shconf -c 0x1 -n 1 -m 0 --no-pci --single-file-segments --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid239533 ] 00:20:51.286 [2024-12-06 17:36:33.086958] nvme_vfio_user.c: 259:nvme_vfio_ctrlr_scan: *DEBUG*: Scan controller : /var/run/vfio-user/domain/vfio-user1/1 00:20:51.286 [2024-12-06 17:36:33.092203] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 0, Size 0x2000, Offset 0x0, Flags 0xf, Cap offset 32 00:20:51.286 [2024-12-06 17:36:33.092232] vfio_user_pci.c: 233:vfio_device_setup_sparse_mmaps: *DEBUG*: Sparse region 0, Size 0x1000, Offset 0x1000, Map addr 0x7f3670596000 00:20:51.286 [2024-12-06 17:36:33.093202] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 1, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:20:51.286 [2024-12-06 17:36:33.094192] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 2, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:20:51.286 [2024-12-06 17:36:33.095198] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 3, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:20:51.286 [2024-12-06 17:36:33.096203] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 4, Size 0x2000, Offset 0x0, Flags 0x3, Cap offset 0 00:20:51.286 [2024-12-06 17:36:33.097206] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 5, Size 0x1000, Offset 0x0, Flags 0x3, Cap offset 0 00:20:51.286 [2024-12-06 17:36:33.098214] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 6, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:20:51.286 [2024-12-06 17:36:33.099223] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 7, Size 0x1000, Offset 0x0, Flags 0x3, Cap offset 0 00:20:51.286 [2024-12-06 17:36:33.100224] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 8, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:20:51.286 [2024-12-06 17:36:33.101231] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 9, Size 0xc000, Offset 0x0, Flags 0xf, Cap offset 32 00:20:51.286 [2024-12-06 17:36:33.101252] vfio_user_pci.c: 233:vfio_device_setup_sparse_mmaps: *DEBUG*: Sparse region 0, Size 0xb000, Offset 0x1000, Map addr 0x7f366f28e000 00:20:51.286 [2024-12-06 17:36:33.102379] vfio_user_pci.c: 65:vfio_add_mr: *DEBUG*: Add memory region: FD 10, VADDR 0x200000200000, IOVA 0x200000200000, Size 0x200000 00:20:51.286 [2024-12-06 17:36:33.117513] vfio_user_pci.c: 386:spdk_vfio_user_setup: *DEBUG*: Device vfio-user0, Path /var/run/vfio-user/domain/vfio-user1/1/cntrl Setup Successfully 00:20:51.286 [2024-12-06 17:36:33.117563] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to connect adminq (no timeout) 00:20:51.545 [2024-12-06 17:36:33.126381] nvme_vfio_user.c: 103:nvme_vfio_ctrlr_get_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x0, value 0x201e0100ff 00:20:51.545 [2024-12-06 17:36:33.126435] nvme_pcie_common.c: 159:nvme_pcie_qpair_construct: *INFO*: max_completions_cap = 64 num_trackers = 192 00:20:51.545 [2024-12-06 17:36:33.126526] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to wait for connect adminq (no timeout) 00:20:51.545 [2024-12-06 17:36:33.126552] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to read vs (no timeout) 00:20:51.545 [2024-12-06 17:36:33.126564] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to read vs wait for vs (no timeout) 00:20:51.545 [2024-12-06 17:36:33.127370] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x8, value 0x10300 00:20:51.545 [2024-12-06 17:36:33.127392] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to read cap (no timeout) 00:20:51.545 [2024-12-06 17:36:33.127404] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to read cap wait for cap (no timeout) 00:20:51.545 [2024-12-06 17:36:33.128371] nvme_vfio_user.c: 103:nvme_vfio_ctrlr_get_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x0, value 0x201e0100ff 00:20:51.545 [2024-12-06 17:36:33.128390] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to check en (no timeout) 00:20:51.545 [2024-12-06 17:36:33.128403] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to check en wait for cc (timeout 15000 ms) 00:20:51.545 [2024-12-06 17:36:33.129376] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x14, value 0x0 00:20:51.545 [2024-12-06 17:36:33.129394] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to disable and wait for CSTS.RDY = 0 (timeout 15000 ms) 00:20:51.545 [2024-12-06 17:36:33.130382] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x1c, value 0x0 00:20:51.545 [2024-12-06 17:36:33.130402] nvme_ctrlr.c:3906:nvme_ctrlr_process_init_wait_for_ready_0: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] CC.EN = 0 && CSTS.RDY = 0 00:20:51.545 [2024-12-06 17:36:33.130411] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to controller is disabled (timeout 15000 ms) 00:20:51.545 [2024-12-06 17:36:33.130423] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to enable controller by writing CC.EN = 1 (timeout 15000 ms) 00:20:51.545 [2024-12-06 17:36:33.130537] nvme_ctrlr.c:4104:nvme_ctrlr_process_init: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] Setting CC.EN = 1 00:20:51.545 [2024-12-06 17:36:33.130546] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to enable controller by writing CC.EN = 1 reg (timeout 15000 ms) 00:20:51.545 [2024-12-06 17:36:33.130554] nvme_vfio_user.c: 61:nvme_vfio_ctrlr_set_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x28, value 0x2000003c0000 00:20:51.545 [2024-12-06 17:36:33.131398] nvme_vfio_user.c: 61:nvme_vfio_ctrlr_set_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x30, value 0x2000003be000 00:20:51.545 [2024-12-06 17:36:33.132391] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x24, value 0xff00ff 00:20:51.545 [2024-12-06 17:36:33.133398] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x14, value 0x460001 00:20:51.545 [2024-12-06 17:36:33.134392] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:20:51.545 [2024-12-06 17:36:33.134487] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to wait for CSTS.RDY = 1 (timeout 15000 ms) 00:20:51.545 [2024-12-06 17:36:33.135410] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x1c, value 0x1 00:20:51.545 [2024-12-06 17:36:33.135428] nvme_ctrlr.c:3941:nvme_ctrlr_process_init_enable_wait_for_ready_1: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] CC.EN = 1 && CSTS.RDY = 1 - controller is ready 00:20:51.545 [2024-12-06 17:36:33.135437] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to reset admin queue (timeout 30000 ms) 00:20:51.545 [2024-12-06 17:36:33.135461] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to identify controller (no timeout) 00:20:51.545 [2024-12-06 17:36:33.135479] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to wait for identify controller (timeout 30000 ms) 00:20:51.545 [2024-12-06 17:36:33.135508] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:20:51.545 [2024-12-06 17:36:33.135526] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:20:51.545 [2024-12-06 17:36:33.135532] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:20:51.545 [2024-12-06 17:36:33.135549] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:0 cdw10:00000001 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:20:51.545 [2024-12-06 17:36:33.135622] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0001 p:1 m:0 dnr:0 00:20:51.545 [2024-12-06 17:36:33.135638] nvme_ctrlr.c:2081:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] transport max_xfer_size 131072 00:20:51.545 [2024-12-06 17:36:33.135674] nvme_ctrlr.c:2085:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] MDTS max_xfer_size 131072 00:20:51.545 [2024-12-06 17:36:33.135684] nvme_ctrlr.c:2088:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] CNTLID 0x0001 00:20:51.545 [2024-12-06 17:36:33.135692] nvme_ctrlr.c:2099:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] Identify CNTLID 0x0001 != Connect CNTLID 0x0000 00:20:51.545 [2024-12-06 17:36:33.135700] nvme_ctrlr.c:2112:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] transport max_sges 1 00:20:51.545 [2024-12-06 17:36:33.135718] nvme_ctrlr.c:2127:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] fuses compare and write: 1 00:20:51.545 [2024-12-06 17:36:33.135726] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to configure AER (timeout 30000 ms) 00:20:51.545 [2024-12-06 17:36:33.135743] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to wait for configure aer (timeout 30000 ms) 00:20:51.545 [2024-12-06 17:36:33.135759] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES ASYNC EVENT CONFIGURATION cid:191 cdw10:0000000b PRP1 0x0 PRP2 0x0 00:20:51.545 [2024-12-06 17:36:33.135779] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0002 p:1 m:0 dnr:0 00:20:51.545 [2024-12-06 17:36:33.135795] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:190 nsid:0 cdw10:00000000 cdw11:00000000 00:20:51.545 [2024-12-06 17:36:33.135808] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:189 nsid:0 cdw10:00000000 cdw11:00000000 00:20:51.545 [2024-12-06 17:36:33.135820] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:188 nsid:0 cdw10:00000000 cdw11:00000000 00:20:51.545 [2024-12-06 17:36:33.135831] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:187 nsid:0 cdw10:00000000 cdw11:00000000 00:20:51.545 [2024-12-06 17:36:33.135840] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to set keep alive timeout (timeout 30000 ms) 00:20:51.545 [2024-12-06 17:36:33.135856] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to wait for set keep alive timeout (timeout 30000 ms) 00:20:51.545 [2024-12-06 17:36:33.135870] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES KEEP ALIVE TIMER cid:191 cdw10:0000000f PRP1 0x0 PRP2 0x0 00:20:51.545 [2024-12-06 17:36:33.135882] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0007 p:1 m:0 dnr:0 00:20:51.545 [2024-12-06 17:36:33.135892] nvme_ctrlr.c:3047:nvme_ctrlr_set_keep_alive_timeout_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] Controller adjusted keep alive timeout to 0 ms 00:20:51.545 [2024-12-06 17:36:33.135901] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to identify controller iocs specific (timeout 30000 ms) 00:20:51.545 [2024-12-06 17:36:33.135911] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to set number of queues (timeout 30000 ms) 00:20:51.545 [2024-12-06 17:36:33.135921] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to wait for set number of queues (timeout 30000 ms) 00:20:51.545 [2024-12-06 17:36:33.135933] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES NUMBER OF QUEUES cid:191 cdw10:00000007 PRP1 0x0 PRP2 0x0 00:20:51.545 [2024-12-06 17:36:33.135956] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:7e007e sqhd:0008 p:1 m:0 dnr:0 00:20:51.546 [2024-12-06 17:36:33.136036] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to identify active ns (timeout 30000 ms) 00:20:51.546 [2024-12-06 17:36:33.136052] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to wait for identify active ns (timeout 30000 ms) 00:20:51.546 [2024-12-06 17:36:33.136064] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f9000 len:4096 00:20:51.546 [2024-12-06 17:36:33.136072] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f9000 00:20:51.546 [2024-12-06 17:36:33.136078] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:20:51.546 [2024-12-06 17:36:33.136088] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:0 cdw10:00000002 cdw11:00000000 PRP1 0x2000002f9000 PRP2 0x0 00:20:51.546 [2024-12-06 17:36:33.136104] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0009 p:1 m:0 dnr:0 00:20:51.546 [2024-12-06 17:36:33.136119] nvme_ctrlr.c:4735:spdk_nvme_ctrlr_get_ns: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] Namespace 1 was added 00:20:51.546 [2024-12-06 17:36:33.136143] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to identify ns (timeout 30000 ms) 00:20:51.546 [2024-12-06 17:36:33.136157] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to wait for identify ns (timeout 30000 ms) 00:20:51.546 [2024-12-06 17:36:33.136169] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:20:51.546 [2024-12-06 17:36:33.136176] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:20:51.546 [2024-12-06 17:36:33.136182] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:20:51.546 [2024-12-06 17:36:33.136191] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:1 cdw10:00000000 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:20:51.546 [2024-12-06 17:36:33.136222] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000a p:1 m:0 dnr:0 00:20:51.546 [2024-12-06 17:36:33.136243] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to identify namespace id descriptors (timeout 30000 ms) 00:20:51.546 [2024-12-06 17:36:33.136257] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to wait for identify namespace id descriptors (timeout 30000 ms) 00:20:51.546 [2024-12-06 17:36:33.136268] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:20:51.546 [2024-12-06 17:36:33.136276] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:20:51.546 [2024-12-06 17:36:33.136282] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:20:51.546 [2024-12-06 17:36:33.136291] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:1 cdw10:00000003 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:20:51.546 [2024-12-06 17:36:33.136305] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000b p:1 m:0 dnr:0 00:20:51.546 [2024-12-06 17:36:33.136318] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to identify ns iocs specific (timeout 30000 ms) 00:20:51.546 [2024-12-06 17:36:33.136329] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to set supported log pages (timeout 30000 ms) 00:20:51.546 [2024-12-06 17:36:33.136342] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to set supported features (timeout 30000 ms) 00:20:51.546 [2024-12-06 17:36:33.136355] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to set host behavior support feature (timeout 30000 ms) 00:20:51.546 [2024-12-06 17:36:33.136363] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to set doorbell buffer config (timeout 30000 ms) 00:20:51.546 [2024-12-06 17:36:33.136371] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to set host ID (timeout 30000 ms) 00:20:51.546 [2024-12-06 17:36:33.136379] nvme_ctrlr.c:3147:nvme_ctrlr_set_host_id: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] NVMe-oF transport - not sending Set Features - Host ID 00:20:51.546 [2024-12-06 17:36:33.136386] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to transport ready (timeout 30000 ms) 00:20:51.546 [2024-12-06 17:36:33.136394] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to ready (no timeout) 00:20:51.546 [2024-12-06 17:36:33.136419] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ARBITRATION cid:191 cdw10:00000001 PRP1 0x0 PRP2 0x0 00:20:51.546 [2024-12-06 17:36:33.136437] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000c p:1 m:0 dnr:0 00:20:51.546 [2024-12-06 17:36:33.136459] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES POWER MANAGEMENT cid:191 cdw10:00000002 PRP1 0x0 PRP2 0x0 00:20:51.546 [2024-12-06 17:36:33.136472] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000d p:1 m:0 dnr:0 00:20:51.546 [2024-12-06 17:36:33.136488] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES TEMPERATURE THRESHOLD cid:191 cdw10:00000004 PRP1 0x0 PRP2 0x0 00:20:51.546 [2024-12-06 17:36:33.136502] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000e p:1 m:0 dnr:0 00:20:51.546 [2024-12-06 17:36:33.136519] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES NUMBER OF QUEUES cid:191 cdw10:00000007 PRP1 0x0 PRP2 0x0 00:20:51.546 [2024-12-06 17:36:33.136530] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:7e007e sqhd:000f p:1 m:0 dnr:0 00:20:51.546 [2024-12-06 17:36:33.136551] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f6000 len:8192 00:20:51.546 [2024-12-06 17:36:33.136562] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f6000 00:20:51.546 [2024-12-06 17:36:33.136568] nvme_pcie_common.c:1275:nvme_pcie_prp_list_append: *DEBUG*: prp[0] = 0x2000002f7000 00:20:51.546 [2024-12-06 17:36:33.136573] nvme_pcie_common.c:1291:nvme_pcie_prp_list_append: *DEBUG*: prp2 = 0x2000002f7000 00:20:51.546 [2024-12-06 17:36:33.136579] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 2 00:20:51.546 [2024-12-06 17:36:33.136588] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:191 nsid:ffffffff cdw10:07ff0001 cdw11:00000000 PRP1 0x2000002f6000 PRP2 0x2000002f7000 00:20:51.546 [2024-12-06 17:36:33.136599] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fc000 len:512 00:20:51.546 [2024-12-06 17:36:33.136607] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fc000 00:20:51.546 [2024-12-06 17:36:33.136613] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:20:51.546 [2024-12-06 17:36:33.136621] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:186 nsid:ffffffff cdw10:007f0002 cdw11:00000000 PRP1 0x2000002fc000 PRP2 0x0 00:20:51.546 [2024-12-06 17:36:33.136632] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:512 00:20:51.546 [2024-12-06 17:36:33.136639] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:20:51.546 [2024-12-06 17:36:33.136658] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:20:51.546 [2024-12-06 17:36:33.136679] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:185 nsid:ffffffff cdw10:007f0003 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:20:51.546 [2024-12-06 17:36:33.136693] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f4000 len:4096 00:20:51.546 [2024-12-06 17:36:33.136701] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f4000 00:20:51.546 [2024-12-06 17:36:33.136707] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:20:51.546 [2024-12-06 17:36:33.136716] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:184 nsid:ffffffff cdw10:03ff0005 cdw11:00000000 PRP1 0x2000002f4000 PRP2 0x0 00:20:51.546 [2024-12-06 17:36:33.136729] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0010 p:1 m:0 dnr:0 00:20:51.546 [2024-12-06 17:36:33.136752] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:186 cdw0:0 sqhd:0011 p:1 m:0 dnr:0 00:20:51.546 [2024-12-06 17:36:33.136771] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:185 cdw0:0 sqhd:0012 p:1 m:0 dnr:0 00:20:51.546 [2024-12-06 17:36:33.136783] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:184 cdw0:0 sqhd:0013 p:1 m:0 dnr:0 00:20:51.546 ===================================================== 00:20:51.546 NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:20:51.546 ===================================================== 00:20:51.546 Controller Capabilities/Features 00:20:51.546 ================================ 00:20:51.546 Vendor ID: 4e58 00:20:51.546 Subsystem Vendor ID: 4e58 00:20:51.546 Serial Number: SPDK1 00:20:51.546 Model Number: SPDK bdev Controller 00:20:51.546 Firmware Version: 25.01 00:20:51.546 Recommended Arb Burst: 6 00:20:51.546 IEEE OUI Identifier: 8d 6b 50 00:20:51.546 Multi-path I/O 00:20:51.546 May have multiple subsystem ports: Yes 00:20:51.546 May have multiple controllers: Yes 00:20:51.546 Associated with SR-IOV VF: No 00:20:51.546 Max Data Transfer Size: 131072 00:20:51.546 Max Number of Namespaces: 32 00:20:51.546 Max Number of I/O Queues: 127 00:20:51.546 NVMe Specification Version (VS): 1.3 00:20:51.546 NVMe Specification Version (Identify): 1.3 00:20:51.546 Maximum Queue Entries: 256 00:20:51.546 Contiguous Queues Required: Yes 00:20:51.546 Arbitration Mechanisms Supported 00:20:51.546 Weighted Round Robin: Not Supported 00:20:51.546 Vendor Specific: Not Supported 00:20:51.546 Reset Timeout: 15000 ms 00:20:51.546 Doorbell Stride: 4 bytes 00:20:51.546 NVM Subsystem Reset: Not Supported 00:20:51.546 Command Sets Supported 00:20:51.546 NVM Command Set: Supported 00:20:51.546 Boot Partition: Not Supported 00:20:51.546 Memory Page Size Minimum: 4096 bytes 00:20:51.546 Memory Page Size Maximum: 4096 bytes 00:20:51.547 Persistent Memory Region: Not Supported 00:20:51.547 Optional Asynchronous Events Supported 00:20:51.547 Namespace Attribute Notices: Supported 00:20:51.547 Firmware Activation Notices: Not Supported 00:20:51.547 ANA Change Notices: Not Supported 00:20:51.547 PLE Aggregate Log Change Notices: Not Supported 00:20:51.547 LBA Status Info Alert Notices: Not Supported 00:20:51.547 EGE Aggregate Log Change Notices: Not Supported 00:20:51.547 Normal NVM Subsystem Shutdown event: Not Supported 00:20:51.547 Zone Descriptor Change Notices: Not Supported 00:20:51.547 Discovery Log Change Notices: Not Supported 00:20:51.547 Controller Attributes 00:20:51.547 128-bit Host Identifier: Supported 00:20:51.547 Non-Operational Permissive Mode: Not Supported 00:20:51.547 NVM Sets: Not Supported 00:20:51.547 Read Recovery Levels: Not Supported 00:20:51.547 Endurance Groups: Not Supported 00:20:51.547 Predictable Latency Mode: Not Supported 00:20:51.547 Traffic Based Keep ALive: Not Supported 00:20:51.547 Namespace Granularity: Not Supported 00:20:51.547 SQ Associations: Not Supported 00:20:51.547 UUID List: Not Supported 00:20:51.547 Multi-Domain Subsystem: Not Supported 00:20:51.547 Fixed Capacity Management: Not Supported 00:20:51.547 Variable Capacity Management: Not Supported 00:20:51.547 Delete Endurance Group: Not Supported 00:20:51.547 Delete NVM Set: Not Supported 00:20:51.547 Extended LBA Formats Supported: Not Supported 00:20:51.547 Flexible Data Placement Supported: Not Supported 00:20:51.547 00:20:51.547 Controller Memory Buffer Support 00:20:51.547 ================================ 00:20:51.547 Supported: No 00:20:51.547 00:20:51.547 Persistent Memory Region Support 00:20:51.547 ================================ 00:20:51.547 Supported: No 00:20:51.547 00:20:51.547 Admin Command Set Attributes 00:20:51.547 ============================ 00:20:51.547 Security Send/Receive: Not Supported 00:20:51.547 Format NVM: Not Supported 00:20:51.547 Firmware Activate/Download: Not Supported 00:20:51.547 Namespace Management: Not Supported 00:20:51.547 Device Self-Test: Not Supported 00:20:51.547 Directives: Not Supported 00:20:51.547 NVMe-MI: Not Supported 00:20:51.547 Virtualization Management: Not Supported 00:20:51.547 Doorbell Buffer Config: Not Supported 00:20:51.547 Get LBA Status Capability: Not Supported 00:20:51.547 Command & Feature Lockdown Capability: Not Supported 00:20:51.547 Abort Command Limit: 4 00:20:51.547 Async Event Request Limit: 4 00:20:51.547 Number of Firmware Slots: N/A 00:20:51.547 Firmware Slot 1 Read-Only: N/A 00:20:51.547 Firmware Activation Without Reset: N/A 00:20:51.547 Multiple Update Detection Support: N/A 00:20:51.547 Firmware Update Granularity: No Information Provided 00:20:51.547 Per-Namespace SMART Log: No 00:20:51.547 Asymmetric Namespace Access Log Page: Not Supported 00:20:51.547 Subsystem NQN: nqn.2019-07.io.spdk:cnode1 00:20:51.547 Command Effects Log Page: Supported 00:20:51.547 Get Log Page Extended Data: Supported 00:20:51.547 Telemetry Log Pages: Not Supported 00:20:51.547 Persistent Event Log Pages: Not Supported 00:20:51.547 Supported Log Pages Log Page: May Support 00:20:51.547 Commands Supported & Effects Log Page: Not Supported 00:20:51.547 Feature Identifiers & Effects Log Page:May Support 00:20:51.547 NVMe-MI Commands & Effects Log Page: May Support 00:20:51.547 Data Area 4 for Telemetry Log: Not Supported 00:20:51.547 Error Log Page Entries Supported: 128 00:20:51.547 Keep Alive: Supported 00:20:51.547 Keep Alive Granularity: 10000 ms 00:20:51.547 00:20:51.547 NVM Command Set Attributes 00:20:51.547 ========================== 00:20:51.547 Submission Queue Entry Size 00:20:51.547 Max: 64 00:20:51.547 Min: 64 00:20:51.547 Completion Queue Entry Size 00:20:51.547 Max: 16 00:20:51.547 Min: 16 00:20:51.547 Number of Namespaces: 32 00:20:51.547 Compare Command: Supported 00:20:51.547 Write Uncorrectable Command: Not Supported 00:20:51.547 Dataset Management Command: Supported 00:20:51.547 Write Zeroes Command: Supported 00:20:51.547 Set Features Save Field: Not Supported 00:20:51.547 Reservations: Not Supported 00:20:51.547 Timestamp: Not Supported 00:20:51.547 Copy: Supported 00:20:51.547 Volatile Write Cache: Present 00:20:51.547 Atomic Write Unit (Normal): 1 00:20:51.547 Atomic Write Unit (PFail): 1 00:20:51.547 Atomic Compare & Write Unit: 1 00:20:51.547 Fused Compare & Write: Supported 00:20:51.547 Scatter-Gather List 00:20:51.547 SGL Command Set: Supported (Dword aligned) 00:20:51.547 SGL Keyed: Not Supported 00:20:51.547 SGL Bit Bucket Descriptor: Not Supported 00:20:51.547 SGL Metadata Pointer: Not Supported 00:20:51.547 Oversized SGL: Not Supported 00:20:51.547 SGL Metadata Address: Not Supported 00:20:51.547 SGL Offset: Not Supported 00:20:51.547 Transport SGL Data Block: Not Supported 00:20:51.547 Replay Protected Memory Block: Not Supported 00:20:51.547 00:20:51.547 Firmware Slot Information 00:20:51.547 ========================= 00:20:51.547 Active slot: 1 00:20:51.547 Slot 1 Firmware Revision: 25.01 00:20:51.547 00:20:51.547 00:20:51.547 Commands Supported and Effects 00:20:51.547 ============================== 00:20:51.547 Admin Commands 00:20:51.547 -------------- 00:20:51.547 Get Log Page (02h): Supported 00:20:51.547 Identify (06h): Supported 00:20:51.547 Abort (08h): Supported 00:20:51.547 Set Features (09h): Supported 00:20:51.547 Get Features (0Ah): Supported 00:20:51.547 Asynchronous Event Request (0Ch): Supported 00:20:51.547 Keep Alive (18h): Supported 00:20:51.547 I/O Commands 00:20:51.547 ------------ 00:20:51.547 Flush (00h): Supported LBA-Change 00:20:51.547 Write (01h): Supported LBA-Change 00:20:51.547 Read (02h): Supported 00:20:51.547 Compare (05h): Supported 00:20:51.547 Write Zeroes (08h): Supported LBA-Change 00:20:51.547 Dataset Management (09h): Supported LBA-Change 00:20:51.547 Copy (19h): Supported LBA-Change 00:20:51.547 00:20:51.547 Error Log 00:20:51.547 ========= 00:20:51.547 00:20:51.547 Arbitration 00:20:51.547 =========== 00:20:51.547 Arbitration Burst: 1 00:20:51.547 00:20:51.547 Power Management 00:20:51.547 ================ 00:20:51.547 Number of Power States: 1 00:20:51.547 Current Power State: Power State #0 00:20:51.547 Power State #0: 00:20:51.547 Max Power: 0.00 W 00:20:51.547 Non-Operational State: Operational 00:20:51.547 Entry Latency: Not Reported 00:20:51.547 Exit Latency: Not Reported 00:20:51.547 Relative Read Throughput: 0 00:20:51.547 Relative Read Latency: 0 00:20:51.547 Relative Write Throughput: 0 00:20:51.547 Relative Write Latency: 0 00:20:51.547 Idle Power: Not Reported 00:20:51.547 Active Power: Not Reported 00:20:51.547 Non-Operational Permissive Mode: Not Supported 00:20:51.547 00:20:51.547 Health Information 00:20:51.547 ================== 00:20:51.547 Critical Warnings: 00:20:51.547 Available Spare Space: OK 00:20:51.547 Temperature: OK 00:20:51.547 Device Reliability: OK 00:20:51.547 Read Only: No 00:20:51.547 Volatile Memory Backup: OK 00:20:51.547 Current Temperature: 0 Kelvin (-273 Celsius) 00:20:51.547 Temperature Threshold: 0 Kelvin (-273 Celsius) 00:20:51.547 Available Spare: 0% 00:20:51.547 Available Sp[2024-12-06 17:36:33.136904] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ERROR_RECOVERY cid:184 cdw10:00000005 PRP1 0x0 PRP2 0x0 00:20:51.547 [2024-12-06 17:36:33.136925] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:184 cdw0:0 sqhd:0014 p:1 m:0 dnr:0 00:20:51.547 [2024-12-06 17:36:33.136995] nvme_ctrlr.c:4399:nvme_ctrlr_destruct_async: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] Prepare to destruct SSD 00:20:51.547 [2024-12-06 17:36:33.137038] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:190 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:20:51.547 [2024-12-06 17:36:33.137049] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:189 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:20:51.547 [2024-12-06 17:36:33.137059] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:188 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:20:51.547 [2024-12-06 17:36:33.137068] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:187 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:20:51.547 [2024-12-06 17:36:33.137422] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x14, value 0x460001 00:20:51.547 [2024-12-06 17:36:33.137441] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x14, value 0x464001 00:20:51.547 [2024-12-06 17:36:33.138423] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:20:51.547 [2024-12-06 17:36:33.138511] nvme_ctrlr.c:1151:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] RTD3E = 0 us 00:20:51.547 [2024-12-06 17:36:33.138525] nvme_ctrlr.c:1154:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] shutdown timeout = 10000 ms 00:20:51.548 [2024-12-06 17:36:33.139435] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x1c, value 0x9 00:20:51.548 [2024-12-06 17:36:33.139457] nvme_ctrlr.c:1273:nvme_ctrlr_shutdown_poll_async: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] shutdown complete in 0 milliseconds 00:20:51.548 [2024-12-06 17:36:33.139508] vfio_user_pci.c: 399:spdk_vfio_user_release: *DEBUG*: Release file /var/run/vfio-user/domain/vfio-user1/1/cntrl 00:20:51.548 [2024-12-06 17:36:33.141472] vfio_user_pci.c: 96:vfio_remove_mr: *DEBUG*: Remove memory region: FD 10, VADDR 0x200000200000, IOVA 0x200000200000, Size 0x200000 00:20:51.548 are Threshold: 0% 00:20:51.548 Life Percentage Used: 0% 00:20:51.548 Data Units Read: 0 00:20:51.548 Data Units Written: 0 00:20:51.548 Host Read Commands: 0 00:20:51.548 Host Write Commands: 0 00:20:51.548 Controller Busy Time: 0 minutes 00:20:51.548 Power Cycles: 0 00:20:51.548 Power On Hours: 0 hours 00:20:51.548 Unsafe Shutdowns: 0 00:20:51.548 Unrecoverable Media Errors: 0 00:20:51.548 Lifetime Error Log Entries: 0 00:20:51.548 Warning Temperature Time: 0 minutes 00:20:51.548 Critical Temperature Time: 0 minutes 00:20:51.548 00:20:51.548 Number of Queues 00:20:51.548 ================ 00:20:51.548 Number of I/O Submission Queues: 127 00:20:51.548 Number of I/O Completion Queues: 127 00:20:51.548 00:20:51.548 Active Namespaces 00:20:51.548 ================= 00:20:51.548 Namespace ID:1 00:20:51.548 Error Recovery Timeout: Unlimited 00:20:51.548 Command Set Identifier: NVM (00h) 00:20:51.548 Deallocate: Supported 00:20:51.548 Deallocated/Unwritten Error: Not Supported 00:20:51.548 Deallocated Read Value: Unknown 00:20:51.548 Deallocate in Write Zeroes: Not Supported 00:20:51.548 Deallocated Guard Field: 0xFFFF 00:20:51.548 Flush: Supported 00:20:51.548 Reservation: Supported 00:20:51.548 Namespace Sharing Capabilities: Multiple Controllers 00:20:51.548 Size (in LBAs): 131072 (0GiB) 00:20:51.548 Capacity (in LBAs): 131072 (0GiB) 00:20:51.548 Utilization (in LBAs): 131072 (0GiB) 00:20:51.548 NGUID: 99BB49AF08BC4B73A3D6CDD99170A2EF 00:20:51.548 UUID: 99bb49af-08bc-4b73-a3d6-cdd99170a2ef 00:20:51.548 Thin Provisioning: Not Supported 00:20:51.548 Per-NS Atomic Units: Yes 00:20:51.548 Atomic Boundary Size (Normal): 0 00:20:51.548 Atomic Boundary Size (PFail): 0 00:20:51.548 Atomic Boundary Offset: 0 00:20:51.548 Maximum Single Source Range Length: 65535 00:20:51.548 Maximum Copy Length: 65535 00:20:51.548 Maximum Source Range Count: 1 00:20:51.548 NGUID/EUI64 Never Reused: No 00:20:51.548 Namespace Write Protected: No 00:20:51.548 Number of LBA Formats: 1 00:20:51.548 Current LBA Format: LBA Format #00 00:20:51.548 LBA Format #00: Data Size: 512 Metadata Size: 0 00:20:51.548 00:20:51.548 17:36:33 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -s 256 -g -q 128 -o 4096 -w read -t 5 -c 0x2 00:20:51.806 [2024-12-06 17:36:33.394560] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:20:57.069 Initializing NVMe Controllers 00:20:57.069 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:20:57.069 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) NSID 1 with lcore 1 00:20:57.069 Initialization complete. Launching workers. 00:20:57.069 ======================================================== 00:20:57.069 Latency(us) 00:20:57.069 Device Information : IOPS MiB/s Average min max 00:20:57.069 VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) NSID 1 from core 1: 31426.56 122.76 4076.65 1229.24 10289.82 00:20:57.069 ======================================================== 00:20:57.069 Total : 31426.56 122.76 4076.65 1229.24 10289.82 00:20:57.069 00:20:57.069 [2024-12-06 17:36:38.419000] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:20:57.069 17:36:38 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@85 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -s 256 -g -q 128 -o 4096 -w write -t 5 -c 0x2 00:20:57.069 [2024-12-06 17:36:38.669205] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:21:02.334 Initializing NVMe Controllers 00:21:02.334 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:21:02.334 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) NSID 1 with lcore 1 00:21:02.334 Initialization complete. Launching workers. 00:21:02.334 ======================================================== 00:21:02.334 Latency(us) 00:21:02.334 Device Information : IOPS MiB/s Average min max 00:21:02.334 VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) NSID 1 from core 1: 16027.20 62.61 7994.84 4987.73 15848.45 00:21:02.334 ======================================================== 00:21:02.334 Total : 16027.20 62.61 7994.84 4987.73 15848.45 00:21:02.334 00:21:02.334 [2024-12-06 17:36:43.705634] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:21:02.334 17:36:43 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -g -q 32 -o 4096 -w randrw -M 50 -t 5 -c 0xE 00:21:02.334 [2024-12-06 17:36:43.941815] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:21:07.599 [2024-12-06 17:36:49.010004] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:21:07.599 Initializing NVMe Controllers 00:21:07.599 Attaching to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:21:07.599 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:21:07.599 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) with lcore 1 00:21:07.599 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) with lcore 2 00:21:07.599 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) with lcore 3 00:21:07.599 Initialization complete. Launching workers. 00:21:07.599 Starting thread on core 2 00:21:07.599 Starting thread on core 3 00:21:07.599 Starting thread on core 1 00:21:07.599 17:36:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@87 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration -t 3 -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -d 256 -g 00:21:07.599 [2024-12-06 17:36:49.334184] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:21:10.884 [2024-12-06 17:36:52.387068] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:21:10.884 Initializing NVMe Controllers 00:21:10.884 Attaching to /var/run/vfio-user/domain/vfio-user1/1 00:21:10.884 Attached to /var/run/vfio-user/domain/vfio-user1/1 00:21:10.884 Associating SPDK bdev Controller (SPDK1 ) with lcore 0 00:21:10.884 Associating SPDK bdev Controller (SPDK1 ) with lcore 1 00:21:10.884 Associating SPDK bdev Controller (SPDK1 ) with lcore 2 00:21:10.884 Associating SPDK bdev Controller (SPDK1 ) with lcore 3 00:21:10.884 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration run with configuration: 00:21:10.884 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration -q 64 -s 131072 -w randrw -M 50 -l 0 -t 3 -c 0xf -m 0 -a 0 -b 0 -n 100000 -i -1 00:21:10.884 Initialization complete. Launching workers. 00:21:10.884 Starting thread on core 1 with urgent priority queue 00:21:10.884 Starting thread on core 2 with urgent priority queue 00:21:10.884 Starting thread on core 3 with urgent priority queue 00:21:10.884 Starting thread on core 0 with urgent priority queue 00:21:10.884 SPDK bdev Controller (SPDK1 ) core 0: 5292.33 IO/s 18.90 secs/100000 ios 00:21:10.884 SPDK bdev Controller (SPDK1 ) core 1: 5719.00 IO/s 17.49 secs/100000 ios 00:21:10.884 SPDK bdev Controller (SPDK1 ) core 2: 5821.33 IO/s 17.18 secs/100000 ios 00:21:10.884 SPDK bdev Controller (SPDK1 ) core 3: 4517.00 IO/s 22.14 secs/100000 ios 00:21:10.884 ======================================================== 00:21:10.884 00:21:10.884 17:36:52 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/hello_world -d 256 -g -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' 00:21:10.884 [2024-12-06 17:36:52.704165] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:21:11.142 Initializing NVMe Controllers 00:21:11.142 Attaching to /var/run/vfio-user/domain/vfio-user1/1 00:21:11.142 Attached to /var/run/vfio-user/domain/vfio-user1/1 00:21:11.142 Namespace ID: 1 size: 0GB 00:21:11.142 Initialization complete. 00:21:11.142 INFO: using host memory buffer for IO 00:21:11.142 Hello world! 00:21:11.142 [2024-12-06 17:36:52.737765] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:21:11.142 17:36:52 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/overhead/overhead -o 4096 -t 1 -H -g -d 256 -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' 00:21:11.398 [2024-12-06 17:36:53.038161] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:21:12.331 Initializing NVMe Controllers 00:21:12.331 Attaching to /var/run/vfio-user/domain/vfio-user1/1 00:21:12.331 Attached to /var/run/vfio-user/domain/vfio-user1/1 00:21:12.331 Initialization complete. Launching workers. 00:21:12.331 submit (in ns) avg, min, max = 8392.4, 3523.3, 4018184.4 00:21:12.331 complete (in ns) avg, min, max = 27750.3, 2077.8, 4073778.9 00:21:12.331 00:21:12.331 Submit histogram 00:21:12.331 ================ 00:21:12.331 Range in us Cumulative Count 00:21:12.331 3.508 - 3.532: 0.0787% ( 10) 00:21:12.331 3.532 - 3.556: 0.5117% ( 55) 00:21:12.331 3.556 - 3.579: 1.9602% ( 184) 00:21:12.331 3.579 - 3.603: 4.4163% ( 312) 00:21:12.331 3.603 - 3.627: 10.2810% ( 745) 00:21:12.331 3.627 - 3.650: 18.0981% ( 993) 00:21:12.331 3.650 - 3.674: 28.1823% ( 1281) 00:21:12.331 3.674 - 3.698: 36.7236% ( 1085) 00:21:12.331 3.698 - 3.721: 44.8792% ( 1036) 00:21:12.331 3.721 - 3.745: 50.3346% ( 693) 00:21:12.331 3.745 - 3.769: 55.2153% ( 620) 00:21:12.331 3.769 - 3.793: 59.0569% ( 488) 00:21:12.331 3.793 - 3.816: 62.8119% ( 477) 00:21:12.331 3.816 - 3.840: 66.5984% ( 481) 00:21:12.331 3.840 - 3.864: 70.1960% ( 457) 00:21:12.331 3.864 - 3.887: 74.4312% ( 538) 00:21:12.331 3.887 - 3.911: 78.8160% ( 557) 00:21:12.331 3.911 - 3.935: 82.7521% ( 500) 00:21:12.331 3.935 - 3.959: 85.2082% ( 312) 00:21:12.331 3.959 - 3.982: 87.1763% ( 250) 00:21:12.331 3.982 - 4.006: 88.9869% ( 230) 00:21:12.331 4.006 - 4.030: 90.5298% ( 196) 00:21:12.331 4.030 - 4.053: 91.6949% ( 148) 00:21:12.331 4.053 - 4.077: 92.8127% ( 142) 00:21:12.331 4.077 - 4.101: 93.8125% ( 127) 00:21:12.331 4.101 - 4.124: 94.5997% ( 100) 00:21:12.331 4.124 - 4.148: 95.1822% ( 74) 00:21:12.331 4.148 - 4.172: 95.7569% ( 73) 00:21:12.331 4.172 - 4.196: 96.1112% ( 45) 00:21:12.331 4.196 - 4.219: 96.3001% ( 24) 00:21:12.331 4.219 - 4.243: 96.5284% ( 29) 00:21:12.331 4.243 - 4.267: 96.6701% ( 18) 00:21:12.331 4.267 - 4.290: 96.7882% ( 15) 00:21:12.331 4.290 - 4.314: 96.8905% ( 13) 00:21:12.331 4.314 - 4.338: 96.9535% ( 8) 00:21:12.331 4.338 - 4.361: 96.9928% ( 5) 00:21:12.331 4.361 - 4.385: 97.0479% ( 7) 00:21:12.331 4.385 - 4.409: 97.0794% ( 4) 00:21:12.331 4.409 - 4.433: 97.1030% ( 3) 00:21:12.331 4.433 - 4.456: 97.1424% ( 5) 00:21:12.331 4.456 - 4.480: 97.1896% ( 6) 00:21:12.331 4.504 - 4.527: 97.1975% ( 1) 00:21:12.331 4.527 - 4.551: 97.2054% ( 1) 00:21:12.331 4.551 - 4.575: 97.2133% ( 1) 00:21:12.331 4.693 - 4.717: 97.2369% ( 3) 00:21:12.331 4.717 - 4.741: 97.2605% ( 3) 00:21:12.331 4.741 - 4.764: 97.2841% ( 3) 00:21:12.331 4.764 - 4.788: 97.3156% ( 4) 00:21:12.331 4.788 - 4.812: 97.3471% ( 4) 00:21:12.331 4.812 - 4.836: 97.3786% ( 4) 00:21:12.331 4.836 - 4.859: 97.4101% ( 4) 00:21:12.331 4.859 - 4.883: 97.4337% ( 3) 00:21:12.331 4.883 - 4.907: 97.5045% ( 9) 00:21:12.331 4.907 - 4.930: 97.5360% ( 4) 00:21:12.331 4.930 - 4.954: 97.5990% ( 8) 00:21:12.331 4.954 - 4.978: 97.6620% ( 8) 00:21:12.331 4.978 - 5.001: 97.7092% ( 6) 00:21:12.331 5.001 - 5.025: 97.7407% ( 4) 00:21:12.331 5.025 - 5.049: 97.7879% ( 6) 00:21:12.331 5.049 - 5.073: 97.8430% ( 7) 00:21:12.331 5.073 - 5.096: 97.8824% ( 5) 00:21:12.331 5.096 - 5.120: 97.8903% ( 1) 00:21:12.331 5.120 - 5.144: 97.8981% ( 1) 00:21:12.331 5.144 - 5.167: 97.9139% ( 2) 00:21:12.331 5.167 - 5.191: 97.9375% ( 3) 00:21:12.331 5.191 - 5.215: 97.9611% ( 3) 00:21:12.331 5.215 - 5.239: 97.9847% ( 3) 00:21:12.331 5.239 - 5.262: 98.0005% ( 2) 00:21:12.331 5.262 - 5.286: 98.0241% ( 3) 00:21:12.331 5.286 - 5.310: 98.0320% ( 1) 00:21:12.331 5.310 - 5.333: 98.0556% ( 3) 00:21:12.331 5.333 - 5.357: 98.0713% ( 2) 00:21:12.331 5.357 - 5.381: 98.0792% ( 1) 00:21:12.331 5.381 - 5.404: 98.0949% ( 2) 00:21:12.331 5.428 - 5.452: 98.1107% ( 2) 00:21:12.331 5.452 - 5.476: 98.1186% ( 1) 00:21:12.331 5.499 - 5.523: 98.1264% ( 1) 00:21:12.331 5.784 - 5.807: 98.1343% ( 1) 00:21:12.331 5.855 - 5.879: 98.1422% ( 1) 00:21:12.331 5.950 - 5.973: 98.1658% ( 3) 00:21:12.331 6.210 - 6.258: 98.1737% ( 1) 00:21:12.331 6.258 - 6.305: 98.1815% ( 1) 00:21:12.331 6.447 - 6.495: 98.1973% ( 2) 00:21:12.331 6.495 - 6.542: 98.2130% ( 2) 00:21:12.331 6.969 - 7.016: 98.2209% ( 1) 00:21:12.331 7.206 - 7.253: 98.2288% ( 1) 00:21:12.331 7.253 - 7.301: 98.2366% ( 1) 00:21:12.331 7.301 - 7.348: 98.2524% ( 2) 00:21:12.331 7.348 - 7.396: 98.2603% ( 1) 00:21:12.331 7.396 - 7.443: 98.2681% ( 1) 00:21:12.331 7.490 - 7.538: 98.2760% ( 1) 00:21:12.331 7.680 - 7.727: 98.2839% ( 1) 00:21:12.331 7.727 - 7.775: 98.2917% ( 1) 00:21:12.331 7.775 - 7.822: 98.2996% ( 1) 00:21:12.331 7.822 - 7.870: 98.3075% ( 1) 00:21:12.331 7.917 - 7.964: 98.3154% ( 1) 00:21:12.331 8.012 - 8.059: 98.3390% ( 3) 00:21:12.331 8.059 - 8.107: 98.3468% ( 1) 00:21:12.331 8.107 - 8.154: 98.3547% ( 1) 00:21:12.331 8.154 - 8.201: 98.3626% ( 1) 00:21:12.331 8.201 - 8.249: 98.3705% ( 1) 00:21:12.331 8.249 - 8.296: 98.3783% ( 1) 00:21:12.331 8.344 - 8.391: 98.3862% ( 1) 00:21:12.331 8.391 - 8.439: 98.3941% ( 1) 00:21:12.331 8.533 - 8.581: 98.4177% ( 3) 00:21:12.331 8.581 - 8.628: 98.4256% ( 1) 00:21:12.331 8.628 - 8.676: 98.4334% ( 1) 00:21:12.331 8.676 - 8.723: 98.4413% ( 1) 00:21:12.331 8.770 - 8.818: 98.4492% ( 1) 00:21:12.331 8.818 - 8.865: 98.4571% ( 1) 00:21:12.331 8.865 - 8.913: 98.4649% ( 1) 00:21:12.331 8.913 - 8.960: 98.4728% ( 1) 00:21:12.331 8.960 - 9.007: 98.4807% ( 1) 00:21:12.331 9.055 - 9.102: 98.4964% ( 2) 00:21:12.331 9.339 - 9.387: 98.5122% ( 2) 00:21:12.331 9.387 - 9.434: 98.5200% ( 1) 00:21:12.331 9.434 - 9.481: 98.5279% ( 1) 00:21:12.331 9.481 - 9.529: 98.5437% ( 2) 00:21:12.331 9.529 - 9.576: 98.5594% ( 2) 00:21:12.331 9.576 - 9.624: 98.5673% ( 1) 00:21:12.331 9.766 - 9.813: 98.5830% ( 2) 00:21:12.331 10.098 - 10.145: 98.5988% ( 2) 00:21:12.331 10.382 - 10.430: 98.6066% ( 1) 00:21:12.331 10.430 - 10.477: 98.6145% ( 1) 00:21:12.331 10.524 - 10.572: 98.6224% ( 1) 00:21:12.331 10.761 - 10.809: 98.6302% ( 1) 00:21:12.331 10.904 - 10.951: 98.6381% ( 1) 00:21:12.331 10.951 - 10.999: 98.6460% ( 1) 00:21:12.331 11.046 - 11.093: 98.6539% ( 1) 00:21:12.331 11.093 - 11.141: 98.6617% ( 1) 00:21:12.331 11.283 - 11.330: 98.6696% ( 1) 00:21:12.331 11.567 - 11.615: 98.6775% ( 1) 00:21:12.331 11.710 - 11.757: 98.6853% ( 1) 00:21:12.331 11.804 - 11.852: 98.6932% ( 1) 00:21:12.331 11.852 - 11.899: 98.7011% ( 1) 00:21:12.331 11.899 - 11.947: 98.7090% ( 1) 00:21:12.331 11.947 - 11.994: 98.7168% ( 1) 00:21:12.331 12.136 - 12.231: 98.7247% ( 1) 00:21:12.331 12.326 - 12.421: 98.7326% ( 1) 00:21:12.331 12.610 - 12.705: 98.7483% ( 2) 00:21:12.331 13.084 - 13.179: 98.7641% ( 2) 00:21:12.331 13.179 - 13.274: 98.7719% ( 1) 00:21:12.331 13.369 - 13.464: 98.7798% ( 1) 00:21:12.331 13.653 - 13.748: 98.7956% ( 2) 00:21:12.331 13.748 - 13.843: 98.8034% ( 1) 00:21:12.331 13.938 - 14.033: 98.8192% ( 2) 00:21:12.331 14.317 - 14.412: 98.8270% ( 1) 00:21:12.331 14.412 - 14.507: 98.8349% ( 1) 00:21:12.331 14.601 - 14.696: 98.8428% ( 1) 00:21:12.331 14.791 - 14.886: 98.8507% ( 1) 00:21:12.331 14.981 - 15.076: 98.8585% ( 1) 00:21:12.331 15.170 - 15.265: 98.8664% ( 1) 00:21:12.331 15.550 - 15.644: 98.8743% ( 1) 00:21:12.331 16.972 - 17.067: 98.8822% ( 1) 00:21:12.331 17.161 - 17.256: 98.8900% ( 1) 00:21:12.331 17.256 - 17.351: 98.9215% ( 4) 00:21:12.331 17.351 - 17.446: 98.9294% ( 1) 00:21:12.331 17.446 - 17.541: 98.9924% ( 8) 00:21:12.331 17.541 - 17.636: 99.0317% ( 5) 00:21:12.331 17.636 - 17.730: 99.0711% ( 5) 00:21:12.331 17.730 - 17.825: 99.1341% ( 8) 00:21:12.331 17.825 - 17.920: 99.1419% ( 1) 00:21:12.331 17.920 - 18.015: 99.2521% ( 14) 00:21:12.331 18.015 - 18.110: 99.3230% ( 9) 00:21:12.331 18.110 - 18.204: 99.4096% ( 11) 00:21:12.331 18.204 - 18.299: 99.4568% ( 6) 00:21:12.331 18.299 - 18.394: 99.5355% ( 10) 00:21:12.331 18.394 - 18.489: 99.5985% ( 8) 00:21:12.331 18.489 - 18.584: 99.6536% ( 7) 00:21:12.331 18.584 - 18.679: 99.7087% ( 7) 00:21:12.331 18.679 - 18.773: 99.7323% ( 3) 00:21:12.331 18.773 - 18.868: 99.7717% ( 5) 00:21:12.331 18.868 - 18.963: 99.7875% ( 2) 00:21:12.331 19.058 - 19.153: 99.8032% ( 2) 00:21:12.331 19.247 - 19.342: 99.8189% ( 2) 00:21:12.331 19.342 - 19.437: 99.8347% ( 2) 00:21:12.331 19.532 - 19.627: 99.8426% ( 1) 00:21:12.331 20.006 - 20.101: 99.8504% ( 1) 00:21:12.331 20.101 - 20.196: 99.8583% ( 1) 00:21:12.331 20.196 - 20.290: 99.8662% ( 1) 00:21:12.331 20.575 - 20.670: 99.8740% ( 1) 00:21:12.331 27.686 - 27.876: 99.8819% ( 1) 00:21:12.331 28.255 - 28.444: 99.8898% ( 1) 00:21:12.331 3980.705 - 4004.978: 99.9764% ( 11) 00:21:12.331 4004.978 - 4029.250: 100.0000% ( 3) 00:21:12.331 00:21:12.331 Complete histogram 00:21:12.331 ================== 00:21:12.331 Range in us Cumulative Count 00:21:12.331 2.074 - 2.086: 1.4563% ( 185) 00:21:12.331 2.086 - 2.098: 26.1041% ( 3131) 00:21:12.331 2.098 - 2.110: 35.6373% ( 1211) 00:21:12.331 2.110 - 2.121: 41.3446% ( 725) 00:21:12.331 2.121 - 2.133: 55.2625% ( 1768) 00:21:12.331 2.133 - 2.145: 58.4586% ( 406) 00:21:12.331 2.145 - 2.157: 62.3239% ( 491) 00:21:12.331 2.157 - 2.169: 71.5579% ( 1173) 00:21:12.331 2.169 - 2.181: 73.4000% ( 234) 00:21:12.331 2.181 - 2.193: 76.2812% ( 366) 00:21:12.331 2.193 - 2.204: 80.5636% ( 544) 00:21:12.331 2.204 - 2.216: 81.3745% ( 103) 00:21:12.331 2.216 - 2.228: 82.7600% ( 176) 00:21:12.331 2.228 - 2.240: 86.9952% ( 538) 00:21:12.331 2.240 - 2.252: 89.6166% ( 333) 00:21:12.331 2.252 - 2.264: 90.9391% ( 168) 00:21:12.331 2.264 - 2.276: 92.8048% ( 237) 00:21:12.331 2.276 - 2.287: 93.4976% ( 88) 00:21:12.331 2.287 - 2.299: 93.8597% ( 46) 00:21:12.331 2.299 - 2.311: 94.4108% ( 70) 00:21:12.331 2.311 - 2.323: 95.3003% ( 113) 00:21:12.331 2.323 - 2.335: 95.5601% ( 33) 00:21:12.331 2.335 - 2.347: 95.5916% ( 4) 00:21:12.331 2.347 - 2.359: 95.6152% ( 3) 00:21:12.331 2.359 - 2.370: 95.6546% ( 5) 00:21:12.331 2.370 - 2.382: 95.7569% ( 13) 00:21:12.331 2.382 - 2.394: 95.9931% ( 30) 00:21:12.331 2.394 - 2.406: 96.2450% ( 32) 00:21:12.331 2.406 - 2.418: 96.4024% ( 20) 00:21:12.331 2.418 - 2.430: 96.5835% ( 23) 00:21:12.331 2.430 - 2.441: 96.7803% ( 25) 00:21:12.331 2.441 - 2.453: 97.0322% ( 32) 00:21:12.331 2.453 - 2.465: 97.2369% ( 26) 00:21:12.331 2.465 - 2.477: 97.3707% ( 17) 00:21:12.331 2.477 - 2.489: 97.5675% ( 25) 00:21:12.331 2.489 - 2.501: 97.7407% ( 22) 00:21:12.331 2.501 - 2.513: 97.9454% ( 26) 00:21:12.331 2.513 - 2.524: 98.0320% ( 11) 00:21:12.331 2.524 - 2.536: 98.1028% ( 9) 00:21:12.331 2.536 - 2.548: 98.1894% ( 11) 00:21:12.331 2.548 - 2.560: 98.2445% ( 7) 00:21:12.331 2.560 - 2.572: 98.3075% ( 8) 00:21:12.331 2.572 - 2.584: 98.3468% ( 5) 00:21:12.331 2.584 - 2.596: 98.3705% ( 3) 00:21:12.331 2.596 - 2.607: 98.3941% ( 3) 00:21:12.331 2.607 - 2.619: 98.4098% ( 2) 00:21:12.331 2.619 - 2.631: 98.4256% ( 2) 00:21:12.331 2.655 - 2.667: 98.4571% ( 4) 00:21:12.331 2.667 - 2.679: 98.4649% ( 1) 00:21:12.331 2.679 - 2.690: 98.4728% ( 1) 00:21:12.331 2.690 - 2.702: 98.4807% ( 1) 00:21:12.331 2.761 - 2.773: 98.4885% ( 1) 00:21:12.331 3.390 - 3.413: 98.4964% ( 1) 00:21:12.331 3.413 - 3.437: 98.5043% ( 1) 00:21:12.331 3.437 - 3.461: 98.5122% ( 1) 00:21:12.331 3.461 - 3.484: 98.5200% ( 1) 00:21:12.331 3.484 - 3.508: 98.5437% ( 3) 00:21:12.331 3.532 - 3.556: 98.5515% ( 1) 00:21:12.332 3.556 - 3.579: 98.5594% ( 1) 00:21:12.332 3.579 - 3.603: 98.5673% ( 1) 00:21:12.332 3.627 - 3.650: 98.5830% ( 2) 00:21:12.332 3.650 - 3.674: 98.5909% ( 1) 00:21:12.332 3.674 - 3.698: 98.5988% ( 1) 00:21:12.332 3.698 - 3.721: 98.6145% ( 2) 00:21:12.332 3.745 - 3.769: 98.6224% ( 1) 00:21:12.332 3.840 - 3.864: 98.6302% ( 1) 00:21:12.332 3.864 - 3.887: 9[2024-12-06 17:36:54.061313] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:21:12.332 8.6381% ( 1) 00:21:12.332 3.935 - 3.959: 98.6460% ( 1) 00:21:12.332 3.982 - 4.006: 98.6539% ( 1) 00:21:12.332 4.030 - 4.053: 98.6617% ( 1) 00:21:12.332 4.053 - 4.077: 98.6696% ( 1) 00:21:12.332 4.077 - 4.101: 98.6775% ( 1) 00:21:12.332 4.101 - 4.124: 98.6853% ( 1) 00:21:12.332 4.124 - 4.148: 98.6932% ( 1) 00:21:12.332 4.290 - 4.314: 98.7011% ( 1) 00:21:12.332 4.314 - 4.338: 98.7090% ( 1) 00:21:12.332 4.741 - 4.764: 98.7168% ( 1) 00:21:12.332 5.736 - 5.760: 98.7247% ( 1) 00:21:12.332 6.400 - 6.447: 98.7326% ( 1) 00:21:12.332 6.732 - 6.779: 98.7405% ( 1) 00:21:12.332 6.827 - 6.874: 98.7483% ( 1) 00:21:12.332 6.874 - 6.921: 98.7562% ( 1) 00:21:12.332 7.443 - 7.490: 98.7641% ( 1) 00:21:12.332 7.585 - 7.633: 98.7719% ( 1) 00:21:12.332 7.964 - 8.012: 98.7798% ( 1) 00:21:12.332 8.059 - 8.107: 98.7877% ( 1) 00:21:12.332 8.249 - 8.296: 98.7956% ( 1) 00:21:12.332 8.486 - 8.533: 98.8034% ( 1) 00:21:12.332 9.719 - 9.766: 98.8113% ( 1) 00:21:12.332 10.050 - 10.098: 98.8192% ( 1) 00:21:12.332 10.287 - 10.335: 98.8270% ( 1) 00:21:12.332 12.231 - 12.326: 98.8349% ( 1) 00:21:12.332 15.455 - 15.550: 98.8585% ( 3) 00:21:12.332 15.550 - 15.644: 98.8664% ( 1) 00:21:12.332 15.644 - 15.739: 98.8743% ( 1) 00:21:12.332 15.739 - 15.834: 98.9294% ( 7) 00:21:12.332 15.834 - 15.929: 98.9609% ( 4) 00:21:12.332 15.929 - 16.024: 98.9924% ( 4) 00:21:12.332 16.024 - 16.119: 99.0160% ( 3) 00:21:12.332 16.119 - 16.213: 99.0396% ( 3) 00:21:12.332 16.213 - 16.308: 99.0711% ( 4) 00:21:12.332 16.308 - 16.403: 99.1104% ( 5) 00:21:12.332 16.403 - 16.498: 99.1419% ( 4) 00:21:12.332 16.498 - 16.593: 99.1813% ( 5) 00:21:12.332 16.593 - 16.687: 99.2049% ( 3) 00:21:12.332 16.687 - 16.782: 99.2600% ( 7) 00:21:12.332 16.782 - 16.877: 99.2836% ( 3) 00:21:12.332 16.877 - 16.972: 99.2915% ( 1) 00:21:12.332 17.161 - 17.256: 99.3073% ( 2) 00:21:12.332 17.256 - 17.351: 99.3230% ( 2) 00:21:12.332 17.446 - 17.541: 99.3309% ( 1) 00:21:12.332 17.636 - 17.730: 99.3387% ( 1) 00:21:12.332 17.730 - 17.825: 99.3466% ( 1) 00:21:12.332 18.204 - 18.299: 99.3545% ( 1) 00:21:12.332 18.868 - 18.963: 99.3624% ( 1) 00:21:12.332 3956.433 - 3980.705: 99.3702% ( 1) 00:21:12.332 3980.705 - 4004.978: 99.9055% ( 68) 00:21:12.332 4004.978 - 4029.250: 99.9921% ( 11) 00:21:12.332 4053.523 - 4077.796: 100.0000% ( 1) 00:21:12.332 00:21:12.332 17:36:54 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@90 -- # aer_vfio_user /var/run/vfio-user/domain/vfio-user1/1 nqn.2019-07.io.spdk:cnode1 1 00:21:12.332 17:36:54 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@22 -- # local traddr=/var/run/vfio-user/domain/vfio-user1/1 00:21:12.332 17:36:54 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@23 -- # local subnqn=nqn.2019-07.io.spdk:cnode1 00:21:12.332 17:36:54 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@24 -- # local malloc_num=Malloc3 00:21:12.332 17:36:54 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_get_subsystems 00:21:12.589 [ 00:21:12.589 { 00:21:12.589 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:21:12.589 "subtype": "Discovery", 00:21:12.589 "listen_addresses": [], 00:21:12.589 "allow_any_host": true, 00:21:12.589 "hosts": [] 00:21:12.589 }, 00:21:12.589 { 00:21:12.589 "nqn": "nqn.2019-07.io.spdk:cnode1", 00:21:12.589 "subtype": "NVMe", 00:21:12.589 "listen_addresses": [ 00:21:12.589 { 00:21:12.589 "trtype": "VFIOUSER", 00:21:12.589 "adrfam": "IPv4", 00:21:12.589 "traddr": "/var/run/vfio-user/domain/vfio-user1/1", 00:21:12.589 "trsvcid": "0" 00:21:12.589 } 00:21:12.589 ], 00:21:12.589 "allow_any_host": true, 00:21:12.589 "hosts": [], 00:21:12.589 "serial_number": "SPDK1", 00:21:12.589 "model_number": "SPDK bdev Controller", 00:21:12.589 "max_namespaces": 32, 00:21:12.589 "min_cntlid": 1, 00:21:12.589 "max_cntlid": 65519, 00:21:12.589 "namespaces": [ 00:21:12.589 { 00:21:12.589 "nsid": 1, 00:21:12.589 "bdev_name": "Malloc1", 00:21:12.589 "name": "Malloc1", 00:21:12.589 "nguid": "99BB49AF08BC4B73A3D6CDD99170A2EF", 00:21:12.589 "uuid": "99bb49af-08bc-4b73-a3d6-cdd99170a2ef" 00:21:12.589 } 00:21:12.589 ] 00:21:12.589 }, 00:21:12.589 { 00:21:12.589 "nqn": "nqn.2019-07.io.spdk:cnode2", 00:21:12.589 "subtype": "NVMe", 00:21:12.589 "listen_addresses": [ 00:21:12.589 { 00:21:12.589 "trtype": "VFIOUSER", 00:21:12.589 "adrfam": "IPv4", 00:21:12.589 "traddr": "/var/run/vfio-user/domain/vfio-user2/2", 00:21:12.589 "trsvcid": "0" 00:21:12.589 } 00:21:12.589 ], 00:21:12.589 "allow_any_host": true, 00:21:12.589 "hosts": [], 00:21:12.589 "serial_number": "SPDK2", 00:21:12.589 "model_number": "SPDK bdev Controller", 00:21:12.589 "max_namespaces": 32, 00:21:12.589 "min_cntlid": 1, 00:21:12.589 "max_cntlid": 65519, 00:21:12.589 "namespaces": [ 00:21:12.589 { 00:21:12.589 "nsid": 1, 00:21:12.589 "bdev_name": "Malloc2", 00:21:12.589 "name": "Malloc2", 00:21:12.589 "nguid": "064BC5F836974A08AA1E011B1094E1E2", 00:21:12.589 "uuid": "064bc5f8-3697-4a08-aa1e-011b1094e1e2" 00:21:12.589 } 00:21:12.589 ] 00:21:12.589 } 00:21:12.589 ] 00:21:12.589 17:36:54 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@27 -- # AER_TOUCH_FILE=/tmp/aer_touch_file 00:21:12.589 17:36:54 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@34 -- # aerpid=242050 00:21:12.589 17:36:54 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/aer/aer -r ' trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -n 2 -g -t /tmp/aer_touch_file 00:21:12.589 17:36:54 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@37 -- # waitforfile /tmp/aer_touch_file 00:21:12.589 17:36:54 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1269 -- # local i=0 00:21:12.589 17:36:54 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1270 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:21:12.589 17:36:54 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1271 -- # '[' 0 -lt 200 ']' 00:21:12.589 17:36:54 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1272 -- # i=1 00:21:12.589 17:36:54 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1273 -- # sleep 0.1 00:21:12.846 17:36:54 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1270 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:21:12.846 17:36:54 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1271 -- # '[' 1 -lt 200 ']' 00:21:12.846 17:36:54 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1272 -- # i=2 00:21:12.846 17:36:54 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1273 -- # sleep 0.1 00:21:12.846 [2024-12-06 17:36:54.552858] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:21:12.846 17:36:54 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1270 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:21:12.846 17:36:54 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1276 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:21:12.846 17:36:54 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1280 -- # return 0 00:21:12.846 17:36:54 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@38 -- # rm -f /tmp/aer_touch_file 00:21:12.846 17:36:54 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 --name Malloc3 00:21:13.103 Malloc3 00:21:13.103 17:36:54 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode1 Malloc3 -n 2 00:21:13.369 [2024-12-06 17:36:55.156420] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:21:13.369 17:36:55 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_get_subsystems 00:21:13.369 Asynchronous Event Request test 00:21:13.369 Attaching to /var/run/vfio-user/domain/vfio-user1/1 00:21:13.369 Attached to /var/run/vfio-user/domain/vfio-user1/1 00:21:13.369 Registering asynchronous event callbacks... 00:21:13.369 Starting namespace attribute notice tests for all controllers... 00:21:13.369 /var/run/vfio-user/domain/vfio-user1/1: aer_cb for log page 4, aen_event_type: 0x02, aen_event_info: 0x00 00:21:13.369 aer_cb - Changed Namespace 00:21:13.369 Cleaning up... 00:21:13.635 [ 00:21:13.635 { 00:21:13.635 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:21:13.635 "subtype": "Discovery", 00:21:13.635 "listen_addresses": [], 00:21:13.635 "allow_any_host": true, 00:21:13.635 "hosts": [] 00:21:13.635 }, 00:21:13.635 { 00:21:13.635 "nqn": "nqn.2019-07.io.spdk:cnode1", 00:21:13.635 "subtype": "NVMe", 00:21:13.635 "listen_addresses": [ 00:21:13.635 { 00:21:13.635 "trtype": "VFIOUSER", 00:21:13.635 "adrfam": "IPv4", 00:21:13.635 "traddr": "/var/run/vfio-user/domain/vfio-user1/1", 00:21:13.635 "trsvcid": "0" 00:21:13.635 } 00:21:13.635 ], 00:21:13.635 "allow_any_host": true, 00:21:13.635 "hosts": [], 00:21:13.635 "serial_number": "SPDK1", 00:21:13.635 "model_number": "SPDK bdev Controller", 00:21:13.635 "max_namespaces": 32, 00:21:13.635 "min_cntlid": 1, 00:21:13.635 "max_cntlid": 65519, 00:21:13.635 "namespaces": [ 00:21:13.635 { 00:21:13.635 "nsid": 1, 00:21:13.635 "bdev_name": "Malloc1", 00:21:13.635 "name": "Malloc1", 00:21:13.635 "nguid": "99BB49AF08BC4B73A3D6CDD99170A2EF", 00:21:13.635 "uuid": "99bb49af-08bc-4b73-a3d6-cdd99170a2ef" 00:21:13.635 }, 00:21:13.635 { 00:21:13.635 "nsid": 2, 00:21:13.635 "bdev_name": "Malloc3", 00:21:13.635 "name": "Malloc3", 00:21:13.635 "nguid": "166E6D91FF4143A8B1281BD92280B027", 00:21:13.635 "uuid": "166e6d91-ff41-43a8-b128-1bd92280b027" 00:21:13.635 } 00:21:13.635 ] 00:21:13.635 }, 00:21:13.635 { 00:21:13.635 "nqn": "nqn.2019-07.io.spdk:cnode2", 00:21:13.635 "subtype": "NVMe", 00:21:13.635 "listen_addresses": [ 00:21:13.635 { 00:21:13.635 "trtype": "VFIOUSER", 00:21:13.635 "adrfam": "IPv4", 00:21:13.635 "traddr": "/var/run/vfio-user/domain/vfio-user2/2", 00:21:13.635 "trsvcid": "0" 00:21:13.635 } 00:21:13.635 ], 00:21:13.635 "allow_any_host": true, 00:21:13.635 "hosts": [], 00:21:13.635 "serial_number": "SPDK2", 00:21:13.635 "model_number": "SPDK bdev Controller", 00:21:13.635 "max_namespaces": 32, 00:21:13.635 "min_cntlid": 1, 00:21:13.635 "max_cntlid": 65519, 00:21:13.635 "namespaces": [ 00:21:13.635 { 00:21:13.635 "nsid": 1, 00:21:13.635 "bdev_name": "Malloc2", 00:21:13.635 "name": "Malloc2", 00:21:13.635 "nguid": "064BC5F836974A08AA1E011B1094E1E2", 00:21:13.635 "uuid": "064bc5f8-3697-4a08-aa1e-011b1094e1e2" 00:21:13.635 } 00:21:13.635 ] 00:21:13.635 } 00:21:13.635 ] 00:21:13.635 17:36:55 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@44 -- # wait 242050 00:21:13.635 17:36:55 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@80 -- # for i in $(seq 1 $NUM_DEVICES) 00:21:13.635 17:36:55 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@81 -- # test_traddr=/var/run/vfio-user/domain/vfio-user2/2 00:21:13.635 17:36:55 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@82 -- # test_subnqn=nqn.2019-07.io.spdk:cnode2 00:21:13.635 17:36:55 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -g -L nvme -L nvme_vfio -L vfio_pci 00:21:13.635 [2024-12-06 17:36:55.459784] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:21:13.635 [2024-12-06 17:36:55.459823] [ DPDK EAL parameters: identify --no-shconf -c 0x1 -n 1 -m 0 --no-pci --single-file-segments --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid242184 ] 00:21:13.894 [2024-12-06 17:36:55.509571] nvme_vfio_user.c: 259:nvme_vfio_ctrlr_scan: *DEBUG*: Scan controller : /var/run/vfio-user/domain/vfio-user2/2 00:21:13.894 [2024-12-06 17:36:55.517981] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 0, Size 0x2000, Offset 0x0, Flags 0xf, Cap offset 32 00:21:13.894 [2024-12-06 17:36:55.518011] vfio_user_pci.c: 233:vfio_device_setup_sparse_mmaps: *DEBUG*: Sparse region 0, Size 0x1000, Offset 0x1000, Map addr 0x7f203e6cb000 00:21:13.894 [2024-12-06 17:36:55.518974] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 1, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:21:13.894 [2024-12-06 17:36:55.519979] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 2, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:21:13.894 [2024-12-06 17:36:55.520964] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 3, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:21:13.894 [2024-12-06 17:36:55.521998] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 4, Size 0x2000, Offset 0x0, Flags 0x3, Cap offset 0 00:21:13.894 [2024-12-06 17:36:55.522986] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 5, Size 0x1000, Offset 0x0, Flags 0x3, Cap offset 0 00:21:13.895 [2024-12-06 17:36:55.523989] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 6, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:21:13.895 [2024-12-06 17:36:55.525000] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 7, Size 0x1000, Offset 0x0, Flags 0x3, Cap offset 0 00:21:13.895 [2024-12-06 17:36:55.526021] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 8, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:21:13.895 [2024-12-06 17:36:55.527035] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 9, Size 0xc000, Offset 0x0, Flags 0xf, Cap offset 32 00:21:13.895 [2024-12-06 17:36:55.527058] vfio_user_pci.c: 233:vfio_device_setup_sparse_mmaps: *DEBUG*: Sparse region 0, Size 0xb000, Offset 0x1000, Map addr 0x7f203d3c3000 00:21:13.895 [2024-12-06 17:36:55.528242] vfio_user_pci.c: 65:vfio_add_mr: *DEBUG*: Add memory region: FD 10, VADDR 0x200000200000, IOVA 0x200000200000, Size 0x200000 00:21:13.895 [2024-12-06 17:36:55.543014] vfio_user_pci.c: 386:spdk_vfio_user_setup: *DEBUG*: Device vfio-user0, Path /var/run/vfio-user/domain/vfio-user2/2/cntrl Setup Successfully 00:21:13.895 [2024-12-06 17:36:55.543066] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to connect adminq (no timeout) 00:21:13.895 [2024-12-06 17:36:55.548190] nvme_vfio_user.c: 103:nvme_vfio_ctrlr_get_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x0, value 0x201e0100ff 00:21:13.895 [2024-12-06 17:36:55.548240] nvme_pcie_common.c: 159:nvme_pcie_qpair_construct: *INFO*: max_completions_cap = 64 num_trackers = 192 00:21:13.895 [2024-12-06 17:36:55.548323] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to wait for connect adminq (no timeout) 00:21:13.895 [2024-12-06 17:36:55.548343] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to read vs (no timeout) 00:21:13.895 [2024-12-06 17:36:55.548353] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to read vs wait for vs (no timeout) 00:21:13.895 [2024-12-06 17:36:55.549193] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x8, value 0x10300 00:21:13.895 [2024-12-06 17:36:55.549213] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to read cap (no timeout) 00:21:13.895 [2024-12-06 17:36:55.549225] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to read cap wait for cap (no timeout) 00:21:13.895 [2024-12-06 17:36:55.550196] nvme_vfio_user.c: 103:nvme_vfio_ctrlr_get_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x0, value 0x201e0100ff 00:21:13.895 [2024-12-06 17:36:55.550215] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to check en (no timeout) 00:21:13.895 [2024-12-06 17:36:55.550228] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to check en wait for cc (timeout 15000 ms) 00:21:13.895 [2024-12-06 17:36:55.551204] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x14, value 0x0 00:21:13.895 [2024-12-06 17:36:55.551224] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to disable and wait for CSTS.RDY = 0 (timeout 15000 ms) 00:21:13.895 [2024-12-06 17:36:55.552208] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x1c, value 0x0 00:21:13.895 [2024-12-06 17:36:55.552227] nvme_ctrlr.c:3906:nvme_ctrlr_process_init_wait_for_ready_0: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] CC.EN = 0 && CSTS.RDY = 0 00:21:13.895 [2024-12-06 17:36:55.552236] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to controller is disabled (timeout 15000 ms) 00:21:13.895 [2024-12-06 17:36:55.552248] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to enable controller by writing CC.EN = 1 (timeout 15000 ms) 00:21:13.895 [2024-12-06 17:36:55.552356] nvme_ctrlr.c:4104:nvme_ctrlr_process_init: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] Setting CC.EN = 1 00:21:13.895 [2024-12-06 17:36:55.552364] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to enable controller by writing CC.EN = 1 reg (timeout 15000 ms) 00:21:13.895 [2024-12-06 17:36:55.552372] nvme_vfio_user.c: 61:nvme_vfio_ctrlr_set_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x28, value 0x2000003c0000 00:21:13.895 [2024-12-06 17:36:55.553211] nvme_vfio_user.c: 61:nvme_vfio_ctrlr_set_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x30, value 0x2000003be000 00:21:13.895 [2024-12-06 17:36:55.554214] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x24, value 0xff00ff 00:21:13.895 [2024-12-06 17:36:55.555225] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x14, value 0x460001 00:21:13.895 [2024-12-06 17:36:55.556220] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:21:13.895 [2024-12-06 17:36:55.556301] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to wait for CSTS.RDY = 1 (timeout 15000 ms) 00:21:13.895 [2024-12-06 17:36:55.557239] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x1c, value 0x1 00:21:13.895 [2024-12-06 17:36:55.557258] nvme_ctrlr.c:3941:nvme_ctrlr_process_init_enable_wait_for_ready_1: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] CC.EN = 1 && CSTS.RDY = 1 - controller is ready 00:21:13.895 [2024-12-06 17:36:55.557267] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to reset admin queue (timeout 30000 ms) 00:21:13.895 [2024-12-06 17:36:55.557290] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to identify controller (no timeout) 00:21:13.895 [2024-12-06 17:36:55.557307] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to wait for identify controller (timeout 30000 ms) 00:21:13.895 [2024-12-06 17:36:55.557329] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:21:13.895 [2024-12-06 17:36:55.557339] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:21:13.895 [2024-12-06 17:36:55.557345] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:21:13.895 [2024-12-06 17:36:55.557361] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:0 cdw10:00000001 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:21:13.895 [2024-12-06 17:36:55.564680] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0001 p:1 m:0 dnr:0 00:21:13.895 [2024-12-06 17:36:55.564702] nvme_ctrlr.c:2081:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] transport max_xfer_size 131072 00:21:13.895 [2024-12-06 17:36:55.564715] nvme_ctrlr.c:2085:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] MDTS max_xfer_size 131072 00:21:13.895 [2024-12-06 17:36:55.564723] nvme_ctrlr.c:2088:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] CNTLID 0x0001 00:21:13.895 [2024-12-06 17:36:55.564731] nvme_ctrlr.c:2099:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] Identify CNTLID 0x0001 != Connect CNTLID 0x0000 00:21:13.895 [2024-12-06 17:36:55.564739] nvme_ctrlr.c:2112:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] transport max_sges 1 00:21:13.895 [2024-12-06 17:36:55.564746] nvme_ctrlr.c:2127:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] fuses compare and write: 1 00:21:13.895 [2024-12-06 17:36:55.564754] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to configure AER (timeout 30000 ms) 00:21:13.895 [2024-12-06 17:36:55.564766] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to wait for configure aer (timeout 30000 ms) 00:21:13.895 [2024-12-06 17:36:55.564782] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES ASYNC EVENT CONFIGURATION cid:191 cdw10:0000000b PRP1 0x0 PRP2 0x0 00:21:13.895 [2024-12-06 17:36:55.572682] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0002 p:1 m:0 dnr:0 00:21:13.895 [2024-12-06 17:36:55.572706] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:190 nsid:0 cdw10:00000000 cdw11:00000000 00:21:13.895 [2024-12-06 17:36:55.572719] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:189 nsid:0 cdw10:00000000 cdw11:00000000 00:21:13.895 [2024-12-06 17:36:55.572735] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:188 nsid:0 cdw10:00000000 cdw11:00000000 00:21:13.895 [2024-12-06 17:36:55.572748] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:187 nsid:0 cdw10:00000000 cdw11:00000000 00:21:13.895 [2024-12-06 17:36:55.572757] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to set keep alive timeout (timeout 30000 ms) 00:21:13.895 [2024-12-06 17:36:55.572772] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to wait for set keep alive timeout (timeout 30000 ms) 00:21:13.895 [2024-12-06 17:36:55.572788] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES KEEP ALIVE TIMER cid:191 cdw10:0000000f PRP1 0x0 PRP2 0x0 00:21:13.895 [2024-12-06 17:36:55.580674] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0007 p:1 m:0 dnr:0 00:21:13.895 [2024-12-06 17:36:55.580692] nvme_ctrlr.c:3047:nvme_ctrlr_set_keep_alive_timeout_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] Controller adjusted keep alive timeout to 0 ms 00:21:13.895 [2024-12-06 17:36:55.580701] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to identify controller iocs specific (timeout 30000 ms) 00:21:13.895 [2024-12-06 17:36:55.580712] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to set number of queues (timeout 30000 ms) 00:21:13.895 [2024-12-06 17:36:55.580722] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to wait for set number of queues (timeout 30000 ms) 00:21:13.895 [2024-12-06 17:36:55.580736] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES NUMBER OF QUEUES cid:191 cdw10:00000007 PRP1 0x0 PRP2 0x0 00:21:13.895 [2024-12-06 17:36:55.588676] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:7e007e sqhd:0008 p:1 m:0 dnr:0 00:21:13.895 [2024-12-06 17:36:55.588766] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to identify active ns (timeout 30000 ms) 00:21:13.895 [2024-12-06 17:36:55.588782] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to wait for identify active ns (timeout 30000 ms) 00:21:13.895 [2024-12-06 17:36:55.588795] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f9000 len:4096 00:21:13.895 [2024-12-06 17:36:55.588804] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f9000 00:21:13.895 [2024-12-06 17:36:55.588810] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:21:13.895 [2024-12-06 17:36:55.588819] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:0 cdw10:00000002 cdw11:00000000 PRP1 0x2000002f9000 PRP2 0x0 00:21:13.895 [2024-12-06 17:36:55.596675] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0009 p:1 m:0 dnr:0 00:21:13.895 [2024-12-06 17:36:55.596697] nvme_ctrlr.c:4735:spdk_nvme_ctrlr_get_ns: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] Namespace 1 was added 00:21:13.895 [2024-12-06 17:36:55.596716] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to identify ns (timeout 30000 ms) 00:21:13.895 [2024-12-06 17:36:55.596731] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to wait for identify ns (timeout 30000 ms) 00:21:13.895 [2024-12-06 17:36:55.596743] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:21:13.895 [2024-12-06 17:36:55.596751] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:21:13.896 [2024-12-06 17:36:55.596757] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:21:13.896 [2024-12-06 17:36:55.596772] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:1 cdw10:00000000 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:21:13.896 [2024-12-06 17:36:55.604676] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000a p:1 m:0 dnr:0 00:21:13.896 [2024-12-06 17:36:55.604703] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to identify namespace id descriptors (timeout 30000 ms) 00:21:13.896 [2024-12-06 17:36:55.604720] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to wait for identify namespace id descriptors (timeout 30000 ms) 00:21:13.896 [2024-12-06 17:36:55.604732] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:21:13.896 [2024-12-06 17:36:55.604741] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:21:13.896 [2024-12-06 17:36:55.604746] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:21:13.896 [2024-12-06 17:36:55.604756] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:1 cdw10:00000003 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:21:13.896 [2024-12-06 17:36:55.612676] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000b p:1 m:0 dnr:0 00:21:13.896 [2024-12-06 17:36:55.612697] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to identify ns iocs specific (timeout 30000 ms) 00:21:13.896 [2024-12-06 17:36:55.612709] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to set supported log pages (timeout 30000 ms) 00:21:13.896 [2024-12-06 17:36:55.612723] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to set supported features (timeout 30000 ms) 00:21:13.896 [2024-12-06 17:36:55.612736] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to set host behavior support feature (timeout 30000 ms) 00:21:13.896 [2024-12-06 17:36:55.612745] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to set doorbell buffer config (timeout 30000 ms) 00:21:13.896 [2024-12-06 17:36:55.612753] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to set host ID (timeout 30000 ms) 00:21:13.896 [2024-12-06 17:36:55.612761] nvme_ctrlr.c:3147:nvme_ctrlr_set_host_id: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] NVMe-oF transport - not sending Set Features - Host ID 00:21:13.896 [2024-12-06 17:36:55.612769] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to transport ready (timeout 30000 ms) 00:21:13.896 [2024-12-06 17:36:55.612777] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to ready (no timeout) 00:21:13.896 [2024-12-06 17:36:55.612801] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ARBITRATION cid:191 cdw10:00000001 PRP1 0x0 PRP2 0x0 00:21:13.896 [2024-12-06 17:36:55.620679] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000c p:1 m:0 dnr:0 00:21:13.896 [2024-12-06 17:36:55.620705] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES POWER MANAGEMENT cid:191 cdw10:00000002 PRP1 0x0 PRP2 0x0 00:21:13.896 [2024-12-06 17:36:55.628675] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000d p:1 m:0 dnr:0 00:21:13.896 [2024-12-06 17:36:55.628701] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES TEMPERATURE THRESHOLD cid:191 cdw10:00000004 PRP1 0x0 PRP2 0x0 00:21:13.896 [2024-12-06 17:36:55.636692] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000e p:1 m:0 dnr:0 00:21:13.896 [2024-12-06 17:36:55.636717] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES NUMBER OF QUEUES cid:191 cdw10:00000007 PRP1 0x0 PRP2 0x0 00:21:13.896 [2024-12-06 17:36:55.644690] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:7e007e sqhd:000f p:1 m:0 dnr:0 00:21:13.896 [2024-12-06 17:36:55.644721] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f6000 len:8192 00:21:13.896 [2024-12-06 17:36:55.644732] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f6000 00:21:13.896 [2024-12-06 17:36:55.644738] nvme_pcie_common.c:1275:nvme_pcie_prp_list_append: *DEBUG*: prp[0] = 0x2000002f7000 00:21:13.896 [2024-12-06 17:36:55.644744] nvme_pcie_common.c:1291:nvme_pcie_prp_list_append: *DEBUG*: prp2 = 0x2000002f7000 00:21:13.896 [2024-12-06 17:36:55.644750] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 2 00:21:13.896 [2024-12-06 17:36:55.644759] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:191 nsid:ffffffff cdw10:07ff0001 cdw11:00000000 PRP1 0x2000002f6000 PRP2 0x2000002f7000 00:21:13.896 [2024-12-06 17:36:55.644771] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fc000 len:512 00:21:13.896 [2024-12-06 17:36:55.644779] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fc000 00:21:13.896 [2024-12-06 17:36:55.644785] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:21:13.896 [2024-12-06 17:36:55.644794] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:186 nsid:ffffffff cdw10:007f0002 cdw11:00000000 PRP1 0x2000002fc000 PRP2 0x0 00:21:13.896 [2024-12-06 17:36:55.644805] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:512 00:21:13.896 [2024-12-06 17:36:55.644812] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:21:13.896 [2024-12-06 17:36:55.644818] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:21:13.896 [2024-12-06 17:36:55.644827] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:185 nsid:ffffffff cdw10:007f0003 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:21:13.896 [2024-12-06 17:36:55.644838] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f4000 len:4096 00:21:13.896 [2024-12-06 17:36:55.644846] nvme_pcie_common.c:1266:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f4000 00:21:13.896 [2024-12-06 17:36:55.644852] nvme_pcie_common.c:1326:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:21:13.896 [2024-12-06 17:36:55.644860] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:184 nsid:ffffffff cdw10:03ff0005 cdw11:00000000 PRP1 0x2000002f4000 PRP2 0x0 00:21:13.896 [2024-12-06 17:36:55.652694] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0010 p:1 m:0 dnr:0 00:21:13.896 [2024-12-06 17:36:55.652722] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:186 cdw0:0 sqhd:0011 p:1 m:0 dnr:0 00:21:13.896 [2024-12-06 17:36:55.652740] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:185 cdw0:0 sqhd:0012 p:1 m:0 dnr:0 00:21:13.896 [2024-12-06 17:36:55.652753] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:184 cdw0:0 sqhd:0013 p:1 m:0 dnr:0 00:21:13.896 ===================================================== 00:21:13.896 NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:21:13.896 ===================================================== 00:21:13.896 Controller Capabilities/Features 00:21:13.896 ================================ 00:21:13.896 Vendor ID: 4e58 00:21:13.896 Subsystem Vendor ID: 4e58 00:21:13.896 Serial Number: SPDK2 00:21:13.896 Model Number: SPDK bdev Controller 00:21:13.896 Firmware Version: 25.01 00:21:13.896 Recommended Arb Burst: 6 00:21:13.896 IEEE OUI Identifier: 8d 6b 50 00:21:13.896 Multi-path I/O 00:21:13.896 May have multiple subsystem ports: Yes 00:21:13.896 May have multiple controllers: Yes 00:21:13.896 Associated with SR-IOV VF: No 00:21:13.896 Max Data Transfer Size: 131072 00:21:13.896 Max Number of Namespaces: 32 00:21:13.896 Max Number of I/O Queues: 127 00:21:13.896 NVMe Specification Version (VS): 1.3 00:21:13.896 NVMe Specification Version (Identify): 1.3 00:21:13.896 Maximum Queue Entries: 256 00:21:13.896 Contiguous Queues Required: Yes 00:21:13.896 Arbitration Mechanisms Supported 00:21:13.896 Weighted Round Robin: Not Supported 00:21:13.896 Vendor Specific: Not Supported 00:21:13.896 Reset Timeout: 15000 ms 00:21:13.896 Doorbell Stride: 4 bytes 00:21:13.896 NVM Subsystem Reset: Not Supported 00:21:13.896 Command Sets Supported 00:21:13.896 NVM Command Set: Supported 00:21:13.896 Boot Partition: Not Supported 00:21:13.896 Memory Page Size Minimum: 4096 bytes 00:21:13.896 Memory Page Size Maximum: 4096 bytes 00:21:13.896 Persistent Memory Region: Not Supported 00:21:13.896 Optional Asynchronous Events Supported 00:21:13.896 Namespace Attribute Notices: Supported 00:21:13.896 Firmware Activation Notices: Not Supported 00:21:13.896 ANA Change Notices: Not Supported 00:21:13.896 PLE Aggregate Log Change Notices: Not Supported 00:21:13.896 LBA Status Info Alert Notices: Not Supported 00:21:13.896 EGE Aggregate Log Change Notices: Not Supported 00:21:13.896 Normal NVM Subsystem Shutdown event: Not Supported 00:21:13.896 Zone Descriptor Change Notices: Not Supported 00:21:13.896 Discovery Log Change Notices: Not Supported 00:21:13.896 Controller Attributes 00:21:13.896 128-bit Host Identifier: Supported 00:21:13.896 Non-Operational Permissive Mode: Not Supported 00:21:13.896 NVM Sets: Not Supported 00:21:13.896 Read Recovery Levels: Not Supported 00:21:13.896 Endurance Groups: Not Supported 00:21:13.896 Predictable Latency Mode: Not Supported 00:21:13.896 Traffic Based Keep ALive: Not Supported 00:21:13.896 Namespace Granularity: Not Supported 00:21:13.896 SQ Associations: Not Supported 00:21:13.896 UUID List: Not Supported 00:21:13.896 Multi-Domain Subsystem: Not Supported 00:21:13.896 Fixed Capacity Management: Not Supported 00:21:13.896 Variable Capacity Management: Not Supported 00:21:13.896 Delete Endurance Group: Not Supported 00:21:13.896 Delete NVM Set: Not Supported 00:21:13.896 Extended LBA Formats Supported: Not Supported 00:21:13.896 Flexible Data Placement Supported: Not Supported 00:21:13.896 00:21:13.896 Controller Memory Buffer Support 00:21:13.896 ================================ 00:21:13.896 Supported: No 00:21:13.896 00:21:13.896 Persistent Memory Region Support 00:21:13.896 ================================ 00:21:13.896 Supported: No 00:21:13.896 00:21:13.896 Admin Command Set Attributes 00:21:13.896 ============================ 00:21:13.896 Security Send/Receive: Not Supported 00:21:13.896 Format NVM: Not Supported 00:21:13.896 Firmware Activate/Download: Not Supported 00:21:13.896 Namespace Management: Not Supported 00:21:13.896 Device Self-Test: Not Supported 00:21:13.897 Directives: Not Supported 00:21:13.897 NVMe-MI: Not Supported 00:21:13.897 Virtualization Management: Not Supported 00:21:13.897 Doorbell Buffer Config: Not Supported 00:21:13.897 Get LBA Status Capability: Not Supported 00:21:13.897 Command & Feature Lockdown Capability: Not Supported 00:21:13.897 Abort Command Limit: 4 00:21:13.897 Async Event Request Limit: 4 00:21:13.897 Number of Firmware Slots: N/A 00:21:13.897 Firmware Slot 1 Read-Only: N/A 00:21:13.897 Firmware Activation Without Reset: N/A 00:21:13.897 Multiple Update Detection Support: N/A 00:21:13.897 Firmware Update Granularity: No Information Provided 00:21:13.897 Per-Namespace SMART Log: No 00:21:13.897 Asymmetric Namespace Access Log Page: Not Supported 00:21:13.897 Subsystem NQN: nqn.2019-07.io.spdk:cnode2 00:21:13.897 Command Effects Log Page: Supported 00:21:13.897 Get Log Page Extended Data: Supported 00:21:13.897 Telemetry Log Pages: Not Supported 00:21:13.897 Persistent Event Log Pages: Not Supported 00:21:13.897 Supported Log Pages Log Page: May Support 00:21:13.897 Commands Supported & Effects Log Page: Not Supported 00:21:13.897 Feature Identifiers & Effects Log Page:May Support 00:21:13.897 NVMe-MI Commands & Effects Log Page: May Support 00:21:13.897 Data Area 4 for Telemetry Log: Not Supported 00:21:13.897 Error Log Page Entries Supported: 128 00:21:13.897 Keep Alive: Supported 00:21:13.897 Keep Alive Granularity: 10000 ms 00:21:13.897 00:21:13.897 NVM Command Set Attributes 00:21:13.897 ========================== 00:21:13.897 Submission Queue Entry Size 00:21:13.897 Max: 64 00:21:13.897 Min: 64 00:21:13.897 Completion Queue Entry Size 00:21:13.897 Max: 16 00:21:13.897 Min: 16 00:21:13.897 Number of Namespaces: 32 00:21:13.897 Compare Command: Supported 00:21:13.897 Write Uncorrectable Command: Not Supported 00:21:13.897 Dataset Management Command: Supported 00:21:13.897 Write Zeroes Command: Supported 00:21:13.897 Set Features Save Field: Not Supported 00:21:13.897 Reservations: Not Supported 00:21:13.897 Timestamp: Not Supported 00:21:13.897 Copy: Supported 00:21:13.897 Volatile Write Cache: Present 00:21:13.897 Atomic Write Unit (Normal): 1 00:21:13.897 Atomic Write Unit (PFail): 1 00:21:13.897 Atomic Compare & Write Unit: 1 00:21:13.897 Fused Compare & Write: Supported 00:21:13.897 Scatter-Gather List 00:21:13.897 SGL Command Set: Supported (Dword aligned) 00:21:13.897 SGL Keyed: Not Supported 00:21:13.897 SGL Bit Bucket Descriptor: Not Supported 00:21:13.897 SGL Metadata Pointer: Not Supported 00:21:13.897 Oversized SGL: Not Supported 00:21:13.897 SGL Metadata Address: Not Supported 00:21:13.897 SGL Offset: Not Supported 00:21:13.897 Transport SGL Data Block: Not Supported 00:21:13.897 Replay Protected Memory Block: Not Supported 00:21:13.897 00:21:13.897 Firmware Slot Information 00:21:13.897 ========================= 00:21:13.897 Active slot: 1 00:21:13.897 Slot 1 Firmware Revision: 25.01 00:21:13.897 00:21:13.897 00:21:13.897 Commands Supported and Effects 00:21:13.897 ============================== 00:21:13.897 Admin Commands 00:21:13.897 -------------- 00:21:13.897 Get Log Page (02h): Supported 00:21:13.897 Identify (06h): Supported 00:21:13.897 Abort (08h): Supported 00:21:13.897 Set Features (09h): Supported 00:21:13.897 Get Features (0Ah): Supported 00:21:13.897 Asynchronous Event Request (0Ch): Supported 00:21:13.897 Keep Alive (18h): Supported 00:21:13.897 I/O Commands 00:21:13.897 ------------ 00:21:13.897 Flush (00h): Supported LBA-Change 00:21:13.897 Write (01h): Supported LBA-Change 00:21:13.897 Read (02h): Supported 00:21:13.897 Compare (05h): Supported 00:21:13.897 Write Zeroes (08h): Supported LBA-Change 00:21:13.897 Dataset Management (09h): Supported LBA-Change 00:21:13.897 Copy (19h): Supported LBA-Change 00:21:13.897 00:21:13.897 Error Log 00:21:13.897 ========= 00:21:13.897 00:21:13.897 Arbitration 00:21:13.897 =========== 00:21:13.897 Arbitration Burst: 1 00:21:13.897 00:21:13.897 Power Management 00:21:13.897 ================ 00:21:13.897 Number of Power States: 1 00:21:13.897 Current Power State: Power State #0 00:21:13.897 Power State #0: 00:21:13.897 Max Power: 0.00 W 00:21:13.897 Non-Operational State: Operational 00:21:13.897 Entry Latency: Not Reported 00:21:13.897 Exit Latency: Not Reported 00:21:13.897 Relative Read Throughput: 0 00:21:13.897 Relative Read Latency: 0 00:21:13.897 Relative Write Throughput: 0 00:21:13.897 Relative Write Latency: 0 00:21:13.897 Idle Power: Not Reported 00:21:13.897 Active Power: Not Reported 00:21:13.897 Non-Operational Permissive Mode: Not Supported 00:21:13.897 00:21:13.897 Health Information 00:21:13.897 ================== 00:21:13.897 Critical Warnings: 00:21:13.897 Available Spare Space: OK 00:21:13.897 Temperature: OK 00:21:13.897 Device Reliability: OK 00:21:13.897 Read Only: No 00:21:13.897 Volatile Memory Backup: OK 00:21:13.897 Current Temperature: 0 Kelvin (-273 Celsius) 00:21:13.897 Temperature Threshold: 0 Kelvin (-273 Celsius) 00:21:13.897 Available Spare: 0% 00:21:13.897 Available Sp[2024-12-06 17:36:55.652869] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ERROR_RECOVERY cid:184 cdw10:00000005 PRP1 0x0 PRP2 0x0 00:21:13.897 [2024-12-06 17:36:55.660680] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:184 cdw0:0 sqhd:0014 p:1 m:0 dnr:0 00:21:13.897 [2024-12-06 17:36:55.660730] nvme_ctrlr.c:4399:nvme_ctrlr_destruct_async: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] Prepare to destruct SSD 00:21:13.897 [2024-12-06 17:36:55.660748] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:190 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:13.897 [2024-12-06 17:36:55.660759] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:189 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:13.897 [2024-12-06 17:36:55.660768] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:188 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:13.897 [2024-12-06 17:36:55.660781] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:187 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:13.897 [2024-12-06 17:36:55.660864] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x14, value 0x460001 00:21:13.897 [2024-12-06 17:36:55.660884] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x14, value 0x464001 00:21:13.897 [2024-12-06 17:36:55.661869] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:21:13.897 [2024-12-06 17:36:55.661940] nvme_ctrlr.c:1151:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] RTD3E = 0 us 00:21:13.897 [2024-12-06 17:36:55.661955] nvme_ctrlr.c:1154:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] shutdown timeout = 10000 ms 00:21:13.897 [2024-12-06 17:36:55.662877] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x1c, value 0x9 00:21:13.897 [2024-12-06 17:36:55.662900] nvme_ctrlr.c:1273:nvme_ctrlr_shutdown_poll_async: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] shutdown complete in 0 milliseconds 00:21:13.897 [2024-12-06 17:36:55.662966] vfio_user_pci.c: 399:spdk_vfio_user_release: *DEBUG*: Release file /var/run/vfio-user/domain/vfio-user2/2/cntrl 00:21:13.897 [2024-12-06 17:36:55.664140] vfio_user_pci.c: 96:vfio_remove_mr: *DEBUG*: Remove memory region: FD 10, VADDR 0x200000200000, IOVA 0x200000200000, Size 0x200000 00:21:13.897 are Threshold: 0% 00:21:13.897 Life Percentage Used: 0% 00:21:13.897 Data Units Read: 0 00:21:13.897 Data Units Written: 0 00:21:13.897 Host Read Commands: 0 00:21:13.897 Host Write Commands: 0 00:21:13.897 Controller Busy Time: 0 minutes 00:21:13.897 Power Cycles: 0 00:21:13.897 Power On Hours: 0 hours 00:21:13.897 Unsafe Shutdowns: 0 00:21:13.897 Unrecoverable Media Errors: 0 00:21:13.897 Lifetime Error Log Entries: 0 00:21:13.897 Warning Temperature Time: 0 minutes 00:21:13.897 Critical Temperature Time: 0 minutes 00:21:13.897 00:21:13.897 Number of Queues 00:21:13.897 ================ 00:21:13.897 Number of I/O Submission Queues: 127 00:21:13.897 Number of I/O Completion Queues: 127 00:21:13.897 00:21:13.897 Active Namespaces 00:21:13.897 ================= 00:21:13.897 Namespace ID:1 00:21:13.897 Error Recovery Timeout: Unlimited 00:21:13.897 Command Set Identifier: NVM (00h) 00:21:13.897 Deallocate: Supported 00:21:13.897 Deallocated/Unwritten Error: Not Supported 00:21:13.897 Deallocated Read Value: Unknown 00:21:13.897 Deallocate in Write Zeroes: Not Supported 00:21:13.897 Deallocated Guard Field: 0xFFFF 00:21:13.897 Flush: Supported 00:21:13.897 Reservation: Supported 00:21:13.897 Namespace Sharing Capabilities: Multiple Controllers 00:21:13.897 Size (in LBAs): 131072 (0GiB) 00:21:13.897 Capacity (in LBAs): 131072 (0GiB) 00:21:13.897 Utilization (in LBAs): 131072 (0GiB) 00:21:13.897 NGUID: 064BC5F836974A08AA1E011B1094E1E2 00:21:13.897 UUID: 064bc5f8-3697-4a08-aa1e-011b1094e1e2 00:21:13.897 Thin Provisioning: Not Supported 00:21:13.897 Per-NS Atomic Units: Yes 00:21:13.897 Atomic Boundary Size (Normal): 0 00:21:13.897 Atomic Boundary Size (PFail): 0 00:21:13.897 Atomic Boundary Offset: 0 00:21:13.897 Maximum Single Source Range Length: 65535 00:21:13.897 Maximum Copy Length: 65535 00:21:13.898 Maximum Source Range Count: 1 00:21:13.898 NGUID/EUI64 Never Reused: No 00:21:13.898 Namespace Write Protected: No 00:21:13.898 Number of LBA Formats: 1 00:21:13.898 Current LBA Format: LBA Format #00 00:21:13.898 LBA Format #00: Data Size: 512 Metadata Size: 0 00:21:13.898 00:21:13.898 17:36:55 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -s 256 -g -q 128 -o 4096 -w read -t 5 -c 0x2 00:21:14.155 [2024-12-06 17:36:55.909457] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:21:19.421 Initializing NVMe Controllers 00:21:19.421 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:21:19.421 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) NSID 1 with lcore 1 00:21:19.421 Initialization complete. Launching workers. 00:21:19.421 ======================================================== 00:21:19.421 Latency(us) 00:21:19.421 Device Information : IOPS MiB/s Average min max 00:21:19.421 VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) NSID 1 from core 1: 31341.38 122.43 4083.42 1219.29 10383.51 00:21:19.421 ======================================================== 00:21:19.421 Total : 31341.38 122.43 4083.42 1219.29 10383.51 00:21:19.421 00:21:19.421 [2024-12-06 17:37:01.015063] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:21:19.421 17:37:01 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@85 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -s 256 -g -q 128 -o 4096 -w write -t 5 -c 0x2 00:21:19.677 [2024-12-06 17:37:01.266727] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:21:24.934 Initializing NVMe Controllers 00:21:24.934 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:21:24.934 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) NSID 1 with lcore 1 00:21:24.934 Initialization complete. Launching workers. 00:21:24.934 ======================================================== 00:21:24.934 Latency(us) 00:21:24.934 Device Information : IOPS MiB/s Average min max 00:21:24.934 VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) NSID 1 from core 1: 29623.06 115.72 4320.31 1252.71 7824.29 00:21:24.934 ======================================================== 00:21:24.934 Total : 29623.06 115.72 4320.31 1252.71 7824.29 00:21:24.934 00:21:24.934 [2024-12-06 17:37:06.292087] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:21:24.934 17:37:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -g -q 32 -o 4096 -w randrw -M 50 -t 5 -c 0xE 00:21:24.934 [2024-12-06 17:37:06.511381] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:21:30.205 [2024-12-06 17:37:11.633830] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:21:30.205 Initializing NVMe Controllers 00:21:30.205 Attaching to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:21:30.205 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:21:30.205 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) with lcore 1 00:21:30.205 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) with lcore 2 00:21:30.205 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) with lcore 3 00:21:30.205 Initialization complete. Launching workers. 00:21:30.205 Starting thread on core 2 00:21:30.205 Starting thread on core 3 00:21:30.205 Starting thread on core 1 00:21:30.205 17:37:11 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@87 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration -t 3 -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -d 256 -g 00:21:30.205 [2024-12-06 17:37:11.958235] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:21:33.490 [2024-12-06 17:37:15.032983] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:21:33.490 Initializing NVMe Controllers 00:21:33.490 Attaching to /var/run/vfio-user/domain/vfio-user2/2 00:21:33.490 Attached to /var/run/vfio-user/domain/vfio-user2/2 00:21:33.490 Associating SPDK bdev Controller (SPDK2 ) with lcore 0 00:21:33.490 Associating SPDK bdev Controller (SPDK2 ) with lcore 1 00:21:33.490 Associating SPDK bdev Controller (SPDK2 ) with lcore 2 00:21:33.490 Associating SPDK bdev Controller (SPDK2 ) with lcore 3 00:21:33.490 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration run with configuration: 00:21:33.490 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration -q 64 -s 131072 -w randrw -M 50 -l 0 -t 3 -c 0xf -m 0 -a 0 -b 0 -n 100000 -i -1 00:21:33.490 Initialization complete. Launching workers. 00:21:33.490 Starting thread on core 1 with urgent priority queue 00:21:33.490 Starting thread on core 2 with urgent priority queue 00:21:33.490 Starting thread on core 3 with urgent priority queue 00:21:33.490 Starting thread on core 0 with urgent priority queue 00:21:33.490 SPDK bdev Controller (SPDK2 ) core 0: 4978.67 IO/s 20.09 secs/100000 ios 00:21:33.490 SPDK bdev Controller (SPDK2 ) core 1: 5465.33 IO/s 18.30 secs/100000 ios 00:21:33.490 SPDK bdev Controller (SPDK2 ) core 2: 5262.67 IO/s 19.00 secs/100000 ios 00:21:33.490 SPDK bdev Controller (SPDK2 ) core 3: 5715.67 IO/s 17.50 secs/100000 ios 00:21:33.490 ======================================================== 00:21:33.490 00:21:33.490 17:37:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/hello_world -d 256 -g -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' 00:21:33.748 [2024-12-06 17:37:15.346310] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:21:33.748 Initializing NVMe Controllers 00:21:33.748 Attaching to /var/run/vfio-user/domain/vfio-user2/2 00:21:33.748 Attached to /var/run/vfio-user/domain/vfio-user2/2 00:21:33.748 Namespace ID: 1 size: 0GB 00:21:33.748 Initialization complete. 00:21:33.748 INFO: using host memory buffer for IO 00:21:33.748 Hello world! 00:21:33.748 [2024-12-06 17:37:15.356376] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:21:33.748 17:37:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/overhead/overhead -o 4096 -t 1 -H -g -d 256 -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' 00:21:34.006 [2024-12-06 17:37:15.671990] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:21:34.941 Initializing NVMe Controllers 00:21:34.941 Attaching to /var/run/vfio-user/domain/vfio-user2/2 00:21:34.941 Attached to /var/run/vfio-user/domain/vfio-user2/2 00:21:34.941 Initialization complete. Launching workers. 00:21:34.941 submit (in ns) avg, min, max = 7444.3, 3501.1, 4015582.2 00:21:34.941 complete (in ns) avg, min, max = 27123.4, 2046.7, 4018937.8 00:21:34.941 00:21:34.941 Submit histogram 00:21:34.941 ================ 00:21:34.941 Range in us Cumulative Count 00:21:34.941 3.484 - 3.508: 0.0868% ( 11) 00:21:34.941 3.508 - 3.532: 0.5761% ( 62) 00:21:34.941 3.532 - 3.556: 2.0835% ( 191) 00:21:34.941 3.556 - 3.579: 4.9404% ( 362) 00:21:34.941 3.579 - 3.603: 11.2777% ( 803) 00:21:34.941 3.603 - 3.627: 20.0221% ( 1108) 00:21:34.941 3.627 - 3.650: 29.1295% ( 1154) 00:21:34.941 3.650 - 3.674: 38.1817% ( 1147) 00:21:34.941 3.674 - 3.698: 45.4345% ( 919) 00:21:34.941 3.698 - 3.721: 53.0818% ( 969) 00:21:34.941 3.721 - 3.745: 58.2353% ( 653) 00:21:34.941 3.745 - 3.769: 62.9074% ( 592) 00:21:34.941 3.769 - 3.793: 66.7982% ( 493) 00:21:34.941 3.793 - 3.816: 70.7205% ( 497) 00:21:34.941 3.816 - 3.840: 73.6406% ( 370) 00:21:34.941 3.840 - 3.864: 77.2394% ( 456) 00:21:34.941 3.864 - 3.887: 80.6645% ( 434) 00:21:34.941 3.887 - 3.911: 83.7898% ( 396) 00:21:34.941 3.911 - 3.935: 86.3862% ( 329) 00:21:34.941 3.935 - 3.959: 88.1541% ( 224) 00:21:34.941 3.959 - 3.982: 89.8114% ( 210) 00:21:34.941 3.982 - 4.006: 91.4450% ( 207) 00:21:34.941 4.006 - 4.030: 92.6604% ( 154) 00:21:34.941 4.030 - 4.053: 93.6232% ( 122) 00:21:34.941 4.053 - 4.077: 94.4045% ( 99) 00:21:34.941 4.077 - 4.101: 95.1148% ( 90) 00:21:34.941 4.101 - 4.124: 95.5489% ( 55) 00:21:34.941 4.124 - 4.148: 95.7935% ( 31) 00:21:34.941 4.148 - 4.172: 96.0066% ( 27) 00:21:34.941 4.172 - 4.196: 96.1487% ( 18) 00:21:34.941 4.196 - 4.219: 96.2355% ( 11) 00:21:34.941 4.219 - 4.243: 96.3302% ( 12) 00:21:34.941 4.243 - 4.267: 96.4091% ( 10) 00:21:34.941 4.267 - 4.290: 96.5117% ( 13) 00:21:34.941 4.290 - 4.314: 96.6143% ( 13) 00:21:34.941 4.314 - 4.338: 96.6932% ( 10) 00:21:34.941 4.338 - 4.361: 96.7879% ( 12) 00:21:34.941 4.361 - 4.385: 96.8274% ( 5) 00:21:34.941 4.385 - 4.409: 96.8590% ( 4) 00:21:34.941 4.409 - 4.433: 96.8826% ( 3) 00:21:34.941 4.433 - 4.456: 96.9063% ( 3) 00:21:34.941 4.456 - 4.480: 96.9221% ( 2) 00:21:34.941 4.480 - 4.504: 96.9379% ( 2) 00:21:34.941 4.504 - 4.527: 96.9458% ( 1) 00:21:34.941 4.527 - 4.551: 96.9537% ( 1) 00:21:34.941 4.646 - 4.670: 96.9616% ( 1) 00:21:34.941 4.670 - 4.693: 96.9695% ( 1) 00:21:34.941 4.693 - 4.717: 96.9931% ( 3) 00:21:34.941 4.717 - 4.741: 97.0247% ( 4) 00:21:34.941 4.741 - 4.764: 97.0563% ( 4) 00:21:34.941 4.764 - 4.788: 97.0799% ( 3) 00:21:34.941 4.788 - 4.812: 97.1115% ( 4) 00:21:34.941 4.812 - 4.836: 97.1431% ( 4) 00:21:34.941 4.836 - 4.859: 97.2141% ( 9) 00:21:34.941 4.859 - 4.883: 97.2851% ( 9) 00:21:34.941 4.883 - 4.907: 97.3641% ( 10) 00:21:34.941 4.907 - 4.930: 97.4114% ( 6) 00:21:34.941 4.930 - 4.954: 97.4588% ( 6) 00:21:34.941 4.954 - 4.978: 97.5061% ( 6) 00:21:34.941 4.978 - 5.001: 97.6087% ( 13) 00:21:34.941 5.001 - 5.025: 97.6718% ( 8) 00:21:34.941 5.025 - 5.049: 97.7034% ( 4) 00:21:34.941 5.049 - 5.073: 97.7192% ( 2) 00:21:34.941 5.073 - 5.096: 97.7666% ( 6) 00:21:34.941 5.096 - 5.120: 97.7823% ( 2) 00:21:34.941 5.120 - 5.144: 97.7902% ( 1) 00:21:34.941 5.144 - 5.167: 97.7981% ( 1) 00:21:34.941 5.167 - 5.191: 97.8060% ( 1) 00:21:34.941 5.191 - 5.215: 97.8139% ( 1) 00:21:34.941 5.215 - 5.239: 97.8297% ( 2) 00:21:34.941 5.239 - 5.262: 97.8455% ( 2) 00:21:34.941 5.262 - 5.286: 97.8534% ( 1) 00:21:34.941 5.310 - 5.333: 97.8613% ( 1) 00:21:34.941 5.381 - 5.404: 97.8692% ( 1) 00:21:34.941 5.641 - 5.665: 97.8770% ( 1) 00:21:34.941 5.713 - 5.736: 97.8849% ( 1) 00:21:34.941 5.736 - 5.760: 97.9007% ( 2) 00:21:34.941 5.760 - 5.784: 97.9086% ( 1) 00:21:34.941 5.807 - 5.831: 97.9165% ( 1) 00:21:34.941 5.950 - 5.973: 97.9323% ( 2) 00:21:34.941 6.021 - 6.044: 97.9481% ( 2) 00:21:34.941 6.068 - 6.116: 97.9639% ( 2) 00:21:34.941 6.116 - 6.163: 97.9717% ( 1) 00:21:34.941 6.210 - 6.258: 97.9796% ( 1) 00:21:34.941 6.305 - 6.353: 97.9954% ( 2) 00:21:34.941 6.353 - 6.400: 98.0033% ( 1) 00:21:34.941 6.400 - 6.447: 98.0112% ( 1) 00:21:34.941 6.827 - 6.874: 98.0191% ( 1) 00:21:34.941 6.921 - 6.969: 98.0270% ( 1) 00:21:34.941 6.969 - 7.016: 98.0349% ( 1) 00:21:34.941 7.016 - 7.064: 98.0428% ( 1) 00:21:34.941 7.111 - 7.159: 98.0507% ( 1) 00:21:34.941 7.206 - 7.253: 98.0586% ( 1) 00:21:34.941 7.253 - 7.301: 98.0665% ( 1) 00:21:34.941 7.301 - 7.348: 98.0743% ( 1) 00:21:34.941 7.396 - 7.443: 98.0822% ( 1) 00:21:34.941 7.443 - 7.490: 98.0901% ( 1) 00:21:34.941 7.490 - 7.538: 98.0980% ( 1) 00:21:34.941 7.538 - 7.585: 98.1138% ( 2) 00:21:34.941 7.633 - 7.680: 98.1217% ( 1) 00:21:34.941 7.775 - 7.822: 98.1296% ( 1) 00:21:34.941 7.822 - 7.870: 98.1375% ( 1) 00:21:34.941 7.917 - 7.964: 98.1533% ( 2) 00:21:34.941 8.012 - 8.059: 98.1612% ( 1) 00:21:34.941 8.059 - 8.107: 98.1690% ( 1) 00:21:34.941 8.107 - 8.154: 98.1848% ( 2) 00:21:34.941 8.154 - 8.201: 98.2006% ( 2) 00:21:34.941 8.201 - 8.249: 98.2164% ( 2) 00:21:34.941 8.249 - 8.296: 98.2243% ( 1) 00:21:34.941 8.344 - 8.391: 98.2322% ( 1) 00:21:34.941 8.391 - 8.439: 98.2401% ( 1) 00:21:34.941 8.439 - 8.486: 98.2559% ( 2) 00:21:34.941 8.486 - 8.533: 98.2795% ( 3) 00:21:34.941 8.581 - 8.628: 98.3032% ( 3) 00:21:34.941 8.628 - 8.676: 98.3111% ( 1) 00:21:34.941 8.723 - 8.770: 98.3269% ( 2) 00:21:34.941 9.007 - 9.055: 98.3427% ( 2) 00:21:34.941 9.055 - 9.102: 98.3506% ( 1) 00:21:34.941 9.102 - 9.150: 98.3585% ( 1) 00:21:34.941 9.150 - 9.197: 98.3900% ( 4) 00:21:34.941 9.292 - 9.339: 98.4058% ( 2) 00:21:34.941 9.434 - 9.481: 98.4216% ( 2) 00:21:34.941 9.481 - 9.529: 98.4374% ( 2) 00:21:34.941 9.624 - 9.671: 98.4453% ( 1) 00:21:34.941 9.671 - 9.719: 98.4532% ( 1) 00:21:34.941 9.813 - 9.861: 98.4611% ( 1) 00:21:34.941 9.861 - 9.908: 98.4689% ( 1) 00:21:34.941 9.908 - 9.956: 98.4768% ( 1) 00:21:34.941 9.956 - 10.003: 98.4847% ( 1) 00:21:34.941 10.003 - 10.050: 98.5005% ( 2) 00:21:34.941 10.050 - 10.098: 98.5084% ( 1) 00:21:34.941 10.193 - 10.240: 98.5163% ( 1) 00:21:34.941 10.335 - 10.382: 98.5242% ( 1) 00:21:34.941 10.382 - 10.430: 98.5400% ( 2) 00:21:34.941 10.430 - 10.477: 98.5479% ( 1) 00:21:34.941 10.572 - 10.619: 98.5558% ( 1) 00:21:34.941 10.619 - 10.667: 98.5636% ( 1) 00:21:34.941 10.667 - 10.714: 98.5715% ( 1) 00:21:34.941 10.714 - 10.761: 98.5873% ( 2) 00:21:34.941 10.809 - 10.856: 98.6031% ( 2) 00:21:34.941 11.236 - 11.283: 98.6110% ( 1) 00:21:34.941 11.330 - 11.378: 98.6189% ( 1) 00:21:34.941 11.425 - 11.473: 98.6268% ( 1) 00:21:34.941 11.520 - 11.567: 98.6347% ( 1) 00:21:34.942 11.615 - 11.662: 98.6505% ( 2) 00:21:34.942 11.710 - 11.757: 98.6584% ( 1) 00:21:34.942 11.757 - 11.804: 98.6662% ( 1) 00:21:34.942 11.899 - 11.947: 98.6741% ( 1) 00:21:34.942 11.947 - 11.994: 98.6899% ( 2) 00:21:34.942 12.089 - 12.136: 98.6978% ( 1) 00:21:34.942 12.136 - 12.231: 98.7057% ( 1) 00:21:34.942 12.231 - 12.326: 98.7136% ( 1) 00:21:34.942 12.421 - 12.516: 98.7294% ( 2) 00:21:34.942 12.516 - 12.610: 98.7452% ( 2) 00:21:34.942 12.800 - 12.895: 98.7531% ( 1) 00:21:34.942 12.895 - 12.990: 98.7610% ( 1) 00:21:34.942 12.990 - 13.084: 98.7688% ( 1) 00:21:34.942 13.084 - 13.179: 98.7767% ( 1) 00:21:34.942 13.179 - 13.274: 98.7925% ( 2) 00:21:34.942 13.274 - 13.369: 98.8162% ( 3) 00:21:34.942 13.464 - 13.559: 98.8241% ( 1) 00:21:34.942 13.748 - 13.843: 98.8399% ( 2) 00:21:34.942 13.843 - 13.938: 98.8478% ( 1) 00:21:34.942 14.317 - 14.412: 98.8557% ( 1) 00:21:34.942 14.412 - 14.507: 98.8635% ( 1) 00:21:34.942 14.507 - 14.601: 98.8793% ( 2) 00:21:34.942 14.601 - 14.696: 98.8872% ( 1) 00:21:34.942 14.696 - 14.791: 98.9030% ( 2) 00:21:34.942 14.981 - 15.076: 98.9109% ( 1) 00:21:34.942 16.119 - 16.213: 98.9188% ( 1) 00:21:34.942 16.213 - 16.308: 98.9267% ( 1) 00:21:34.942 16.782 - 16.877: 98.9346% ( 1) 00:21:34.942 17.067 - 17.161: 98.9425% ( 1) 00:21:34.942 17.161 - 17.256: 98.9661% ( 3) 00:21:34.942 17.351 - 17.446: 99.0056% ( 5) 00:21:34.942 17.446 - 17.541: 99.0214% ( 2) 00:21:34.942 17.541 - 17.636: 99.0845% ( 8) 00:21:34.942 17.636 - 17.730: 99.1556% ( 9) 00:21:34.942 17.730 - 17.825: 99.2187% ( 8) 00:21:34.942 17.825 - 17.920: 99.2660% ( 6) 00:21:34.942 17.920 - 18.015: 99.3213% ( 7) 00:21:34.942 18.015 - 18.110: 99.3844% ( 8) 00:21:34.942 18.110 - 18.204: 99.4239% ( 5) 00:21:34.942 18.204 - 18.299: 99.4712% ( 6) 00:21:34.942 18.299 - 18.394: 99.5107% ( 5) 00:21:34.942 18.394 - 18.489: 99.5659% ( 7) 00:21:34.942 18.489 - 18.584: 99.6291% ( 8) 00:21:34.942 18.584 - 18.679: 99.6922% ( 8) 00:21:34.942 18.679 - 18.773: 99.7317% ( 5) 00:21:34.942 18.773 - 18.868: 99.7475% ( 2) 00:21:34.942 18.868 - 18.963: 99.7553% ( 1) 00:21:34.942 18.963 - 19.058: 99.7632% ( 1) 00:21:34.942 19.153 - 19.247: 99.7711% ( 1) 00:21:34.942 19.437 - 19.532: 99.7869% ( 2) 00:21:34.942 19.627 - 19.721: 99.7948% ( 1) 00:21:34.942 19.721 - 19.816: 99.8027% ( 1) 00:21:34.942 20.006 - 20.101: 99.8106% ( 1) 00:21:34.942 20.101 - 20.196: 99.8185% ( 1) 00:21:34.942 22.092 - 22.187: 99.8264% ( 1) 00:21:34.942 22.281 - 22.376: 99.8343% ( 1) 00:21:34.942 25.600 - 25.790: 99.8501% ( 2) 00:21:34.942 26.548 - 26.738: 99.8579% ( 1) 00:21:34.942 27.876 - 28.065: 99.8658% ( 1) 00:21:34.942 28.444 - 28.634: 99.8737% ( 1) 00:21:34.942 28.634 - 28.824: 99.8816% ( 1) 00:21:34.942 29.772 - 29.961: 99.8895% ( 1) 00:21:34.942 31.289 - 31.479: 99.8974% ( 1) 00:21:34.942 34.133 - 34.323: 99.9053% ( 1) 00:21:34.942 35.650 - 35.840: 99.9132% ( 1) 00:21:34.942 3980.705 - 4004.978: 99.9842% ( 9) 00:21:34.942 4004.978 - 4029.250: 100.0000% ( 2) 00:21:34.942 00:21:34.942 Complete histogram 00:21:34.942 ================== 00:21:34.942 Range in us Cumulative Count 00:21:34.942 2.039 - 2.050: 0.3078% ( 39) 00:21:34.942 2.050 - 2.062: 14.6397% ( 1816) 00:21:34.942 2.062 - 2.074: 27.2591% ( 1599) 00:21:34.942 2.074 - 2.086: 35.1511% ( 1000) 00:21:34.942 2.086 - 2.098: 49.6646% ( 1839) 00:21:34.942 2.098 - 2.110: 56.3807% ( 851) 00:21:34.942 2.110 - 2.121: 61.2817% ( 621) 00:21:34.942 2.121 - 2.133: 70.3338% ( 1147) 00:21:34.942 2.133 - 2.145: 74.7849% ( 564) 00:21:34.942 2.145 - 2.157: 78.5889% ( 482) 00:21:34.942 2.157 - 2.169: 84.6421% ( 767) 00:21:34.942 2.169 - 2.181: 86.9071% ( 287) 00:21:34.942 2.181 - 2.193: 88.4145% ( 191) 00:21:34.942 2.193 - 2.204: 90.2060% ( 227) 00:21:34.942 2.204 - 2.216: 92.1711% ( 249) 00:21:34.942 2.216 - 2.228: 93.5996% ( 181) 00:21:34.942 2.228 - 2.240: 94.4835% ( 112) 00:21:34.942 2.240 - 2.252: 94.9096% ( 54) 00:21:34.942 2.252 - 2.264: 95.0833% ( 22) 00:21:34.942 2.264 - 2.276: 95.3121% ( 29) 00:21:34.942 2.276 - 2.287: 95.6752% ( 46) 00:21:34.942 2.287 - 2.299: 95.8172% ( 18) 00:21:34.942 2.299 - 2.311: 95.9356% ( 15) 00:21:34.942 2.311 - 2.323: 96.0066% ( 9) 00:21:34.942 2.323 - 2.335: 96.0619% ( 7) 00:21:34.942 2.335 - 2.347: 96.1250% ( 8) 00:21:34.942 2.347 - 2.359: 96.2592% ( 17) 00:21:34.942 2.359 - 2.370: 96.6064% ( 44) 00:21:34.942 2.370 - 2.382: 96.7643% ( 20) 00:21:34.942 2.382 - 2.394: 97.0878% ( 41) 00:21:34.942 2.394 - 2.406: 97.3720% ( 36) 00:21:34.942 2.406 - 2.418: 97.5219% ( 19) 00:21:34.942 2.418 - 2.430: 97.7350% ( 27) 00:21:34.942 2.430 - 2.441: 97.9086% ( 22) 00:21:34.942 2.441 - 2.453: 98.0112% ( 13) 00:21:34.942 2.453 - 2.465: 98.1533% ( 18) 00:21:34.942 2.465 - 2.477: 98.2401% ( 11) 00:21:34.942 2.477 - 2.489: 98.2716% ( 4) 00:21:34.942 2.489 - 2.501: 98.3269% ( 7) 00:21:34.942 2.501 - 2.513: 98.3663% ( 5) 00:21:34.942 2.513 - 2.524: 9[2024-12-06 17:37:16.773454] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:21:35.200 8.3821% ( 2) 00:21:35.200 2.524 - 2.536: 98.4216% ( 5) 00:21:35.200 2.548 - 2.560: 98.4453% ( 3) 00:21:35.200 2.560 - 2.572: 98.4611% ( 2) 00:21:35.200 2.631 - 2.643: 98.4768% ( 2) 00:21:35.200 2.667 - 2.679: 98.4847% ( 1) 00:21:35.200 3.247 - 3.271: 98.4926% ( 1) 00:21:35.200 3.295 - 3.319: 98.5005% ( 1) 00:21:35.200 3.366 - 3.390: 98.5163% ( 2) 00:21:35.200 3.390 - 3.413: 98.5242% ( 1) 00:21:35.200 3.437 - 3.461: 98.5321% ( 1) 00:21:35.200 3.484 - 3.508: 98.5715% ( 5) 00:21:35.200 3.508 - 3.532: 98.5873% ( 2) 00:21:35.200 3.532 - 3.556: 98.6031% ( 2) 00:21:35.200 3.556 - 3.579: 98.6189% ( 2) 00:21:35.200 3.603 - 3.627: 98.6268% ( 1) 00:21:35.200 3.627 - 3.650: 98.6505% ( 3) 00:21:35.200 3.674 - 3.698: 98.6662% ( 2) 00:21:35.200 3.698 - 3.721: 98.6899% ( 3) 00:21:35.200 3.721 - 3.745: 98.7057% ( 2) 00:21:35.200 3.745 - 3.769: 98.7136% ( 1) 00:21:35.200 3.769 - 3.793: 98.7215% ( 1) 00:21:35.200 3.816 - 3.840: 98.7294% ( 1) 00:21:35.200 3.887 - 3.911: 98.7373% ( 1) 00:21:35.200 3.959 - 3.982: 98.7452% ( 1) 00:21:35.200 3.982 - 4.006: 98.7531% ( 1) 00:21:35.200 4.006 - 4.030: 98.7688% ( 2) 00:21:35.200 4.267 - 4.290: 98.7767% ( 1) 00:21:35.200 4.930 - 4.954: 98.7846% ( 1) 00:21:35.200 5.476 - 5.499: 98.7925% ( 1) 00:21:35.200 5.831 - 5.855: 98.8004% ( 1) 00:21:35.200 5.997 - 6.021: 98.8162% ( 2) 00:21:35.200 6.447 - 6.495: 98.8241% ( 1) 00:21:35.200 6.495 - 6.542: 98.8320% ( 1) 00:21:35.200 6.684 - 6.732: 98.8399% ( 1) 00:21:35.200 7.111 - 7.159: 98.8478% ( 1) 00:21:35.200 7.585 - 7.633: 98.8557% ( 1) 00:21:35.200 7.680 - 7.727: 98.8635% ( 1) 00:21:35.200 7.775 - 7.822: 98.8714% ( 1) 00:21:35.200 8.296 - 8.344: 98.8793% ( 1) 00:21:35.200 8.439 - 8.486: 98.8872% ( 1) 00:21:35.200 8.628 - 8.676: 98.8951% ( 1) 00:21:35.200 10.050 - 10.098: 98.9030% ( 1) 00:21:35.200 13.084 - 13.179: 98.9109% ( 1) 00:21:35.201 15.360 - 15.455: 98.9188% ( 1) 00:21:35.201 15.455 - 15.550: 98.9267% ( 1) 00:21:35.201 15.644 - 15.739: 98.9661% ( 5) 00:21:35.201 15.739 - 15.834: 98.9819% ( 2) 00:21:35.201 15.834 - 15.929: 99.0056% ( 3) 00:21:35.201 15.929 - 16.024: 99.0214% ( 2) 00:21:35.201 16.024 - 16.119: 99.0845% ( 8) 00:21:35.201 16.119 - 16.213: 99.1319% ( 6) 00:21:35.201 16.308 - 16.403: 99.1556% ( 3) 00:21:35.201 16.403 - 16.498: 99.1713% ( 2) 00:21:35.201 16.498 - 16.593: 99.2108% ( 5) 00:21:35.201 16.593 - 16.687: 99.2266% ( 2) 00:21:35.201 16.687 - 16.782: 99.2345% ( 1) 00:21:35.201 16.782 - 16.877: 99.3055% ( 9) 00:21:35.201 16.877 - 16.972: 99.3213% ( 2) 00:21:35.201 16.972 - 17.067: 99.3371% ( 2) 00:21:35.201 17.067 - 17.161: 99.3450% ( 1) 00:21:35.201 17.161 - 17.256: 99.3529% ( 1) 00:21:35.201 17.256 - 17.351: 99.3686% ( 2) 00:21:35.201 107.710 - 108.468: 99.3765% ( 1) 00:21:35.201 3737.979 - 3762.252: 99.3844% ( 1) 00:21:35.201 3883.615 - 3907.887: 99.3923% ( 1) 00:21:35.201 3980.705 - 4004.978: 99.8579% ( 59) 00:21:35.201 4004.978 - 4029.250: 100.0000% ( 18) 00:21:35.201 00:21:35.201 17:37:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@90 -- # aer_vfio_user /var/run/vfio-user/domain/vfio-user2/2 nqn.2019-07.io.spdk:cnode2 2 00:21:35.201 17:37:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@22 -- # local traddr=/var/run/vfio-user/domain/vfio-user2/2 00:21:35.201 17:37:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@23 -- # local subnqn=nqn.2019-07.io.spdk:cnode2 00:21:35.201 17:37:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@24 -- # local malloc_num=Malloc4 00:21:35.201 17:37:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_get_subsystems 00:21:35.460 [ 00:21:35.460 { 00:21:35.460 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:21:35.460 "subtype": "Discovery", 00:21:35.460 "listen_addresses": [], 00:21:35.460 "allow_any_host": true, 00:21:35.460 "hosts": [] 00:21:35.460 }, 00:21:35.460 { 00:21:35.460 "nqn": "nqn.2019-07.io.spdk:cnode1", 00:21:35.460 "subtype": "NVMe", 00:21:35.460 "listen_addresses": [ 00:21:35.460 { 00:21:35.460 "trtype": "VFIOUSER", 00:21:35.460 "adrfam": "IPv4", 00:21:35.460 "traddr": "/var/run/vfio-user/domain/vfio-user1/1", 00:21:35.460 "trsvcid": "0" 00:21:35.460 } 00:21:35.460 ], 00:21:35.460 "allow_any_host": true, 00:21:35.460 "hosts": [], 00:21:35.460 "serial_number": "SPDK1", 00:21:35.460 "model_number": "SPDK bdev Controller", 00:21:35.460 "max_namespaces": 32, 00:21:35.460 "min_cntlid": 1, 00:21:35.460 "max_cntlid": 65519, 00:21:35.460 "namespaces": [ 00:21:35.460 { 00:21:35.460 "nsid": 1, 00:21:35.460 "bdev_name": "Malloc1", 00:21:35.460 "name": "Malloc1", 00:21:35.460 "nguid": "99BB49AF08BC4B73A3D6CDD99170A2EF", 00:21:35.460 "uuid": "99bb49af-08bc-4b73-a3d6-cdd99170a2ef" 00:21:35.460 }, 00:21:35.460 { 00:21:35.460 "nsid": 2, 00:21:35.460 "bdev_name": "Malloc3", 00:21:35.460 "name": "Malloc3", 00:21:35.460 "nguid": "166E6D91FF4143A8B1281BD92280B027", 00:21:35.460 "uuid": "166e6d91-ff41-43a8-b128-1bd92280b027" 00:21:35.460 } 00:21:35.460 ] 00:21:35.460 }, 00:21:35.460 { 00:21:35.460 "nqn": "nqn.2019-07.io.spdk:cnode2", 00:21:35.460 "subtype": "NVMe", 00:21:35.460 "listen_addresses": [ 00:21:35.460 { 00:21:35.460 "trtype": "VFIOUSER", 00:21:35.460 "adrfam": "IPv4", 00:21:35.460 "traddr": "/var/run/vfio-user/domain/vfio-user2/2", 00:21:35.460 "trsvcid": "0" 00:21:35.460 } 00:21:35.460 ], 00:21:35.460 "allow_any_host": true, 00:21:35.460 "hosts": [], 00:21:35.460 "serial_number": "SPDK2", 00:21:35.460 "model_number": "SPDK bdev Controller", 00:21:35.460 "max_namespaces": 32, 00:21:35.460 "min_cntlid": 1, 00:21:35.460 "max_cntlid": 65519, 00:21:35.460 "namespaces": [ 00:21:35.460 { 00:21:35.460 "nsid": 1, 00:21:35.460 "bdev_name": "Malloc2", 00:21:35.460 "name": "Malloc2", 00:21:35.460 "nguid": "064BC5F836974A08AA1E011B1094E1E2", 00:21:35.460 "uuid": "064bc5f8-3697-4a08-aa1e-011b1094e1e2" 00:21:35.460 } 00:21:35.460 ] 00:21:35.460 } 00:21:35.460 ] 00:21:35.460 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@27 -- # AER_TOUCH_FILE=/tmp/aer_touch_file 00:21:35.460 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@34 -- # aerpid=244703 00:21:35.460 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/aer/aer -r ' trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -n 2 -g -t /tmp/aer_touch_file 00:21:35.460 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@37 -- # waitforfile /tmp/aer_touch_file 00:21:35.460 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1269 -- # local i=0 00:21:35.460 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1270 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:21:35.460 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1271 -- # '[' 0 -lt 200 ']' 00:21:35.460 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1272 -- # i=1 00:21:35.460 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1273 -- # sleep 0.1 00:21:35.460 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1270 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:21:35.460 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1271 -- # '[' 1 -lt 200 ']' 00:21:35.460 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1272 -- # i=2 00:21:35.460 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1273 -- # sleep 0.1 00:21:35.718 [2024-12-06 17:37:17.305172] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:21:35.718 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1270 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:21:35.718 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1276 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:21:35.718 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1280 -- # return 0 00:21:35.718 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@38 -- # rm -f /tmp/aer_touch_file 00:21:35.718 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 --name Malloc4 00:21:35.977 Malloc4 00:21:35.977 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode2 Malloc4 -n 2 00:21:36.235 [2024-12-06 17:37:17.890522] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:21:36.235 17:37:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_get_subsystems 00:21:36.235 Asynchronous Event Request test 00:21:36.235 Attaching to /var/run/vfio-user/domain/vfio-user2/2 00:21:36.235 Attached to /var/run/vfio-user/domain/vfio-user2/2 00:21:36.235 Registering asynchronous event callbacks... 00:21:36.235 Starting namespace attribute notice tests for all controllers... 00:21:36.235 /var/run/vfio-user/domain/vfio-user2/2: aer_cb for log page 4, aen_event_type: 0x02, aen_event_info: 0x00 00:21:36.235 aer_cb - Changed Namespace 00:21:36.235 Cleaning up... 00:21:36.493 [ 00:21:36.493 { 00:21:36.493 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:21:36.493 "subtype": "Discovery", 00:21:36.493 "listen_addresses": [], 00:21:36.493 "allow_any_host": true, 00:21:36.493 "hosts": [] 00:21:36.493 }, 00:21:36.493 { 00:21:36.493 "nqn": "nqn.2019-07.io.spdk:cnode1", 00:21:36.493 "subtype": "NVMe", 00:21:36.493 "listen_addresses": [ 00:21:36.493 { 00:21:36.493 "trtype": "VFIOUSER", 00:21:36.493 "adrfam": "IPv4", 00:21:36.493 "traddr": "/var/run/vfio-user/domain/vfio-user1/1", 00:21:36.493 "trsvcid": "0" 00:21:36.493 } 00:21:36.493 ], 00:21:36.493 "allow_any_host": true, 00:21:36.493 "hosts": [], 00:21:36.493 "serial_number": "SPDK1", 00:21:36.493 "model_number": "SPDK bdev Controller", 00:21:36.493 "max_namespaces": 32, 00:21:36.493 "min_cntlid": 1, 00:21:36.493 "max_cntlid": 65519, 00:21:36.493 "namespaces": [ 00:21:36.493 { 00:21:36.493 "nsid": 1, 00:21:36.493 "bdev_name": "Malloc1", 00:21:36.493 "name": "Malloc1", 00:21:36.493 "nguid": "99BB49AF08BC4B73A3D6CDD99170A2EF", 00:21:36.493 "uuid": "99bb49af-08bc-4b73-a3d6-cdd99170a2ef" 00:21:36.493 }, 00:21:36.493 { 00:21:36.493 "nsid": 2, 00:21:36.493 "bdev_name": "Malloc3", 00:21:36.493 "name": "Malloc3", 00:21:36.493 "nguid": "166E6D91FF4143A8B1281BD92280B027", 00:21:36.493 "uuid": "166e6d91-ff41-43a8-b128-1bd92280b027" 00:21:36.493 } 00:21:36.493 ] 00:21:36.493 }, 00:21:36.493 { 00:21:36.493 "nqn": "nqn.2019-07.io.spdk:cnode2", 00:21:36.493 "subtype": "NVMe", 00:21:36.493 "listen_addresses": [ 00:21:36.493 { 00:21:36.493 "trtype": "VFIOUSER", 00:21:36.493 "adrfam": "IPv4", 00:21:36.493 "traddr": "/var/run/vfio-user/domain/vfio-user2/2", 00:21:36.493 "trsvcid": "0" 00:21:36.493 } 00:21:36.493 ], 00:21:36.493 "allow_any_host": true, 00:21:36.493 "hosts": [], 00:21:36.493 "serial_number": "SPDK2", 00:21:36.493 "model_number": "SPDK bdev Controller", 00:21:36.493 "max_namespaces": 32, 00:21:36.493 "min_cntlid": 1, 00:21:36.493 "max_cntlid": 65519, 00:21:36.493 "namespaces": [ 00:21:36.493 { 00:21:36.493 "nsid": 1, 00:21:36.493 "bdev_name": "Malloc2", 00:21:36.493 "name": "Malloc2", 00:21:36.493 "nguid": "064BC5F836974A08AA1E011B1094E1E2", 00:21:36.493 "uuid": "064bc5f8-3697-4a08-aa1e-011b1094e1e2" 00:21:36.493 }, 00:21:36.493 { 00:21:36.493 "nsid": 2, 00:21:36.493 "bdev_name": "Malloc4", 00:21:36.493 "name": "Malloc4", 00:21:36.493 "nguid": "9D869876C49042218F6B483CC3791D9F", 00:21:36.493 "uuid": "9d869876-c490-4221-8f6b-483cc3791d9f" 00:21:36.493 } 00:21:36.493 ] 00:21:36.493 } 00:21:36.493 ] 00:21:36.493 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@44 -- # wait 244703 00:21:36.493 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@105 -- # stop_nvmf_vfio_user 00:21:36.493 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@95 -- # killprocess 239114 00:21:36.493 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@954 -- # '[' -z 239114 ']' 00:21:36.493 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@958 -- # kill -0 239114 00:21:36.494 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@959 -- # uname 00:21:36.494 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:21:36.494 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 239114 00:21:36.494 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:21:36.494 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:21:36.494 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@972 -- # echo 'killing process with pid 239114' 00:21:36.494 killing process with pid 239114 00:21:36.494 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@973 -- # kill 239114 00:21:36.494 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@978 -- # wait 239114 00:21:36.775 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@97 -- # rm -rf /var/run/vfio-user 00:21:36.775 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@99 -- # trap - SIGINT SIGTERM EXIT 00:21:36.775 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@108 -- # setup_nvmf_vfio_user --interrupt-mode '-M -I' 00:21:36.775 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@51 -- # local nvmf_app_args=--interrupt-mode 00:21:36.775 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@52 -- # local 'transport_args=-M -I' 00:21:36.775 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@55 -- # nvmfpid=244965 00:21:36.775 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m '[0,1,2,3]' --interrupt-mode 00:21:36.775 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@57 -- # echo 'Process pid: 244965' 00:21:36.775 Process pid: 244965 00:21:36.775 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@59 -- # trap 'killprocess $nvmfpid; exit 1' SIGINT SIGTERM EXIT 00:21:36.775 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@60 -- # waitforlisten 244965 00:21:36.775 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@835 -- # '[' -z 244965 ']' 00:21:36.775 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:36.775 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@840 -- # local max_retries=100 00:21:36.775 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:36.775 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:36.775 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@844 -- # xtrace_disable 00:21:36.775 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@10 -- # set +x 00:21:36.775 [2024-12-06 17:37:18.565800] thread.c:3005:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:21:36.775 [2024-12-06 17:37:18.566843] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:21:36.775 [2024-12-06 17:37:18.566913] [ DPDK EAL parameters: nvmf -l 0,1,2,3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:21:37.035 [2024-12-06 17:37:18.633698] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:21:37.035 [2024-12-06 17:37:18.676382] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:21:37.035 [2024-12-06 17:37:18.676442] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:21:37.035 [2024-12-06 17:37:18.676469] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:21:37.035 [2024-12-06 17:37:18.676481] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:21:37.035 [2024-12-06 17:37:18.676490] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:21:37.035 [2024-12-06 17:37:18.677873] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:21:37.035 [2024-12-06 17:37:18.677934] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:21:37.035 [2024-12-06 17:37:18.678003] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:21:37.035 [2024-12-06 17:37:18.678006] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:21:37.035 [2024-12-06 17:37:18.758577] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:21:37.035 [2024-12-06 17:37:18.758803] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:21:37.035 [2024-12-06 17:37:18.759077] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:21:37.035 [2024-12-06 17:37:18.759662] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:21:37.035 [2024-12-06 17:37:18.759897] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_003) to intr mode from intr mode. 00:21:37.035 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:21:37.035 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@868 -- # return 0 00:21:37.035 17:37:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@62 -- # sleep 1 00:21:37.974 17:37:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t VFIOUSER -M -I 00:21:38.542 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@66 -- # mkdir -p /var/run/vfio-user 00:21:38.543 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # seq 1 2 00:21:38.543 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # for i in $(seq 1 $NUM_DEVICES) 00:21:38.543 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@69 -- # mkdir -p /var/run/vfio-user/domain/vfio-user1/1 00:21:38.543 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc1 00:21:38.802 Malloc1 00:21:38.802 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2019-07.io.spdk:cnode1 -a -s SPDK1 00:21:39.060 17:37:20 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode1 Malloc1 00:21:39.319 17:37:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2019-07.io.spdk:cnode1 -t VFIOUSER -a /var/run/vfio-user/domain/vfio-user1/1 -s 0 00:21:39.577 17:37:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # for i in $(seq 1 $NUM_DEVICES) 00:21:39.577 17:37:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@69 -- # mkdir -p /var/run/vfio-user/domain/vfio-user2/2 00:21:39.577 17:37:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc2 00:21:40.143 Malloc2 00:21:40.143 17:37:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2019-07.io.spdk:cnode2 -a -s SPDK2 00:21:40.400 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode2 Malloc2 00:21:40.658 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2019-07.io.spdk:cnode2 -t VFIOUSER -a /var/run/vfio-user/domain/vfio-user2/2 -s 0 00:21:40.915 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@109 -- # stop_nvmf_vfio_user 00:21:40.915 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@95 -- # killprocess 244965 00:21:40.915 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@954 -- # '[' -z 244965 ']' 00:21:40.915 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@958 -- # kill -0 244965 00:21:40.915 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@959 -- # uname 00:21:40.915 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:21:40.915 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 244965 00:21:40.915 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:21:40.915 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:21:40.915 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@972 -- # echo 'killing process with pid 244965' 00:21:40.915 killing process with pid 244965 00:21:40.915 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@973 -- # kill 244965 00:21:40.915 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@978 -- # wait 244965 00:21:41.173 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@97 -- # rm -rf /var/run/vfio-user 00:21:41.173 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@99 -- # trap - SIGINT SIGTERM EXIT 00:21:41.173 00:21:41.173 real 0m53.849s 00:21:41.173 user 3m28.252s 00:21:41.173 sys 0m3.943s 00:21:41.173 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1130 -- # xtrace_disable 00:21:41.173 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@10 -- # set +x 00:21:41.173 ************************************ 00:21:41.173 END TEST nvmf_vfio_user 00:21:41.173 ************************************ 00:21:41.173 17:37:22 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@32 -- # run_test nvmf_vfio_user_nvme_compliance /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/compliance/compliance.sh --transport=tcp 00:21:41.173 17:37:22 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:21:41.173 17:37:22 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:21:41.173 17:37:22 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:21:41.173 ************************************ 00:21:41.173 START TEST nvmf_vfio_user_nvme_compliance 00:21:41.173 ************************************ 00:21:41.173 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/compliance/compliance.sh --transport=tcp 00:21:41.173 * Looking for test storage... 00:21:41.173 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/compliance 00:21:41.173 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:21:41.173 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1711 -- # lcov --version 00:21:41.173 17:37:22 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@333 -- # local ver1 ver1_l 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@334 -- # local ver2 ver2_l 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@336 -- # IFS=.-: 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@336 -- # read -ra ver1 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@337 -- # IFS=.-: 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@337 -- # read -ra ver2 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@338 -- # local 'op=<' 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@340 -- # ver1_l=2 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@341 -- # ver2_l=1 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@344 -- # case "$op" in 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@345 -- # : 1 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@364 -- # (( v = 0 )) 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@365 -- # decimal 1 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@353 -- # local d=1 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@355 -- # echo 1 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@365 -- # ver1[v]=1 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@366 -- # decimal 2 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@353 -- # local d=2 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@355 -- # echo 2 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@366 -- # ver2[v]=2 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@368 -- # return 0 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:21:41.433 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:21:41.433 --rc genhtml_branch_coverage=1 00:21:41.433 --rc genhtml_function_coverage=1 00:21:41.433 --rc genhtml_legend=1 00:21:41.433 --rc geninfo_all_blocks=1 00:21:41.433 --rc geninfo_unexecuted_blocks=1 00:21:41.433 00:21:41.433 ' 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:21:41.433 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:21:41.433 --rc genhtml_branch_coverage=1 00:21:41.433 --rc genhtml_function_coverage=1 00:21:41.433 --rc genhtml_legend=1 00:21:41.433 --rc geninfo_all_blocks=1 00:21:41.433 --rc geninfo_unexecuted_blocks=1 00:21:41.433 00:21:41.433 ' 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:21:41.433 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:21:41.433 --rc genhtml_branch_coverage=1 00:21:41.433 --rc genhtml_function_coverage=1 00:21:41.433 --rc genhtml_legend=1 00:21:41.433 --rc geninfo_all_blocks=1 00:21:41.433 --rc geninfo_unexecuted_blocks=1 00:21:41.433 00:21:41.433 ' 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:21:41.433 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:21:41.433 --rc genhtml_branch_coverage=1 00:21:41.433 --rc genhtml_function_coverage=1 00:21:41.433 --rc genhtml_legend=1 00:21:41.433 --rc geninfo_all_blocks=1 00:21:41.433 --rc geninfo_unexecuted_blocks=1 00:21:41.433 00:21:41.433 ' 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@7 -- # uname -s 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@15 -- # shopt -s extglob 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:21:41.433 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- paths/export.sh@5 -- # export PATH 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@51 -- # : 0 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:21:41.434 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@55 -- # have_pci_nics=0 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@11 -- # MALLOC_BDEV_SIZE=64 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@14 -- # export TEST_TRANSPORT=VFIOUSER 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@14 -- # TEST_TRANSPORT=VFIOUSER 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@16 -- # rm -rf /var/run/vfio-user 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@20 -- # nvmfpid=245577 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x7 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@21 -- # echo 'Process pid: 245577' 00:21:41.434 Process pid: 245577 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@23 -- # trap 'killprocess $nvmfpid; exit 1' SIGINT SIGTERM EXIT 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@24 -- # waitforlisten 245577 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@835 -- # '[' -z 245577 ']' 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@840 -- # local max_retries=100 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:41.434 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@844 -- # xtrace_disable 00:21:41.434 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:21:41.434 [2024-12-06 17:37:23.091089] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:21:41.434 [2024-12-06 17:37:23.091179] [ DPDK EAL parameters: nvmf -c 0x7 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:21:41.434 [2024-12-06 17:37:23.157480] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:21:41.434 [2024-12-06 17:37:23.202907] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:21:41.434 [2024-12-06 17:37:23.202978] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:21:41.434 [2024-12-06 17:37:23.203006] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:21:41.434 [2024-12-06 17:37:23.203017] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:21:41.434 [2024-12-06 17:37:23.203032] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:21:41.434 [2024-12-06 17:37:23.204299] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:21:41.434 [2024-12-06 17:37:23.204366] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:21:41.434 [2024-12-06 17:37:23.204363] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:21:41.693 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:21:41.693 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@868 -- # return 0 00:21:41.693 17:37:23 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@26 -- # sleep 1 00:21:42.627 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@28 -- # nqn=nqn.2021-09.io.spdk:cnode0 00:21:42.627 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@29 -- # traddr=/var/run/vfio-user 00:21:42.627 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@31 -- # rpc_cmd nvmf_create_transport -t VFIOUSER 00:21:42.627 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@563 -- # xtrace_disable 00:21:42.627 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:21:42.627 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:21:42.627 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@33 -- # mkdir -p /var/run/vfio-user 00:21:42.627 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@35 -- # rpc_cmd bdev_malloc_create 64 512 -b malloc0 00:21:42.627 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@563 -- # xtrace_disable 00:21:42.627 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:21:42.627 malloc0 00:21:42.627 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:21:42.627 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@36 -- # rpc_cmd nvmf_create_subsystem nqn.2021-09.io.spdk:cnode0 -a -s spdk -m 32 00:21:42.627 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@563 -- # xtrace_disable 00:21:42.627 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:21:42.627 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:21:42.627 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@37 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2021-09.io.spdk:cnode0 malloc0 00:21:42.627 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@563 -- # xtrace_disable 00:21:42.627 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:21:42.627 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:21:42.627 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@38 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2021-09.io.spdk:cnode0 -t VFIOUSER -a /var/run/vfio-user -s 0 00:21:42.627 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@563 -- # xtrace_disable 00:21:42.627 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:21:42.627 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:21:42.627 17:37:24 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/compliance/nvme_compliance -g -r 'trtype:VFIOUSER traddr:/var/run/vfio-user subnqn:nqn.2021-09.io.spdk:cnode0' 00:21:42.885 00:21:42.885 00:21:42.885 CUnit - A unit testing framework for C - Version 2.1-3 00:21:42.885 http://cunit.sourceforge.net/ 00:21:42.885 00:21:42.885 00:21:42.885 Suite: nvme_compliance 00:21:42.885 Test: admin_identify_ctrlr_verify_dptr ...[2024-12-06 17:37:24.566992] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:21:42.885 [2024-12-06 17:37:24.568463] vfio_user.c: 832:nvme_cmd_map_prps: *ERROR*: no PRP2, 3072 remaining 00:21:42.885 [2024-12-06 17:37:24.568487] vfio_user.c:5544:map_admin_cmd_req: *ERROR*: /var/run/vfio-user: map Admin Opc 6 failed 00:21:42.885 [2024-12-06 17:37:24.568515] vfio_user.c:5637:handle_cmd_req: *ERROR*: /var/run/vfio-user: process NVMe command opc 0x6 failed 00:21:42.885 [2024-12-06 17:37:24.570002] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:21:42.885 passed 00:21:42.885 Test: admin_identify_ctrlr_verify_fused ...[2024-12-06 17:37:24.655565] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:21:42.885 [2024-12-06 17:37:24.658586] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:21:42.885 passed 00:21:43.142 Test: admin_identify_ns ...[2024-12-06 17:37:24.745510] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:21:43.142 [2024-12-06 17:37:24.804685] ctrlr.c:2752:_nvmf_ctrlr_get_ns_safe: *ERROR*: Identify Namespace for invalid NSID 0 00:21:43.142 [2024-12-06 17:37:24.812683] ctrlr.c:2752:_nvmf_ctrlr_get_ns_safe: *ERROR*: Identify Namespace for invalid NSID 4294967295 00:21:43.142 [2024-12-06 17:37:24.833807] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:21:43.142 passed 00:21:43.142 Test: admin_get_features_mandatory_features ...[2024-12-06 17:37:24.917658] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:21:43.142 [2024-12-06 17:37:24.920688] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:21:43.142 passed 00:21:43.409 Test: admin_get_features_optional_features ...[2024-12-06 17:37:25.005309] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:21:43.410 [2024-12-06 17:37:25.009330] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:21:43.410 passed 00:21:43.410 Test: admin_set_features_number_of_queues ...[2024-12-06 17:37:25.094620] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:21:43.410 [2024-12-06 17:37:25.196783] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:21:43.410 passed 00:21:43.667 Test: admin_get_log_page_mandatory_logs ...[2024-12-06 17:37:25.282005] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:21:43.667 [2024-12-06 17:37:25.285029] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:21:43.667 passed 00:21:43.667 Test: admin_get_log_page_with_lpo ...[2024-12-06 17:37:25.368309] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:21:43.667 [2024-12-06 17:37:25.435682] ctrlr.c:2699:nvmf_ctrlr_get_log_page: *ERROR*: Get log page: offset (516) > len (512) 00:21:43.667 [2024-12-06 17:37:25.448759] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:21:43.667 passed 00:21:43.924 Test: fabric_property_get ...[2024-12-06 17:37:25.534102] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:21:43.924 [2024-12-06 17:37:25.535383] vfio_user.c:5637:handle_cmd_req: *ERROR*: /var/run/vfio-user: process NVMe command opc 0x7f failed 00:21:43.924 [2024-12-06 17:37:25.537126] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:21:43.924 passed 00:21:43.924 Test: admin_delete_io_sq_use_admin_qid ...[2024-12-06 17:37:25.620641] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:21:43.924 [2024-12-06 17:37:25.621986] vfio_user.c:2329:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O sqid:0 does not exist 00:21:43.924 [2024-12-06 17:37:25.623689] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:21:43.924 passed 00:21:43.924 Test: admin_delete_io_sq_delete_sq_twice ...[2024-12-06 17:37:25.707920] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:21:44.181 [2024-12-06 17:37:25.791676] vfio_user.c:2329:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O sqid:1 does not exist 00:21:44.181 [2024-12-06 17:37:25.807678] vfio_user.c:2329:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O sqid:1 does not exist 00:21:44.181 [2024-12-06 17:37:25.812763] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:21:44.181 passed 00:21:44.181 Test: admin_delete_io_cq_use_admin_qid ...[2024-12-06 17:37:25.897414] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:21:44.181 [2024-12-06 17:37:25.898714] vfio_user.c:2329:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O cqid:0 does not exist 00:21:44.181 [2024-12-06 17:37:25.900433] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:21:44.181 passed 00:21:44.181 Test: admin_delete_io_cq_delete_cq_first ...[2024-12-06 17:37:25.981919] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:21:44.438 [2024-12-06 17:37:26.057678] vfio_user.c:2339:handle_del_io_q: *ERROR*: /var/run/vfio-user: the associated SQ must be deleted first 00:21:44.438 [2024-12-06 17:37:26.081677] vfio_user.c:2329:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O sqid:1 does not exist 00:21:44.438 [2024-12-06 17:37:26.086772] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:21:44.438 passed 00:21:44.438 Test: admin_create_io_cq_verify_iv_pc ...[2024-12-06 17:37:26.169116] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:21:44.438 [2024-12-06 17:37:26.170416] vfio_user.c:2178:handle_create_io_cq: *ERROR*: /var/run/vfio-user: IV is too big 00:21:44.438 [2024-12-06 17:37:26.170467] vfio_user.c:2172:handle_create_io_cq: *ERROR*: /var/run/vfio-user: non-PC CQ not supported 00:21:44.438 [2024-12-06 17:37:26.174147] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:21:44.438 passed 00:21:44.438 Test: admin_create_io_sq_verify_qsize_cqid ...[2024-12-06 17:37:26.255275] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:21:44.696 [2024-12-06 17:37:26.346681] vfio_user.c:2260:handle_create_io_q: *ERROR*: /var/run/vfio-user: invalid I/O queue size 1 00:21:44.696 [2024-12-06 17:37:26.354676] vfio_user.c:2260:handle_create_io_q: *ERROR*: /var/run/vfio-user: invalid I/O queue size 257 00:21:44.696 [2024-12-06 17:37:26.362675] vfio_user.c:2058:handle_create_io_sq: *ERROR*: /var/run/vfio-user: invalid cqid:0 00:21:44.696 [2024-12-06 17:37:26.370674] vfio_user.c:2058:handle_create_io_sq: *ERROR*: /var/run/vfio-user: invalid cqid:128 00:21:44.696 [2024-12-06 17:37:26.399792] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:21:44.696 passed 00:21:44.696 Test: admin_create_io_sq_verify_pc ...[2024-12-06 17:37:26.483404] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:21:44.696 [2024-12-06 17:37:26.499692] vfio_user.c:2071:handle_create_io_sq: *ERROR*: /var/run/vfio-user: non-PC SQ not supported 00:21:44.696 [2024-12-06 17:37:26.517762] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:21:44.952 passed 00:21:44.952 Test: admin_create_io_qp_max_qps ...[2024-12-06 17:37:26.603337] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:21:45.886 [2024-12-06 17:37:27.706683] nvme_ctrlr.c:5523:spdk_nvme_ctrlr_alloc_qid: *ERROR*: [/var/run/vfio-user, 0] No free I/O queue IDs 00:21:46.453 [2024-12-06 17:37:28.085813] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:21:46.453 passed 00:21:46.453 Test: admin_create_io_sq_shared_cq ...[2024-12-06 17:37:28.170215] vfio_user.c:2873:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:21:46.711 [2024-12-06 17:37:28.301672] vfio_user.c:2339:handle_del_io_q: *ERROR*: /var/run/vfio-user: the associated SQ must be deleted first 00:21:46.711 [2024-12-06 17:37:28.338761] vfio_user.c:2835:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:21:46.711 passed 00:21:46.711 00:21:46.711 Run Summary: Type Total Ran Passed Failed Inactive 00:21:46.711 suites 1 1 n/a 0 0 00:21:46.711 tests 18 18 18 0 0 00:21:46.711 asserts 360 360 360 0 n/a 00:21:46.711 00:21:46.711 Elapsed time = 1.562 seconds 00:21:46.711 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@42 -- # killprocess 245577 00:21:46.711 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@954 -- # '[' -z 245577 ']' 00:21:46.711 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@958 -- # kill -0 245577 00:21:46.711 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@959 -- # uname 00:21:46.711 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:21:46.711 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 245577 00:21:46.711 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:21:46.711 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:21:46.711 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@972 -- # echo 'killing process with pid 245577' 00:21:46.711 killing process with pid 245577 00:21:46.711 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@973 -- # kill 245577 00:21:46.711 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@978 -- # wait 245577 00:21:46.970 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@44 -- # rm -rf /var/run/vfio-user 00:21:46.970 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@46 -- # trap - SIGINT SIGTERM EXIT 00:21:46.970 00:21:46.970 real 0m5.749s 00:21:46.970 user 0m16.172s 00:21:46.970 sys 0m0.559s 00:21:46.970 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1130 -- # xtrace_disable 00:21:46.970 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:21:46.970 ************************************ 00:21:46.970 END TEST nvmf_vfio_user_nvme_compliance 00:21:46.970 ************************************ 00:21:46.970 17:37:28 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@33 -- # run_test nvmf_vfio_user_fuzz /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/vfio_user_fuzz.sh --transport=tcp 00:21:46.970 17:37:28 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:21:46.970 17:37:28 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:21:46.970 17:37:28 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:21:46.970 ************************************ 00:21:46.970 START TEST nvmf_vfio_user_fuzz 00:21:46.970 ************************************ 00:21:46.970 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/vfio_user_fuzz.sh --transport=tcp 00:21:46.970 * Looking for test storage... 00:21:46.970 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:21:46.970 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:21:46.970 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1711 -- # lcov --version 00:21:46.970 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@333 -- # local ver1 ver1_l 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@334 -- # local ver2 ver2_l 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@336 -- # IFS=.-: 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@336 -- # read -ra ver1 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@337 -- # IFS=.-: 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@337 -- # read -ra ver2 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@338 -- # local 'op=<' 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@340 -- # ver1_l=2 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@341 -- # ver2_l=1 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@344 -- # case "$op" in 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@345 -- # : 1 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@364 -- # (( v = 0 )) 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@365 -- # decimal 1 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@353 -- # local d=1 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@355 -- # echo 1 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@365 -- # ver1[v]=1 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@366 -- # decimal 2 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@353 -- # local d=2 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@355 -- # echo 2 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@366 -- # ver2[v]=2 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@368 -- # return 0 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:21:47.231 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:21:47.231 --rc genhtml_branch_coverage=1 00:21:47.231 --rc genhtml_function_coverage=1 00:21:47.231 --rc genhtml_legend=1 00:21:47.231 --rc geninfo_all_blocks=1 00:21:47.231 --rc geninfo_unexecuted_blocks=1 00:21:47.231 00:21:47.231 ' 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:21:47.231 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:21:47.231 --rc genhtml_branch_coverage=1 00:21:47.231 --rc genhtml_function_coverage=1 00:21:47.231 --rc genhtml_legend=1 00:21:47.231 --rc geninfo_all_blocks=1 00:21:47.231 --rc geninfo_unexecuted_blocks=1 00:21:47.231 00:21:47.231 ' 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:21:47.231 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:21:47.231 --rc genhtml_branch_coverage=1 00:21:47.231 --rc genhtml_function_coverage=1 00:21:47.231 --rc genhtml_legend=1 00:21:47.231 --rc geninfo_all_blocks=1 00:21:47.231 --rc geninfo_unexecuted_blocks=1 00:21:47.231 00:21:47.231 ' 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:21:47.231 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:21:47.231 --rc genhtml_branch_coverage=1 00:21:47.231 --rc genhtml_function_coverage=1 00:21:47.231 --rc genhtml_legend=1 00:21:47.231 --rc geninfo_all_blocks=1 00:21:47.231 --rc geninfo_unexecuted_blocks=1 00:21:47.231 00:21:47.231 ' 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@7 -- # uname -s 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@15 -- # shopt -s extglob 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- paths/export.sh@5 -- # export PATH 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@51 -- # : 0 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:21:47.231 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:21:47.231 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:21:47.232 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:21:47.232 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@55 -- # have_pci_nics=0 00:21:47.232 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@12 -- # MALLOC_BDEV_SIZE=64 00:21:47.232 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:21:47.232 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@15 -- # nqn=nqn.2021-09.io.spdk:cnode0 00:21:47.232 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@16 -- # traddr=/var/run/vfio-user 00:21:47.232 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@18 -- # export TEST_TRANSPORT=VFIOUSER 00:21:47.232 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@18 -- # TEST_TRANSPORT=VFIOUSER 00:21:47.232 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@20 -- # rm -rf /var/run/vfio-user 00:21:47.232 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@24 -- # nvmfpid=246302 00:21:47.232 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@23 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1 00:21:47.232 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@25 -- # echo 'Process pid: 246302' 00:21:47.232 Process pid: 246302 00:21:47.232 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@27 -- # trap 'killprocess $nvmfpid; exit 1' SIGINT SIGTERM EXIT 00:21:47.232 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@28 -- # waitforlisten 246302 00:21:47.232 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@835 -- # '[' -z 246302 ']' 00:21:47.232 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:47.232 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@840 -- # local max_retries=100 00:21:47.232 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:47.232 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:47.232 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@844 -- # xtrace_disable 00:21:47.232 17:37:28 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:21:47.490 17:37:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:21:47.490 17:37:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@868 -- # return 0 00:21:47.490 17:37:29 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@30 -- # sleep 1 00:21:48.426 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@32 -- # rpc_cmd nvmf_create_transport -t VFIOUSER 00:21:48.426 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@563 -- # xtrace_disable 00:21:48.426 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:21:48.426 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:21:48.426 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@34 -- # mkdir -p /var/run/vfio-user 00:21:48.426 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@36 -- # rpc_cmd bdev_malloc_create 64 512 -b malloc0 00:21:48.426 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@563 -- # xtrace_disable 00:21:48.426 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:21:48.426 malloc0 00:21:48.426 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:21:48.426 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@37 -- # rpc_cmd nvmf_create_subsystem nqn.2021-09.io.spdk:cnode0 -a -s spdk 00:21:48.426 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@563 -- # xtrace_disable 00:21:48.426 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:21:48.426 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:21:48.426 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@38 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2021-09.io.spdk:cnode0 malloc0 00:21:48.426 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@563 -- # xtrace_disable 00:21:48.426 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:21:48.426 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:21:48.426 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@39 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2021-09.io.spdk:cnode0 -t VFIOUSER -a /var/run/vfio-user -s 0 00:21:48.426 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@563 -- # xtrace_disable 00:21:48.427 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:21:48.427 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:21:48.427 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@41 -- # trid='trtype:VFIOUSER subnqn:nqn.2021-09.io.spdk:cnode0 traddr:/var/run/vfio-user' 00:21:48.427 17:37:30 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/fuzz/nvme_fuzz/nvme_fuzz -m 0x2 -t 30 -S 123456 -F 'trtype:VFIOUSER subnqn:nqn.2021-09.io.spdk:cnode0 traddr:/var/run/vfio-user' -N -a 00:22:20.504 Fuzzing completed. Shutting down the fuzz application 00:22:20.504 00:22:20.504 Dumping successful admin opcodes: 00:22:20.504 9, 10, 00:22:20.504 Dumping successful io opcodes: 00:22:20.504 0, 00:22:20.504 NS: 0x20000081ef00 I/O qp, Total commands completed: 664176, total successful commands: 2592, random_seed: 2783911488 00:22:20.504 NS: 0x20000081ef00 admin qp, Total commands completed: 85344, total successful commands: 20, random_seed: 1129269440 00:22:20.504 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@44 -- # rpc_cmd nvmf_delete_subsystem nqn.2021-09.io.spdk:cnode0 00:22:20.504 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:20.504 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:22:20.504 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:20.504 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@46 -- # killprocess 246302 00:22:20.504 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@954 -- # '[' -z 246302 ']' 00:22:20.504 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@958 -- # kill -0 246302 00:22:20.504 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@959 -- # uname 00:22:20.504 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:22:20.504 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 246302 00:22:20.504 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:22:20.504 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:22:20.504 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@972 -- # echo 'killing process with pid 246302' 00:22:20.504 killing process with pid 246302 00:22:20.504 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@973 -- # kill 246302 00:22:20.504 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@978 -- # wait 246302 00:22:20.504 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@48 -- # rm -rf /var/run/vfio-user /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/vfio_user_fuzz_log.txt /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/vfio_user_fuzz_tgt_output.txt 00:22:20.504 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@50 -- # trap - SIGINT SIGTERM EXIT 00:22:20.504 00:22:20.504 real 0m32.183s 00:22:20.504 user 0m30.386s 00:22:20.504 sys 0m29.219s 00:22:20.504 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1130 -- # xtrace_disable 00:22:20.504 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:22:20.504 ************************************ 00:22:20.504 END TEST nvmf_vfio_user_fuzz 00:22:20.504 ************************************ 00:22:20.504 17:38:00 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@37 -- # run_test nvmf_auth_target /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/auth.sh --transport=tcp 00:22:20.504 17:38:00 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:22:20.504 17:38:00 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:22:20.504 17:38:00 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:22:20.504 ************************************ 00:22:20.504 START TEST nvmf_auth_target 00:22:20.504 ************************************ 00:22:20.504 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/auth.sh --transport=tcp 00:22:20.504 * Looking for test storage... 00:22:20.504 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:22:20.504 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:22:20.504 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1711 -- # lcov --version 00:22:20.504 17:38:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:22:20.504 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:22:20.504 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:22:20.504 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@333 -- # local ver1 ver1_l 00:22:20.504 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@334 -- # local ver2 ver2_l 00:22:20.504 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@336 -- # IFS=.-: 00:22:20.504 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@336 -- # read -ra ver1 00:22:20.504 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@337 -- # IFS=.-: 00:22:20.504 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@337 -- # read -ra ver2 00:22:20.504 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@338 -- # local 'op=<' 00:22:20.504 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@340 -- # ver1_l=2 00:22:20.504 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@341 -- # ver2_l=1 00:22:20.504 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:22:20.504 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@344 -- # case "$op" in 00:22:20.504 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@345 -- # : 1 00:22:20.504 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@364 -- # (( v = 0 )) 00:22:20.504 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:22:20.504 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@365 -- # decimal 1 00:22:20.504 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@353 -- # local d=1 00:22:20.504 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:22:20.504 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@355 -- # echo 1 00:22:20.504 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@365 -- # ver1[v]=1 00:22:20.504 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@366 -- # decimal 2 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@353 -- # local d=2 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@355 -- # echo 2 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@366 -- # ver2[v]=2 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@368 -- # return 0 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:22:20.505 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:22:20.505 --rc genhtml_branch_coverage=1 00:22:20.505 --rc genhtml_function_coverage=1 00:22:20.505 --rc genhtml_legend=1 00:22:20.505 --rc geninfo_all_blocks=1 00:22:20.505 --rc geninfo_unexecuted_blocks=1 00:22:20.505 00:22:20.505 ' 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:22:20.505 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:22:20.505 --rc genhtml_branch_coverage=1 00:22:20.505 --rc genhtml_function_coverage=1 00:22:20.505 --rc genhtml_legend=1 00:22:20.505 --rc geninfo_all_blocks=1 00:22:20.505 --rc geninfo_unexecuted_blocks=1 00:22:20.505 00:22:20.505 ' 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:22:20.505 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:22:20.505 --rc genhtml_branch_coverage=1 00:22:20.505 --rc genhtml_function_coverage=1 00:22:20.505 --rc genhtml_legend=1 00:22:20.505 --rc geninfo_all_blocks=1 00:22:20.505 --rc geninfo_unexecuted_blocks=1 00:22:20.505 00:22:20.505 ' 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:22:20.505 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:22:20.505 --rc genhtml_branch_coverage=1 00:22:20.505 --rc genhtml_function_coverage=1 00:22:20.505 --rc genhtml_legend=1 00:22:20.505 --rc geninfo_all_blocks=1 00:22:20.505 --rc geninfo_unexecuted_blocks=1 00:22:20.505 00:22:20.505 ' 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@7 -- # uname -s 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@15 -- # shopt -s extglob 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- paths/export.sh@5 -- # export PATH 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@51 -- # : 0 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:22:20.505 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@55 -- # have_pci_nics=0 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@13 -- # digests=("sha256" "sha384" "sha512") 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@14 -- # dhgroups=("null" "ffdhe2048" "ffdhe3072" "ffdhe4096" "ffdhe6144" "ffdhe8192") 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@15 -- # subnqn=nqn.2024-03.io.spdk:cnode0 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@16 -- # hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@17 -- # hostsock=/var/tmp/host.sock 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@18 -- # keys=() 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@18 -- # ckeys=() 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@86 -- # nvmftestinit 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@476 -- # prepare_net_devs 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@438 -- # local -g is_hw=no 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@440 -- # remove_spdk_ns 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@309 -- # xtrace_disable 00:22:20.505 17:38:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@315 -- # pci_devs=() 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@315 -- # local -a pci_devs 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@316 -- # pci_net_devs=() 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@317 -- # pci_drivers=() 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@317 -- # local -A pci_drivers 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@319 -- # net_devs=() 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@319 -- # local -ga net_devs 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@320 -- # e810=() 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@320 -- # local -ga e810 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@321 -- # x722=() 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@321 -- # local -ga x722 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@322 -- # mlx=() 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@322 -- # local -ga mlx 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:22:21.440 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:22:21.440 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@418 -- # [[ up == up ]] 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:22:21.440 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:22:21.441 Found net devices under 0000:0a:00.0: cvl_0_0 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@418 -- # [[ up == up ]] 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:22:21.441 Found net devices under 0000:0a:00.1: cvl_0_1 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@442 -- # is_hw=yes 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:22:21.441 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:22:21.699 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:22:21.699 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.204 ms 00:22:21.699 00:22:21.699 --- 10.0.0.2 ping statistics --- 00:22:21.699 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:21.699 rtt min/avg/max/mdev = 0.204/0.204/0.204/0.000 ms 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:22:21.699 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:22:21.699 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.112 ms 00:22:21.699 00:22:21.699 --- 10.0.0.1 ping statistics --- 00:22:21.699 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:21.699 rtt min/avg/max/mdev = 0.112/0.112/0.112/0.000 ms 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@450 -- # return 0 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@87 -- # nvmfappstart -L nvmf_auth 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@726 -- # xtrace_disable 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@509 -- # nvmfpid=251695 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -L nvmf_auth 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@510 -- # waitforlisten 251695 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@835 -- # '[' -z 251695 ']' 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@840 -- # local max_retries=100 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@844 -- # xtrace_disable 00:22:21.699 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:21.957 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:22:21.957 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@868 -- # return 0 00:22:21.957 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:22:21.957 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@732 -- # xtrace_disable 00:22:21.957 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:21.957 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:22:21.957 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@89 -- # hostpid=251771 00:22:21.957 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 2 -r /var/tmp/host.sock -L nvme_auth 00:22:21.957 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@91 -- # trap 'dumplogs; cleanup' SIGINT SIGTERM EXIT 00:22:21.957 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@94 -- # gen_dhchap_key null 48 00:22:21.957 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@751 -- # local digest len file key 00:22:21.957 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:22:21.957 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # local -A digests 00:22:21.957 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # digest=null 00:22:21.957 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # len=48 00:22:21.957 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # xxd -p -c0 -l 24 /dev/urandom 00:22:21.957 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # key=d4c75eaf3141388101c890ad53986e42687cf004d6fc1bf7 00:22:21.957 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # mktemp -t spdk.key-null.XXX 00:22:21.957 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-null.b1A 00:22:21.957 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@757 -- # format_dhchap_key d4c75eaf3141388101c890ad53986e42687cf004d6fc1bf7 0 00:22:21.957 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@747 -- # format_key DHHC-1 d4c75eaf3141388101c890ad53986e42687cf004d6fc1bf7 0 00:22:21.957 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # local prefix key digest 00:22:21.957 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:22:21.958 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # key=d4c75eaf3141388101c890ad53986e42687cf004d6fc1bf7 00:22:21.958 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # digest=0 00:22:21.958 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@733 -- # python - 00:22:21.958 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-null.b1A 00:22:21.958 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-null.b1A 00:22:21.958 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@94 -- # keys[0]=/tmp/spdk.key-null.b1A 00:22:21.958 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@94 -- # gen_dhchap_key sha512 64 00:22:21.958 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@751 -- # local digest len file key 00:22:21.958 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:22:21.958 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # local -A digests 00:22:21.958 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # digest=sha512 00:22:21.958 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # len=64 00:22:21.958 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # xxd -p -c0 -l 32 /dev/urandom 00:22:21.958 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # key=159edd8a90737578ba2952877b6745f9bf5ad6efed9a7befae5b523f0c44db8e 00:22:21.958 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # mktemp -t spdk.key-sha512.XXX 00:22:21.958 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-sha512.7q2 00:22:21.958 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@757 -- # format_dhchap_key 159edd8a90737578ba2952877b6745f9bf5ad6efed9a7befae5b523f0c44db8e 3 00:22:21.958 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@747 -- # format_key DHHC-1 159edd8a90737578ba2952877b6745f9bf5ad6efed9a7befae5b523f0c44db8e 3 00:22:21.958 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # local prefix key digest 00:22:21.958 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:22:21.958 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # key=159edd8a90737578ba2952877b6745f9bf5ad6efed9a7befae5b523f0c44db8e 00:22:21.958 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # digest=3 00:22:21.958 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@733 -- # python - 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-sha512.7q2 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-sha512.7q2 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@94 -- # ckeys[0]=/tmp/spdk.key-sha512.7q2 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@95 -- # gen_dhchap_key sha256 32 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@751 -- # local digest len file key 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # local -A digests 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # digest=sha256 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # len=32 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # xxd -p -c0 -l 16 /dev/urandom 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # key=52bf13d6e1be82f057ac5cd225efc11a 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # mktemp -t spdk.key-sha256.XXX 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-sha256.zDo 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@757 -- # format_dhchap_key 52bf13d6e1be82f057ac5cd225efc11a 1 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@747 -- # format_key DHHC-1 52bf13d6e1be82f057ac5cd225efc11a 1 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # local prefix key digest 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # key=52bf13d6e1be82f057ac5cd225efc11a 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # digest=1 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@733 -- # python - 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-sha256.zDo 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-sha256.zDo 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@95 -- # keys[1]=/tmp/spdk.key-sha256.zDo 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@95 -- # gen_dhchap_key sha384 48 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@751 -- # local digest len file key 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # local -A digests 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # digest=sha384 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # len=48 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # xxd -p -c0 -l 24 /dev/urandom 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # key=5fbce36c396e7fa5f9e4f5f2db814eba1a20cb691a66eba9 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # mktemp -t spdk.key-sha384.XXX 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-sha384.OEt 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@757 -- # format_dhchap_key 5fbce36c396e7fa5f9e4f5f2db814eba1a20cb691a66eba9 2 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@747 -- # format_key DHHC-1 5fbce36c396e7fa5f9e4f5f2db814eba1a20cb691a66eba9 2 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # local prefix key digest 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # key=5fbce36c396e7fa5f9e4f5f2db814eba1a20cb691a66eba9 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # digest=2 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@733 -- # python - 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-sha384.OEt 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-sha384.OEt 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@95 -- # ckeys[1]=/tmp/spdk.key-sha384.OEt 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@96 -- # gen_dhchap_key sha384 48 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@751 -- # local digest len file key 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # local -A digests 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # digest=sha384 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # len=48 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # xxd -p -c0 -l 24 /dev/urandom 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # key=a809f2c0721e0767b95e717abb399206077bc39c8948777c 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # mktemp -t spdk.key-sha384.XXX 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-sha384.TnN 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@757 -- # format_dhchap_key a809f2c0721e0767b95e717abb399206077bc39c8948777c 2 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@747 -- # format_key DHHC-1 a809f2c0721e0767b95e717abb399206077bc39c8948777c 2 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # local prefix key digest 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # key=a809f2c0721e0767b95e717abb399206077bc39c8948777c 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # digest=2 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@733 -- # python - 00:22:22.217 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-sha384.TnN 00:22:22.218 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-sha384.TnN 00:22:22.218 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@96 -- # keys[2]=/tmp/spdk.key-sha384.TnN 00:22:22.218 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@96 -- # gen_dhchap_key sha256 32 00:22:22.218 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@751 -- # local digest len file key 00:22:22.218 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:22:22.218 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # local -A digests 00:22:22.218 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # digest=sha256 00:22:22.218 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # len=32 00:22:22.218 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # xxd -p -c0 -l 16 /dev/urandom 00:22:22.218 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # key=4d268dad32a7dbfbbbbb8b09c341249d 00:22:22.218 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # mktemp -t spdk.key-sha256.XXX 00:22:22.218 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-sha256.Avu 00:22:22.218 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@757 -- # format_dhchap_key 4d268dad32a7dbfbbbbb8b09c341249d 1 00:22:22.218 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@747 -- # format_key DHHC-1 4d268dad32a7dbfbbbbb8b09c341249d 1 00:22:22.218 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # local prefix key digest 00:22:22.218 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:22:22.218 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # key=4d268dad32a7dbfbbbbb8b09c341249d 00:22:22.218 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # digest=1 00:22:22.218 17:38:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@733 -- # python - 00:22:22.218 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-sha256.Avu 00:22:22.218 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-sha256.Avu 00:22:22.218 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@96 -- # ckeys[2]=/tmp/spdk.key-sha256.Avu 00:22:22.218 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@97 -- # gen_dhchap_key sha512 64 00:22:22.218 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@751 -- # local digest len file key 00:22:22.218 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:22:22.218 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # local -A digests 00:22:22.218 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # digest=sha512 00:22:22.218 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # len=64 00:22:22.218 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # xxd -p -c0 -l 32 /dev/urandom 00:22:22.218 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # key=6fe2d3a1e70600278839ec681b1a4f4cdb78a0c9ed931b31b79d86eccdf1e36b 00:22:22.218 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # mktemp -t spdk.key-sha512.XXX 00:22:22.218 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-sha512.BLn 00:22:22.218 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@757 -- # format_dhchap_key 6fe2d3a1e70600278839ec681b1a4f4cdb78a0c9ed931b31b79d86eccdf1e36b 3 00:22:22.218 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@747 -- # format_key DHHC-1 6fe2d3a1e70600278839ec681b1a4f4cdb78a0c9ed931b31b79d86eccdf1e36b 3 00:22:22.218 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # local prefix key digest 00:22:22.218 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:22:22.218 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # key=6fe2d3a1e70600278839ec681b1a4f4cdb78a0c9ed931b31b79d86eccdf1e36b 00:22:22.218 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@732 -- # digest=3 00:22:22.218 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@733 -- # python - 00:22:22.476 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-sha512.BLn 00:22:22.476 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-sha512.BLn 00:22:22.476 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@97 -- # keys[3]=/tmp/spdk.key-sha512.BLn 00:22:22.476 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@97 -- # ckeys[3]= 00:22:22.476 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@99 -- # waitforlisten 251695 00:22:22.476 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@835 -- # '[' -z 251695 ']' 00:22:22.476 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:22:22.476 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@840 -- # local max_retries=100 00:22:22.476 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:22:22.476 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:22:22.476 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@844 -- # xtrace_disable 00:22:22.476 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:22.734 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:22:22.734 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@868 -- # return 0 00:22:22.734 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@100 -- # waitforlisten 251771 /var/tmp/host.sock 00:22:22.734 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@835 -- # '[' -z 251771 ']' 00:22:22.734 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/host.sock 00:22:22.734 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@840 -- # local max_retries=100 00:22:22.734 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/host.sock...' 00:22:22.734 Waiting for process to start up and listen on UNIX domain socket /var/tmp/host.sock... 00:22:22.734 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@844 -- # xtrace_disable 00:22:22.734 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:22.992 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:22:22.992 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@868 -- # return 0 00:22:22.992 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@101 -- # rpc_cmd 00:22:22.992 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:22.992 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:22.992 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:22.992 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@108 -- # for i in "${!keys[@]}" 00:22:22.992 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@109 -- # rpc_cmd keyring_file_add_key key0 /tmp/spdk.key-null.b1A 00:22:22.992 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:22.992 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:22.992 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:22.992 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@110 -- # hostrpc keyring_file_add_key key0 /tmp/spdk.key-null.b1A 00:22:22.992 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key key0 /tmp/spdk.key-null.b1A 00:22:23.250 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@111 -- # [[ -n /tmp/spdk.key-sha512.7q2 ]] 00:22:23.250 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@112 -- # rpc_cmd keyring_file_add_key ckey0 /tmp/spdk.key-sha512.7q2 00:22:23.250 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:23.250 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:23.250 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:23.250 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@113 -- # hostrpc keyring_file_add_key ckey0 /tmp/spdk.key-sha512.7q2 00:22:23.250 17:38:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key ckey0 /tmp/spdk.key-sha512.7q2 00:22:23.508 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@108 -- # for i in "${!keys[@]}" 00:22:23.508 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@109 -- # rpc_cmd keyring_file_add_key key1 /tmp/spdk.key-sha256.zDo 00:22:23.508 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:23.508 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:23.508 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:23.508 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@110 -- # hostrpc keyring_file_add_key key1 /tmp/spdk.key-sha256.zDo 00:22:23.508 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key key1 /tmp/spdk.key-sha256.zDo 00:22:23.765 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@111 -- # [[ -n /tmp/spdk.key-sha384.OEt ]] 00:22:23.765 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@112 -- # rpc_cmd keyring_file_add_key ckey1 /tmp/spdk.key-sha384.OEt 00:22:23.765 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:23.765 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:23.765 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:23.765 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@113 -- # hostrpc keyring_file_add_key ckey1 /tmp/spdk.key-sha384.OEt 00:22:23.765 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key ckey1 /tmp/spdk.key-sha384.OEt 00:22:24.023 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@108 -- # for i in "${!keys[@]}" 00:22:24.023 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@109 -- # rpc_cmd keyring_file_add_key key2 /tmp/spdk.key-sha384.TnN 00:22:24.023 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:24.023 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:24.023 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:24.023 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@110 -- # hostrpc keyring_file_add_key key2 /tmp/spdk.key-sha384.TnN 00:22:24.023 17:38:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key key2 /tmp/spdk.key-sha384.TnN 00:22:24.280 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@111 -- # [[ -n /tmp/spdk.key-sha256.Avu ]] 00:22:24.280 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@112 -- # rpc_cmd keyring_file_add_key ckey2 /tmp/spdk.key-sha256.Avu 00:22:24.280 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:24.280 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:24.280 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:24.280 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@113 -- # hostrpc keyring_file_add_key ckey2 /tmp/spdk.key-sha256.Avu 00:22:24.280 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key ckey2 /tmp/spdk.key-sha256.Avu 00:22:24.538 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@108 -- # for i in "${!keys[@]}" 00:22:24.538 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@109 -- # rpc_cmd keyring_file_add_key key3 /tmp/spdk.key-sha512.BLn 00:22:24.538 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:24.538 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:24.538 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:24.538 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@110 -- # hostrpc keyring_file_add_key key3 /tmp/spdk.key-sha512.BLn 00:22:24.538 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key key3 /tmp/spdk.key-sha512.BLn 00:22:24.796 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@111 -- # [[ -n '' ]] 00:22:24.796 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@118 -- # for digest in "${digests[@]}" 00:22:24.796 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:22:24.796 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:22:24.796 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:22:24.796 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:22:25.055 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 null 0 00:22:25.055 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:22:25.055 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:22:25.055 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:22:25.055 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:22:25.055 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:22:25.055 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:22:25.055 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:25.055 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:25.055 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:25.055 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:22:25.055 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:22:25.055 17:38:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:22:25.621 00:22:25.621 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:22:25.621 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:25.621 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:22:25.621 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:25.621 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:22:25.621 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:25.621 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:25.621 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:25.621 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:22:25.621 { 00:22:25.621 "cntlid": 1, 00:22:25.621 "qid": 0, 00:22:25.621 "state": "enabled", 00:22:25.621 "thread": "nvmf_tgt_poll_group_000", 00:22:25.621 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:22:25.621 "listen_address": { 00:22:25.621 "trtype": "TCP", 00:22:25.621 "adrfam": "IPv4", 00:22:25.621 "traddr": "10.0.0.2", 00:22:25.621 "trsvcid": "4420" 00:22:25.621 }, 00:22:25.621 "peer_address": { 00:22:25.621 "trtype": "TCP", 00:22:25.621 "adrfam": "IPv4", 00:22:25.621 "traddr": "10.0.0.1", 00:22:25.621 "trsvcid": "47386" 00:22:25.621 }, 00:22:25.621 "auth": { 00:22:25.621 "state": "completed", 00:22:25.621 "digest": "sha256", 00:22:25.621 "dhgroup": "null" 00:22:25.621 } 00:22:25.621 } 00:22:25.621 ]' 00:22:25.621 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:22:25.879 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:22:25.879 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:22:25.879 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:22:25.879 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:22:25.879 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:22:25.879 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:25.879 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:26.136 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:22:26.136 17:38:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:22:31.398 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:22:31.398 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:22:31.398 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:22:31.398 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:31.398 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:31.398 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:31.398 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:22:31.398 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:22:31.398 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:22:31.398 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 null 1 00:22:31.398 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:22:31.398 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:22:31.398 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:22:31.398 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:22:31.398 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:22:31.398 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:31.398 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:31.398 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:31.398 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:31.398 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:31.398 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:31.398 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:31.398 00:22:31.398 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:22:31.398 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:22:31.398 17:38:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:31.398 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:31.398 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:22:31.398 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:31.398 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:31.656 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:31.656 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:22:31.656 { 00:22:31.656 "cntlid": 3, 00:22:31.656 "qid": 0, 00:22:31.656 "state": "enabled", 00:22:31.656 "thread": "nvmf_tgt_poll_group_000", 00:22:31.656 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:22:31.656 "listen_address": { 00:22:31.656 "trtype": "TCP", 00:22:31.656 "adrfam": "IPv4", 00:22:31.656 "traddr": "10.0.0.2", 00:22:31.656 "trsvcid": "4420" 00:22:31.656 }, 00:22:31.656 "peer_address": { 00:22:31.656 "trtype": "TCP", 00:22:31.656 "adrfam": "IPv4", 00:22:31.656 "traddr": "10.0.0.1", 00:22:31.656 "trsvcid": "47416" 00:22:31.656 }, 00:22:31.656 "auth": { 00:22:31.656 "state": "completed", 00:22:31.656 "digest": "sha256", 00:22:31.656 "dhgroup": "null" 00:22:31.656 } 00:22:31.656 } 00:22:31.656 ]' 00:22:31.656 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:22:31.656 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:22:31.656 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:22:31.656 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:22:31.656 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:22:31.656 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:22:31.656 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:31.656 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:31.914 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:22:31.914 17:38:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:22:32.847 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:22:32.847 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:22:32.847 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:22:32.847 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:32.847 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:32.847 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:32.847 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:22:32.847 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:22:32.847 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:22:33.105 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 null 2 00:22:33.105 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:22:33.105 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:22:33.105 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:22:33.105 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:22:33.105 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:22:33.105 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:22:33.105 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:33.105 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:33.105 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:33.105 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:22:33.105 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:22:33.105 17:38:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:22:33.362 00:22:33.362 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:22:33.362 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:33.362 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:22:33.618 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:33.618 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:22:33.618 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:33.618 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:33.618 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:33.618 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:22:33.618 { 00:22:33.618 "cntlid": 5, 00:22:33.618 "qid": 0, 00:22:33.618 "state": "enabled", 00:22:33.618 "thread": "nvmf_tgt_poll_group_000", 00:22:33.618 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:22:33.618 "listen_address": { 00:22:33.618 "trtype": "TCP", 00:22:33.618 "adrfam": "IPv4", 00:22:33.618 "traddr": "10.0.0.2", 00:22:33.618 "trsvcid": "4420" 00:22:33.618 }, 00:22:33.618 "peer_address": { 00:22:33.618 "trtype": "TCP", 00:22:33.618 "adrfam": "IPv4", 00:22:33.618 "traddr": "10.0.0.1", 00:22:33.618 "trsvcid": "51902" 00:22:33.619 }, 00:22:33.619 "auth": { 00:22:33.619 "state": "completed", 00:22:33.619 "digest": "sha256", 00:22:33.619 "dhgroup": "null" 00:22:33.619 } 00:22:33.619 } 00:22:33.619 ]' 00:22:33.619 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:22:33.875 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:22:33.875 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:22:33.875 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:22:33.875 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:22:33.875 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:22:33.875 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:33.875 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:34.131 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:22:34.131 17:38:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:22:35.060 17:38:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:22:35.060 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:22:35.060 17:38:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:22:35.060 17:38:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:35.060 17:38:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:35.060 17:38:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:35.060 17:38:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:22:35.060 17:38:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:22:35.060 17:38:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:22:35.317 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 null 3 00:22:35.317 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:22:35.317 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:22:35.317 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:22:35.317 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:22:35.317 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:22:35.317 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key3 00:22:35.317 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:35.317 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:35.317 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:35.317 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:22:35.317 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:22:35.317 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:22:35.574 00:22:35.574 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:22:35.574 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:22:35.574 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:35.831 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:35.831 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:22:35.831 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:35.831 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:35.831 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:35.831 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:22:35.831 { 00:22:35.831 "cntlid": 7, 00:22:35.831 "qid": 0, 00:22:35.831 "state": "enabled", 00:22:35.831 "thread": "nvmf_tgt_poll_group_000", 00:22:35.831 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:22:35.831 "listen_address": { 00:22:35.831 "trtype": "TCP", 00:22:35.831 "adrfam": "IPv4", 00:22:35.831 "traddr": "10.0.0.2", 00:22:35.831 "trsvcid": "4420" 00:22:35.831 }, 00:22:35.831 "peer_address": { 00:22:35.831 "trtype": "TCP", 00:22:35.831 "adrfam": "IPv4", 00:22:35.831 "traddr": "10.0.0.1", 00:22:35.831 "trsvcid": "51922" 00:22:35.831 }, 00:22:35.831 "auth": { 00:22:35.831 "state": "completed", 00:22:35.831 "digest": "sha256", 00:22:35.831 "dhgroup": "null" 00:22:35.831 } 00:22:35.831 } 00:22:35.831 ]' 00:22:35.831 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:22:35.831 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:22:35.831 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:22:36.088 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:22:36.088 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:22:36.088 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:22:36.088 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:36.088 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:36.345 17:38:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:22:36.345 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:22:37.276 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:22:37.276 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:22:37.276 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:22:37.276 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:37.276 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:37.276 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:37.276 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:22:37.276 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:22:37.276 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:22:37.276 17:38:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:22:37.533 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe2048 0 00:22:37.533 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:22:37.533 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:22:37.533 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:22:37.533 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:22:37.533 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:22:37.533 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:22:37.533 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:37.533 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:37.533 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:37.533 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:22:37.533 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:22:37.533 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:22:37.791 00:22:37.791 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:22:37.791 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:22:37.791 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:38.048 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:38.048 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:22:38.048 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:38.048 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:38.048 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:38.048 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:22:38.048 { 00:22:38.048 "cntlid": 9, 00:22:38.048 "qid": 0, 00:22:38.048 "state": "enabled", 00:22:38.048 "thread": "nvmf_tgt_poll_group_000", 00:22:38.048 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:22:38.048 "listen_address": { 00:22:38.048 "trtype": "TCP", 00:22:38.048 "adrfam": "IPv4", 00:22:38.048 "traddr": "10.0.0.2", 00:22:38.048 "trsvcid": "4420" 00:22:38.048 }, 00:22:38.048 "peer_address": { 00:22:38.048 "trtype": "TCP", 00:22:38.048 "adrfam": "IPv4", 00:22:38.048 "traddr": "10.0.0.1", 00:22:38.048 "trsvcid": "51942" 00:22:38.048 }, 00:22:38.048 "auth": { 00:22:38.048 "state": "completed", 00:22:38.048 "digest": "sha256", 00:22:38.048 "dhgroup": "ffdhe2048" 00:22:38.048 } 00:22:38.048 } 00:22:38.048 ]' 00:22:38.048 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:22:38.048 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:22:38.048 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:22:38.307 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:22:38.307 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:22:38.307 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:22:38.307 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:38.307 17:38:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:38.565 17:38:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:22:38.565 17:38:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:22:39.498 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:22:39.498 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:22:39.498 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:22:39.498 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:39.498 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:39.498 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:39.498 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:22:39.498 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:22:39.498 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:22:39.756 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe2048 1 00:22:39.756 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:22:39.756 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:22:39.756 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:22:39.756 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:22:39.756 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:22:39.756 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:39.756 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:39.756 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:39.756 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:39.756 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:39.756 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:39.756 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:40.014 00:22:40.014 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:22:40.014 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:22:40.014 17:38:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:40.272 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:40.272 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:22:40.272 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:40.272 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:40.272 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:40.272 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:22:40.272 { 00:22:40.272 "cntlid": 11, 00:22:40.272 "qid": 0, 00:22:40.272 "state": "enabled", 00:22:40.272 "thread": "nvmf_tgt_poll_group_000", 00:22:40.272 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:22:40.272 "listen_address": { 00:22:40.272 "trtype": "TCP", 00:22:40.272 "adrfam": "IPv4", 00:22:40.272 "traddr": "10.0.0.2", 00:22:40.272 "trsvcid": "4420" 00:22:40.272 }, 00:22:40.272 "peer_address": { 00:22:40.272 "trtype": "TCP", 00:22:40.272 "adrfam": "IPv4", 00:22:40.272 "traddr": "10.0.0.1", 00:22:40.272 "trsvcid": "51978" 00:22:40.272 }, 00:22:40.272 "auth": { 00:22:40.272 "state": "completed", 00:22:40.272 "digest": "sha256", 00:22:40.272 "dhgroup": "ffdhe2048" 00:22:40.272 } 00:22:40.272 } 00:22:40.272 ]' 00:22:40.272 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:22:40.272 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:22:40.272 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:22:40.272 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:22:40.272 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:22:40.530 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:22:40.530 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:40.530 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:40.788 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:22:40.788 17:38:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:22:41.722 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:22:41.722 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:22:41.722 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:22:41.722 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:41.722 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:41.722 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:41.722 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:22:41.722 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:22:41.722 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:22:41.722 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe2048 2 00:22:41.722 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:22:41.722 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:22:41.722 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:22:41.722 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:22:41.722 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:22:41.722 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:22:41.722 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:41.722 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:41.722 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:41.722 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:22:41.722 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:22:41.980 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:22:42.238 00:22:42.238 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:22:42.238 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:22:42.238 17:38:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:42.495 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:42.495 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:22:42.495 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:42.495 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:42.495 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:42.495 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:22:42.495 { 00:22:42.495 "cntlid": 13, 00:22:42.495 "qid": 0, 00:22:42.495 "state": "enabled", 00:22:42.495 "thread": "nvmf_tgt_poll_group_000", 00:22:42.495 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:22:42.495 "listen_address": { 00:22:42.495 "trtype": "TCP", 00:22:42.495 "adrfam": "IPv4", 00:22:42.495 "traddr": "10.0.0.2", 00:22:42.495 "trsvcid": "4420" 00:22:42.495 }, 00:22:42.495 "peer_address": { 00:22:42.495 "trtype": "TCP", 00:22:42.495 "adrfam": "IPv4", 00:22:42.495 "traddr": "10.0.0.1", 00:22:42.495 "trsvcid": "52000" 00:22:42.495 }, 00:22:42.495 "auth": { 00:22:42.495 "state": "completed", 00:22:42.495 "digest": "sha256", 00:22:42.495 "dhgroup": "ffdhe2048" 00:22:42.495 } 00:22:42.495 } 00:22:42.496 ]' 00:22:42.496 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:22:42.496 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:22:42.496 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:22:42.496 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:22:42.496 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:22:42.496 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:22:42.496 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:42.496 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:42.753 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:22:42.753 17:38:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:22:43.687 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:22:43.687 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:22:43.687 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:22:43.687 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:43.687 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:43.687 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:43.687 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:22:43.687 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:22:43.687 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:22:43.944 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe2048 3 00:22:43.944 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:22:43.944 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:22:43.944 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:22:43.944 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:22:43.944 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:22:43.944 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key3 00:22:43.944 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:43.944 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:43.944 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:43.944 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:22:43.944 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:22:43.944 17:38:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:22:44.508 00:22:44.508 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:22:44.508 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:22:44.508 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:44.508 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:44.508 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:22:44.508 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:44.508 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:44.508 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:44.509 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:22:44.509 { 00:22:44.509 "cntlid": 15, 00:22:44.509 "qid": 0, 00:22:44.509 "state": "enabled", 00:22:44.509 "thread": "nvmf_tgt_poll_group_000", 00:22:44.509 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:22:44.509 "listen_address": { 00:22:44.509 "trtype": "TCP", 00:22:44.509 "adrfam": "IPv4", 00:22:44.509 "traddr": "10.0.0.2", 00:22:44.509 "trsvcid": "4420" 00:22:44.509 }, 00:22:44.509 "peer_address": { 00:22:44.509 "trtype": "TCP", 00:22:44.509 "adrfam": "IPv4", 00:22:44.509 "traddr": "10.0.0.1", 00:22:44.509 "trsvcid": "33184" 00:22:44.509 }, 00:22:44.509 "auth": { 00:22:44.509 "state": "completed", 00:22:44.509 "digest": "sha256", 00:22:44.509 "dhgroup": "ffdhe2048" 00:22:44.509 } 00:22:44.509 } 00:22:44.509 ]' 00:22:44.509 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:22:44.766 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:22:44.766 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:22:44.766 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:22:44.766 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:22:44.766 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:22:44.766 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:44.766 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:45.024 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:22:45.024 17:38:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:22:45.958 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:22:45.958 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:22:45.958 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:22:45.958 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:45.958 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:45.958 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:45.958 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:22:45.958 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:22:45.958 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:22:45.958 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:22:46.216 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe3072 0 00:22:46.216 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:22:46.216 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:22:46.216 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:22:46.216 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:22:46.216 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:22:46.216 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:22:46.216 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:46.216 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:46.216 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:46.216 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:22:46.216 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:22:46.216 17:38:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:22:46.474 00:22:46.474 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:22:46.474 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:22:46.474 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:46.732 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:46.732 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:22:46.732 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:46.732 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:46.732 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:46.732 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:22:46.732 { 00:22:46.732 "cntlid": 17, 00:22:46.732 "qid": 0, 00:22:46.732 "state": "enabled", 00:22:46.732 "thread": "nvmf_tgt_poll_group_000", 00:22:46.732 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:22:46.732 "listen_address": { 00:22:46.732 "trtype": "TCP", 00:22:46.732 "adrfam": "IPv4", 00:22:46.732 "traddr": "10.0.0.2", 00:22:46.732 "trsvcid": "4420" 00:22:46.732 }, 00:22:46.732 "peer_address": { 00:22:46.732 "trtype": "TCP", 00:22:46.732 "adrfam": "IPv4", 00:22:46.732 "traddr": "10.0.0.1", 00:22:46.732 "trsvcid": "33198" 00:22:46.732 }, 00:22:46.732 "auth": { 00:22:46.732 "state": "completed", 00:22:46.732 "digest": "sha256", 00:22:46.732 "dhgroup": "ffdhe3072" 00:22:46.732 } 00:22:46.732 } 00:22:46.732 ]' 00:22:46.732 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:22:46.990 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:22:46.990 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:22:46.990 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:22:46.990 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:22:46.990 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:22:46.990 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:46.990 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:47.248 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:22:47.248 17:38:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:22:48.182 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:22:48.182 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:22:48.182 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:22:48.182 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:48.182 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:48.182 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:48.182 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:22:48.182 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:22:48.182 17:38:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:22:48.440 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe3072 1 00:22:48.440 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:22:48.440 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:22:48.440 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:22:48.440 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:22:48.440 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:22:48.440 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:48.440 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:48.440 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:48.440 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:48.440 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:48.440 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:48.440 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:48.697 00:22:48.697 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:22:48.697 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:22:48.697 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:48.955 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:48.955 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:22:48.955 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:48.955 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:49.213 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:49.213 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:22:49.213 { 00:22:49.213 "cntlid": 19, 00:22:49.213 "qid": 0, 00:22:49.213 "state": "enabled", 00:22:49.213 "thread": "nvmf_tgt_poll_group_000", 00:22:49.213 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:22:49.213 "listen_address": { 00:22:49.213 "trtype": "TCP", 00:22:49.213 "adrfam": "IPv4", 00:22:49.213 "traddr": "10.0.0.2", 00:22:49.213 "trsvcid": "4420" 00:22:49.213 }, 00:22:49.213 "peer_address": { 00:22:49.213 "trtype": "TCP", 00:22:49.213 "adrfam": "IPv4", 00:22:49.213 "traddr": "10.0.0.1", 00:22:49.213 "trsvcid": "33216" 00:22:49.213 }, 00:22:49.213 "auth": { 00:22:49.213 "state": "completed", 00:22:49.213 "digest": "sha256", 00:22:49.213 "dhgroup": "ffdhe3072" 00:22:49.213 } 00:22:49.213 } 00:22:49.213 ]' 00:22:49.213 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:22:49.213 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:22:49.213 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:22:49.213 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:22:49.213 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:22:49.213 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:22:49.213 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:49.213 17:38:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:49.470 17:38:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:22:49.470 17:38:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:22:50.403 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:22:50.403 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:22:50.403 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:22:50.403 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:50.403 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:50.403 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:50.403 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:22:50.403 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:22:50.403 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:22:50.662 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe3072 2 00:22:50.662 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:22:50.662 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:22:50.662 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:22:50.662 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:22:50.662 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:22:50.662 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:22:50.662 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:50.662 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:50.662 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:50.662 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:22:50.662 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:22:50.662 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:22:50.920 00:22:50.920 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:22:50.920 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:50.920 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:22:51.178 17:38:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:51.178 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:22:51.178 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:51.178 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:51.178 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:51.178 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:22:51.178 { 00:22:51.178 "cntlid": 21, 00:22:51.178 "qid": 0, 00:22:51.178 "state": "enabled", 00:22:51.178 "thread": "nvmf_tgt_poll_group_000", 00:22:51.178 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:22:51.178 "listen_address": { 00:22:51.178 "trtype": "TCP", 00:22:51.178 "adrfam": "IPv4", 00:22:51.178 "traddr": "10.0.0.2", 00:22:51.178 "trsvcid": "4420" 00:22:51.178 }, 00:22:51.178 "peer_address": { 00:22:51.178 "trtype": "TCP", 00:22:51.178 "adrfam": "IPv4", 00:22:51.178 "traddr": "10.0.0.1", 00:22:51.178 "trsvcid": "33226" 00:22:51.178 }, 00:22:51.178 "auth": { 00:22:51.178 "state": "completed", 00:22:51.178 "digest": "sha256", 00:22:51.178 "dhgroup": "ffdhe3072" 00:22:51.178 } 00:22:51.178 } 00:22:51.178 ]' 00:22:51.178 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:22:51.435 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:22:51.435 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:22:51.435 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:22:51.435 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:22:51.435 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:22:51.435 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:51.435 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:51.692 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:22:51.692 17:38:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:22:52.640 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:22:52.640 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:22:52.640 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:22:52.640 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:52.640 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:52.640 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:52.640 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:22:52.640 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:22:52.640 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:22:52.897 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe3072 3 00:22:52.897 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:22:52.897 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:22:52.897 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:22:52.897 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:22:52.897 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:22:52.897 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key3 00:22:52.897 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:52.897 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:52.897 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:52.897 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:22:52.897 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:22:52.897 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:22:53.155 00:22:53.155 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:22:53.155 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:22:53.155 17:38:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:53.411 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:53.411 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:22:53.411 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:53.411 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:53.411 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:53.411 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:22:53.411 { 00:22:53.411 "cntlid": 23, 00:22:53.411 "qid": 0, 00:22:53.411 "state": "enabled", 00:22:53.411 "thread": "nvmf_tgt_poll_group_000", 00:22:53.411 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:22:53.411 "listen_address": { 00:22:53.411 "trtype": "TCP", 00:22:53.411 "adrfam": "IPv4", 00:22:53.411 "traddr": "10.0.0.2", 00:22:53.411 "trsvcid": "4420" 00:22:53.411 }, 00:22:53.411 "peer_address": { 00:22:53.411 "trtype": "TCP", 00:22:53.411 "adrfam": "IPv4", 00:22:53.411 "traddr": "10.0.0.1", 00:22:53.411 "trsvcid": "33252" 00:22:53.411 }, 00:22:53.411 "auth": { 00:22:53.411 "state": "completed", 00:22:53.411 "digest": "sha256", 00:22:53.411 "dhgroup": "ffdhe3072" 00:22:53.411 } 00:22:53.411 } 00:22:53.411 ]' 00:22:53.411 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:22:53.411 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:22:53.411 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:22:53.668 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:22:53.668 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:22:53.668 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:22:53.668 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:53.668 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:53.925 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:22:53.925 17:38:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:22:54.858 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:22:54.858 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:22:54.858 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:22:54.858 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:54.858 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:54.858 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:54.858 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:22:54.858 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:22:54.858 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:22:54.858 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:22:55.115 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe4096 0 00:22:55.115 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:22:55.116 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:22:55.116 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:22:55.116 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:22:55.116 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:22:55.116 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:22:55.116 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:55.116 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:55.116 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:55.116 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:22:55.116 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:22:55.116 17:38:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:22:55.374 00:22:55.374 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:22:55.374 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:22:55.374 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:55.939 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:55.939 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:22:55.939 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:55.939 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:55.939 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:55.939 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:22:55.939 { 00:22:55.939 "cntlid": 25, 00:22:55.939 "qid": 0, 00:22:55.939 "state": "enabled", 00:22:55.939 "thread": "nvmf_tgt_poll_group_000", 00:22:55.939 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:22:55.939 "listen_address": { 00:22:55.939 "trtype": "TCP", 00:22:55.939 "adrfam": "IPv4", 00:22:55.939 "traddr": "10.0.0.2", 00:22:55.939 "trsvcid": "4420" 00:22:55.939 }, 00:22:55.939 "peer_address": { 00:22:55.939 "trtype": "TCP", 00:22:55.939 "adrfam": "IPv4", 00:22:55.939 "traddr": "10.0.0.1", 00:22:55.939 "trsvcid": "38810" 00:22:55.939 }, 00:22:55.939 "auth": { 00:22:55.939 "state": "completed", 00:22:55.939 "digest": "sha256", 00:22:55.939 "dhgroup": "ffdhe4096" 00:22:55.939 } 00:22:55.939 } 00:22:55.939 ]' 00:22:55.939 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:22:55.939 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:22:55.939 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:22:55.940 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:22:55.940 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:22:55.940 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:22:55.940 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:55.940 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:56.197 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:22:56.197 17:38:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:22:57.130 17:38:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:22:57.130 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:22:57.130 17:38:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:22:57.130 17:38:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:57.130 17:38:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:57.130 17:38:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:57.130 17:38:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:22:57.130 17:38:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:22:57.130 17:38:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:22:57.388 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe4096 1 00:22:57.388 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:22:57.388 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:22:57.388 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:22:57.388 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:22:57.388 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:22:57.388 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:57.388 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:57.388 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:57.388 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:57.388 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:57.388 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:57.388 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:57.646 00:22:57.646 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:22:57.646 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:22:57.646 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:57.903 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:57.903 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:22:57.903 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:57.903 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:57.903 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:57.903 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:22:57.903 { 00:22:57.903 "cntlid": 27, 00:22:57.903 "qid": 0, 00:22:57.903 "state": "enabled", 00:22:57.903 "thread": "nvmf_tgt_poll_group_000", 00:22:57.903 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:22:57.903 "listen_address": { 00:22:57.903 "trtype": "TCP", 00:22:57.903 "adrfam": "IPv4", 00:22:57.903 "traddr": "10.0.0.2", 00:22:57.903 "trsvcid": "4420" 00:22:57.903 }, 00:22:57.903 "peer_address": { 00:22:57.903 "trtype": "TCP", 00:22:57.903 "adrfam": "IPv4", 00:22:57.903 "traddr": "10.0.0.1", 00:22:57.903 "trsvcid": "38840" 00:22:57.903 }, 00:22:57.903 "auth": { 00:22:57.903 "state": "completed", 00:22:57.903 "digest": "sha256", 00:22:57.903 "dhgroup": "ffdhe4096" 00:22:57.903 } 00:22:57.903 } 00:22:57.903 ]' 00:22:57.903 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:22:57.903 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:22:58.161 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:22:58.161 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:22:58.161 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:22:58.161 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:22:58.161 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:58.161 17:38:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:58.419 17:38:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:22:58.419 17:38:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:22:59.352 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:22:59.353 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:22:59.353 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:22:59.353 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:59.353 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:59.353 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:59.353 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:22:59.353 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:22:59.353 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:22:59.611 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe4096 2 00:22:59.611 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:22:59.611 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:22:59.611 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:22:59.611 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:22:59.611 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:22:59.611 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:22:59.611 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:22:59.611 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:59.611 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:22:59.611 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:22:59.611 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:22:59.611 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:22:59.869 00:23:00.128 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:23:00.128 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:23:00.128 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:23:00.386 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:00.386 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:23:00.386 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:00.386 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:00.386 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:00.386 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:23:00.386 { 00:23:00.386 "cntlid": 29, 00:23:00.386 "qid": 0, 00:23:00.386 "state": "enabled", 00:23:00.386 "thread": "nvmf_tgt_poll_group_000", 00:23:00.387 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:23:00.387 "listen_address": { 00:23:00.387 "trtype": "TCP", 00:23:00.387 "adrfam": "IPv4", 00:23:00.387 "traddr": "10.0.0.2", 00:23:00.387 "trsvcid": "4420" 00:23:00.387 }, 00:23:00.387 "peer_address": { 00:23:00.387 "trtype": "TCP", 00:23:00.387 "adrfam": "IPv4", 00:23:00.387 "traddr": "10.0.0.1", 00:23:00.387 "trsvcid": "38866" 00:23:00.387 }, 00:23:00.387 "auth": { 00:23:00.387 "state": "completed", 00:23:00.387 "digest": "sha256", 00:23:00.387 "dhgroup": "ffdhe4096" 00:23:00.387 } 00:23:00.387 } 00:23:00.387 ]' 00:23:00.387 17:38:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:23:00.387 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:23:00.387 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:23:00.387 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:23:00.387 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:23:00.387 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:23:00.387 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:23:00.387 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:23:00.645 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:23:00.645 17:38:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:23:01.578 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:23:01.578 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:23:01.578 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:23:01.578 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:01.578 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:01.578 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:01.578 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:23:01.578 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:23:01.578 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:23:01.837 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe4096 3 00:23:01.837 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:23:01.837 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:23:01.837 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:23:01.837 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:23:01.837 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:23:01.837 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key3 00:23:01.837 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:01.837 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:01.837 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:01.837 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:23:01.837 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:23:01.837 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:23:02.403 00:23:02.403 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:23:02.403 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:23:02.403 17:38:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:23:02.403 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:02.403 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:23:02.403 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:02.403 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:02.403 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:02.403 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:23:02.404 { 00:23:02.404 "cntlid": 31, 00:23:02.404 "qid": 0, 00:23:02.404 "state": "enabled", 00:23:02.404 "thread": "nvmf_tgt_poll_group_000", 00:23:02.404 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:23:02.404 "listen_address": { 00:23:02.404 "trtype": "TCP", 00:23:02.404 "adrfam": "IPv4", 00:23:02.404 "traddr": "10.0.0.2", 00:23:02.404 "trsvcid": "4420" 00:23:02.404 }, 00:23:02.404 "peer_address": { 00:23:02.404 "trtype": "TCP", 00:23:02.404 "adrfam": "IPv4", 00:23:02.404 "traddr": "10.0.0.1", 00:23:02.404 "trsvcid": "38890" 00:23:02.404 }, 00:23:02.404 "auth": { 00:23:02.404 "state": "completed", 00:23:02.404 "digest": "sha256", 00:23:02.404 "dhgroup": "ffdhe4096" 00:23:02.404 } 00:23:02.404 } 00:23:02.404 ]' 00:23:02.404 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:23:02.662 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:23:02.662 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:23:02.662 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:23:02.662 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:23:02.662 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:23:02.662 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:23:02.662 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:23:02.921 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:23:02.921 17:38:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:23:03.853 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:23:03.853 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:23:03.853 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:23:03.853 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:03.853 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:03.853 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:03.853 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:23:03.853 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:23:03.853 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:23:03.853 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:23:04.111 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe6144 0 00:23:04.111 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:23:04.111 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:23:04.111 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:23:04.111 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:23:04.111 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:23:04.111 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:23:04.111 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:04.111 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:04.111 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:04.111 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:23:04.111 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:23:04.111 17:38:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:23:04.677 00:23:04.677 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:23:04.677 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:23:04.677 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:23:04.934 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:04.934 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:23:04.934 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:04.934 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:04.934 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:04.934 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:23:04.934 { 00:23:04.934 "cntlid": 33, 00:23:04.934 "qid": 0, 00:23:04.934 "state": "enabled", 00:23:04.934 "thread": "nvmf_tgt_poll_group_000", 00:23:04.934 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:23:04.934 "listen_address": { 00:23:04.934 "trtype": "TCP", 00:23:04.934 "adrfam": "IPv4", 00:23:04.934 "traddr": "10.0.0.2", 00:23:04.934 "trsvcid": "4420" 00:23:04.934 }, 00:23:04.934 "peer_address": { 00:23:04.934 "trtype": "TCP", 00:23:04.934 "adrfam": "IPv4", 00:23:04.934 "traddr": "10.0.0.1", 00:23:04.934 "trsvcid": "56708" 00:23:04.934 }, 00:23:04.934 "auth": { 00:23:04.934 "state": "completed", 00:23:04.934 "digest": "sha256", 00:23:04.934 "dhgroup": "ffdhe6144" 00:23:04.934 } 00:23:04.934 } 00:23:04.934 ]' 00:23:04.934 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:23:04.934 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:23:04.934 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:23:04.934 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:23:04.934 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:23:04.934 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:23:04.934 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:23:04.934 17:38:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:23:05.499 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:23:05.499 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:23:06.063 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:23:06.063 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:23:06.063 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:23:06.064 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:06.064 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:06.321 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:06.321 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:23:06.321 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:23:06.321 17:38:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:23:06.578 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe6144 1 00:23:06.578 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:23:06.578 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:23:06.578 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:23:06.578 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:23:06.578 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:23:06.578 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:23:06.578 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:06.578 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:06.578 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:06.578 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:23:06.578 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:23:06.578 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:23:07.148 00:23:07.148 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:23:07.148 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:23:07.148 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:23:07.406 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:07.406 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:23:07.406 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:07.406 17:38:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:07.406 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:07.406 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:23:07.406 { 00:23:07.406 "cntlid": 35, 00:23:07.406 "qid": 0, 00:23:07.406 "state": "enabled", 00:23:07.406 "thread": "nvmf_tgt_poll_group_000", 00:23:07.406 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:23:07.406 "listen_address": { 00:23:07.406 "trtype": "TCP", 00:23:07.406 "adrfam": "IPv4", 00:23:07.406 "traddr": "10.0.0.2", 00:23:07.406 "trsvcid": "4420" 00:23:07.406 }, 00:23:07.406 "peer_address": { 00:23:07.406 "trtype": "TCP", 00:23:07.406 "adrfam": "IPv4", 00:23:07.406 "traddr": "10.0.0.1", 00:23:07.406 "trsvcid": "56726" 00:23:07.406 }, 00:23:07.406 "auth": { 00:23:07.406 "state": "completed", 00:23:07.406 "digest": "sha256", 00:23:07.406 "dhgroup": "ffdhe6144" 00:23:07.406 } 00:23:07.406 } 00:23:07.406 ]' 00:23:07.406 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:23:07.406 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:23:07.406 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:23:07.406 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:23:07.406 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:23:07.406 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:23:07.406 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:23:07.406 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:23:07.664 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:23:07.664 17:38:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:23:08.596 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:23:08.596 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:23:08.596 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:23:08.596 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:08.596 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:08.596 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:08.596 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:23:08.596 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:23:08.596 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:23:08.854 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe6144 2 00:23:08.854 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:23:08.854 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:23:08.854 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:23:08.854 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:23:08.854 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:23:08.854 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:23:08.854 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:08.854 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:08.854 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:08.854 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:23:08.854 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:23:08.854 17:38:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:23:09.420 00:23:09.420 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:23:09.420 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:23:09.420 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:23:09.678 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:09.678 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:23:09.678 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:09.678 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:09.678 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:09.678 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:23:09.678 { 00:23:09.678 "cntlid": 37, 00:23:09.678 "qid": 0, 00:23:09.678 "state": "enabled", 00:23:09.678 "thread": "nvmf_tgt_poll_group_000", 00:23:09.678 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:23:09.678 "listen_address": { 00:23:09.678 "trtype": "TCP", 00:23:09.678 "adrfam": "IPv4", 00:23:09.678 "traddr": "10.0.0.2", 00:23:09.678 "trsvcid": "4420" 00:23:09.678 }, 00:23:09.678 "peer_address": { 00:23:09.678 "trtype": "TCP", 00:23:09.678 "adrfam": "IPv4", 00:23:09.678 "traddr": "10.0.0.1", 00:23:09.678 "trsvcid": "56748" 00:23:09.678 }, 00:23:09.678 "auth": { 00:23:09.678 "state": "completed", 00:23:09.678 "digest": "sha256", 00:23:09.678 "dhgroup": "ffdhe6144" 00:23:09.678 } 00:23:09.678 } 00:23:09.678 ]' 00:23:09.678 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:23:09.678 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:23:09.678 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:23:09.678 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:23:09.678 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:23:09.936 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:23:09.936 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:23:09.936 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:23:10.195 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:23:10.195 17:38:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:23:11.128 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:23:11.128 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:23:11.128 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:23:11.128 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:11.128 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:11.128 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:11.128 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:23:11.128 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:23:11.128 17:38:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:23:11.386 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe6144 3 00:23:11.386 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:23:11.386 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:23:11.386 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:23:11.386 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:23:11.386 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:23:11.386 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key3 00:23:11.386 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:11.386 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:11.386 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:11.386 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:23:11.386 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:23:11.386 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:23:11.953 00:23:11.953 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:23:11.953 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:23:11.953 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:23:12.211 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:12.211 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:23:12.211 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:12.211 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:12.211 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:12.211 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:23:12.211 { 00:23:12.211 "cntlid": 39, 00:23:12.211 "qid": 0, 00:23:12.211 "state": "enabled", 00:23:12.211 "thread": "nvmf_tgt_poll_group_000", 00:23:12.211 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:23:12.211 "listen_address": { 00:23:12.211 "trtype": "TCP", 00:23:12.211 "adrfam": "IPv4", 00:23:12.211 "traddr": "10.0.0.2", 00:23:12.211 "trsvcid": "4420" 00:23:12.211 }, 00:23:12.211 "peer_address": { 00:23:12.211 "trtype": "TCP", 00:23:12.211 "adrfam": "IPv4", 00:23:12.211 "traddr": "10.0.0.1", 00:23:12.211 "trsvcid": "56764" 00:23:12.211 }, 00:23:12.211 "auth": { 00:23:12.211 "state": "completed", 00:23:12.211 "digest": "sha256", 00:23:12.211 "dhgroup": "ffdhe6144" 00:23:12.211 } 00:23:12.211 } 00:23:12.211 ]' 00:23:12.211 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:23:12.211 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:23:12.211 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:23:12.211 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:23:12.211 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:23:12.211 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:23:12.211 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:23:12.211 17:38:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:23:12.469 17:38:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:23:12.469 17:38:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:23:13.407 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:23:13.407 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:23:13.407 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:23:13.407 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:13.407 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:13.407 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:13.407 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:23:13.407 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:23:13.407 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:23:13.407 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:23:13.664 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe8192 0 00:23:13.664 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:23:13.664 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:23:13.664 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:23:13.664 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:23:13.664 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:23:13.664 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:23:13.664 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:13.664 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:13.664 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:13.664 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:23:13.664 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:23:13.664 17:38:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:23:14.597 00:23:14.597 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:23:14.597 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:23:14.597 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:23:14.855 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:14.855 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:23:14.855 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:14.855 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:14.855 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:14.855 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:23:14.855 { 00:23:14.855 "cntlid": 41, 00:23:14.855 "qid": 0, 00:23:14.855 "state": "enabled", 00:23:14.855 "thread": "nvmf_tgt_poll_group_000", 00:23:14.855 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:23:14.855 "listen_address": { 00:23:14.855 "trtype": "TCP", 00:23:14.855 "adrfam": "IPv4", 00:23:14.855 "traddr": "10.0.0.2", 00:23:14.855 "trsvcid": "4420" 00:23:14.855 }, 00:23:14.855 "peer_address": { 00:23:14.855 "trtype": "TCP", 00:23:14.855 "adrfam": "IPv4", 00:23:14.855 "traddr": "10.0.0.1", 00:23:14.855 "trsvcid": "46020" 00:23:14.855 }, 00:23:14.855 "auth": { 00:23:14.855 "state": "completed", 00:23:14.855 "digest": "sha256", 00:23:14.855 "dhgroup": "ffdhe8192" 00:23:14.855 } 00:23:14.855 } 00:23:14.855 ]' 00:23:14.855 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:23:14.855 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:23:14.855 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:23:14.855 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:23:14.855 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:23:14.855 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:23:14.855 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:23:14.856 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:23:15.113 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:23:15.113 17:38:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:23:16.045 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:23:16.045 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:23:16.045 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:23:16.045 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:16.045 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:16.045 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:16.045 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:23:16.045 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:23:16.045 17:38:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:23:16.302 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe8192 1 00:23:16.302 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:23:16.302 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:23:16.303 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:23:16.303 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:23:16.303 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:23:16.303 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:23:16.303 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:16.303 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:16.303 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:16.303 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:23:16.303 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:23:16.303 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:23:17.237 00:23:17.237 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:23:17.237 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:23:17.237 17:38:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:23:17.496 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:17.496 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:23:17.496 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:17.496 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:17.496 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:17.496 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:23:17.496 { 00:23:17.496 "cntlid": 43, 00:23:17.496 "qid": 0, 00:23:17.496 "state": "enabled", 00:23:17.496 "thread": "nvmf_tgt_poll_group_000", 00:23:17.496 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:23:17.496 "listen_address": { 00:23:17.496 "trtype": "TCP", 00:23:17.496 "adrfam": "IPv4", 00:23:17.496 "traddr": "10.0.0.2", 00:23:17.496 "trsvcid": "4420" 00:23:17.496 }, 00:23:17.496 "peer_address": { 00:23:17.496 "trtype": "TCP", 00:23:17.496 "adrfam": "IPv4", 00:23:17.496 "traddr": "10.0.0.1", 00:23:17.496 "trsvcid": "46038" 00:23:17.496 }, 00:23:17.496 "auth": { 00:23:17.496 "state": "completed", 00:23:17.496 "digest": "sha256", 00:23:17.496 "dhgroup": "ffdhe8192" 00:23:17.496 } 00:23:17.496 } 00:23:17.496 ]' 00:23:17.496 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:23:17.496 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:23:17.496 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:23:17.496 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:23:17.496 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:23:17.496 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:23:17.496 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:23:17.496 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:23:17.754 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:23:17.754 17:38:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:23:18.688 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:23:18.688 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:23:18.688 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:23:18.688 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:18.688 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:18.688 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:18.688 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:23:18.688 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:23:18.688 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:23:18.946 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe8192 2 00:23:18.946 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:23:18.946 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:23:18.946 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:23:18.946 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:23:18.946 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:23:18.946 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:23:18.946 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:18.946 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:18.946 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:18.946 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:23:18.946 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:23:18.946 17:39:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:23:19.877 00:23:19.877 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:23:19.877 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:23:19.877 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:23:20.135 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:20.135 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:23:20.135 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:20.135 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:20.135 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:20.135 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:23:20.135 { 00:23:20.135 "cntlid": 45, 00:23:20.135 "qid": 0, 00:23:20.135 "state": "enabled", 00:23:20.135 "thread": "nvmf_tgt_poll_group_000", 00:23:20.135 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:23:20.135 "listen_address": { 00:23:20.135 "trtype": "TCP", 00:23:20.135 "adrfam": "IPv4", 00:23:20.135 "traddr": "10.0.0.2", 00:23:20.135 "trsvcid": "4420" 00:23:20.135 }, 00:23:20.135 "peer_address": { 00:23:20.135 "trtype": "TCP", 00:23:20.135 "adrfam": "IPv4", 00:23:20.135 "traddr": "10.0.0.1", 00:23:20.135 "trsvcid": "46062" 00:23:20.135 }, 00:23:20.135 "auth": { 00:23:20.135 "state": "completed", 00:23:20.135 "digest": "sha256", 00:23:20.135 "dhgroup": "ffdhe8192" 00:23:20.135 } 00:23:20.135 } 00:23:20.135 ]' 00:23:20.135 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:23:20.135 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:23:20.135 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:23:20.394 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:23:20.394 17:39:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:23:20.394 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:23:20.394 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:23:20.394 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:23:20.651 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:23:20.651 17:39:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:23:21.351 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:23:21.351 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:23:21.351 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:23:21.351 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:21.351 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:21.351 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:21.351 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:23:21.351 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:23:21.351 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:23:21.916 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe8192 3 00:23:21.916 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:23:21.916 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:23:21.916 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:23:21.916 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:23:21.916 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:23:21.917 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key3 00:23:21.917 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:21.917 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:21.917 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:21.917 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:23:21.917 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:23:21.917 17:39:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:23:22.481 00:23:22.481 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:23:22.481 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:23:22.481 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:23:22.739 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:22.739 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:23:22.739 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:22.739 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:22.996 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:22.996 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:23:22.996 { 00:23:22.996 "cntlid": 47, 00:23:22.996 "qid": 0, 00:23:22.996 "state": "enabled", 00:23:22.996 "thread": "nvmf_tgt_poll_group_000", 00:23:22.996 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:23:22.996 "listen_address": { 00:23:22.996 "trtype": "TCP", 00:23:22.996 "adrfam": "IPv4", 00:23:22.996 "traddr": "10.0.0.2", 00:23:22.996 "trsvcid": "4420" 00:23:22.996 }, 00:23:22.996 "peer_address": { 00:23:22.996 "trtype": "TCP", 00:23:22.996 "adrfam": "IPv4", 00:23:22.996 "traddr": "10.0.0.1", 00:23:22.996 "trsvcid": "46090" 00:23:22.996 }, 00:23:22.996 "auth": { 00:23:22.996 "state": "completed", 00:23:22.996 "digest": "sha256", 00:23:22.996 "dhgroup": "ffdhe8192" 00:23:22.996 } 00:23:22.996 } 00:23:22.996 ]' 00:23:22.996 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:23:22.996 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:23:22.996 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:23:22.996 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:23:22.997 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:23:22.997 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:23:22.997 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:23:22.997 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:23:23.255 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:23:23.255 17:39:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:23:24.186 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:23:24.186 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:23:24.186 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:23:24.186 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:24.186 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:24.186 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:24.186 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@118 -- # for digest in "${digests[@]}" 00:23:24.186 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:23:24.186 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:23:24.186 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:23:24.186 17:39:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:23:24.444 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 null 0 00:23:24.444 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:23:24.444 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:23:24.444 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:23:24.444 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:23:24.444 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:23:24.444 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:23:24.444 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:24.444 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:24.444 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:24.444 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:23:24.444 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:23:24.444 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:23:24.720 00:23:24.720 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:23:24.720 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:23:24.720 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:23:25.044 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:25.044 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:23:25.044 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:25.044 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:25.044 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:25.044 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:23:25.044 { 00:23:25.044 "cntlid": 49, 00:23:25.044 "qid": 0, 00:23:25.044 "state": "enabled", 00:23:25.044 "thread": "nvmf_tgt_poll_group_000", 00:23:25.044 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:23:25.044 "listen_address": { 00:23:25.044 "trtype": "TCP", 00:23:25.044 "adrfam": "IPv4", 00:23:25.044 "traddr": "10.0.0.2", 00:23:25.044 "trsvcid": "4420" 00:23:25.044 }, 00:23:25.044 "peer_address": { 00:23:25.044 "trtype": "TCP", 00:23:25.044 "adrfam": "IPv4", 00:23:25.044 "traddr": "10.0.0.1", 00:23:25.044 "trsvcid": "33498" 00:23:25.044 }, 00:23:25.044 "auth": { 00:23:25.044 "state": "completed", 00:23:25.044 "digest": "sha384", 00:23:25.044 "dhgroup": "null" 00:23:25.044 } 00:23:25.044 } 00:23:25.044 ]' 00:23:25.044 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:23:25.044 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:23:25.044 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:23:25.044 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:23:25.044 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:23:25.332 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:23:25.332 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:23:25.332 17:39:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:23:25.613 17:39:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:23:25.613 17:39:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:23:26.242 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:23:26.242 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:23:26.534 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:23:26.534 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:26.534 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:26.534 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:26.534 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:23:26.534 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:23:26.535 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:23:26.535 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 null 1 00:23:26.535 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:23:26.535 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:23:26.535 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:23:26.535 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:23:26.535 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:23:26.535 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:23:26.535 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:26.535 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:26.795 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:26.795 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:23:26.795 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:23:26.795 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:23:27.053 00:23:27.053 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:23:27.053 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:23:27.053 17:39:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:23:27.311 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:27.311 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:23:27.311 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:27.311 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:27.311 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:27.311 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:23:27.311 { 00:23:27.311 "cntlid": 51, 00:23:27.311 "qid": 0, 00:23:27.311 "state": "enabled", 00:23:27.311 "thread": "nvmf_tgt_poll_group_000", 00:23:27.311 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:23:27.312 "listen_address": { 00:23:27.312 "trtype": "TCP", 00:23:27.312 "adrfam": "IPv4", 00:23:27.312 "traddr": "10.0.0.2", 00:23:27.312 "trsvcid": "4420" 00:23:27.312 }, 00:23:27.312 "peer_address": { 00:23:27.312 "trtype": "TCP", 00:23:27.312 "adrfam": "IPv4", 00:23:27.312 "traddr": "10.0.0.1", 00:23:27.312 "trsvcid": "33518" 00:23:27.312 }, 00:23:27.312 "auth": { 00:23:27.312 "state": "completed", 00:23:27.312 "digest": "sha384", 00:23:27.312 "dhgroup": "null" 00:23:27.312 } 00:23:27.312 } 00:23:27.312 ]' 00:23:27.312 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:23:27.312 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:23:27.312 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:23:27.312 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:23:27.312 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:23:27.312 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:23:27.312 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:23:27.312 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:23:27.569 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:23:27.569 17:39:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:23:28.504 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:23:28.504 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:23:28.504 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:23:28.762 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:28.762 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:28.762 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:28.762 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:23:28.762 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:23:28.762 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:23:29.020 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 null 2 00:23:29.020 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:23:29.020 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:23:29.020 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:23:29.020 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:23:29.020 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:23:29.020 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:23:29.020 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:29.020 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:29.020 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:29.020 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:23:29.020 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:23:29.020 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:23:29.277 00:23:29.277 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:23:29.277 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:23:29.277 17:39:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:23:29.535 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:29.535 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:23:29.535 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:29.535 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:29.535 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:29.535 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:23:29.535 { 00:23:29.535 "cntlid": 53, 00:23:29.535 "qid": 0, 00:23:29.535 "state": "enabled", 00:23:29.535 "thread": "nvmf_tgt_poll_group_000", 00:23:29.535 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:23:29.535 "listen_address": { 00:23:29.535 "trtype": "TCP", 00:23:29.535 "adrfam": "IPv4", 00:23:29.535 "traddr": "10.0.0.2", 00:23:29.535 "trsvcid": "4420" 00:23:29.535 }, 00:23:29.535 "peer_address": { 00:23:29.535 "trtype": "TCP", 00:23:29.535 "adrfam": "IPv4", 00:23:29.535 "traddr": "10.0.0.1", 00:23:29.535 "trsvcid": "33560" 00:23:29.535 }, 00:23:29.535 "auth": { 00:23:29.535 "state": "completed", 00:23:29.535 "digest": "sha384", 00:23:29.535 "dhgroup": "null" 00:23:29.535 } 00:23:29.535 } 00:23:29.535 ]' 00:23:29.535 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:23:29.535 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:23:29.535 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:23:29.535 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:23:29.535 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:23:29.535 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:23:29.535 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:23:29.535 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:23:29.793 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:23:29.793 17:39:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:23:30.725 17:39:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:23:30.725 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:23:30.725 17:39:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:23:30.725 17:39:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:30.725 17:39:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:30.725 17:39:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:30.725 17:39:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:23:30.725 17:39:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:23:30.725 17:39:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:23:30.982 17:39:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 null 3 00:23:30.982 17:39:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:23:30.982 17:39:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:23:30.982 17:39:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:23:30.982 17:39:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:23:30.982 17:39:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:23:30.982 17:39:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key3 00:23:30.982 17:39:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:30.982 17:39:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:30.982 17:39:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:30.982 17:39:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:23:30.982 17:39:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:23:30.982 17:39:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:23:31.547 00:23:31.547 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:23:31.547 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:23:31.547 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:23:31.804 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:31.804 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:23:31.804 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:31.804 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:31.804 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:31.804 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:23:31.804 { 00:23:31.805 "cntlid": 55, 00:23:31.805 "qid": 0, 00:23:31.805 "state": "enabled", 00:23:31.805 "thread": "nvmf_tgt_poll_group_000", 00:23:31.805 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:23:31.805 "listen_address": { 00:23:31.805 "trtype": "TCP", 00:23:31.805 "adrfam": "IPv4", 00:23:31.805 "traddr": "10.0.0.2", 00:23:31.805 "trsvcid": "4420" 00:23:31.805 }, 00:23:31.805 "peer_address": { 00:23:31.805 "trtype": "TCP", 00:23:31.805 "adrfam": "IPv4", 00:23:31.805 "traddr": "10.0.0.1", 00:23:31.805 "trsvcid": "33580" 00:23:31.805 }, 00:23:31.805 "auth": { 00:23:31.805 "state": "completed", 00:23:31.805 "digest": "sha384", 00:23:31.805 "dhgroup": "null" 00:23:31.805 } 00:23:31.805 } 00:23:31.805 ]' 00:23:31.805 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:23:31.805 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:23:31.805 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:23:31.805 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:23:31.805 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:23:31.805 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:23:31.805 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:23:31.805 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:23:32.063 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:23:32.063 17:39:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:23:32.995 17:39:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:23:32.995 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:23:32.995 17:39:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:23:32.995 17:39:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:32.995 17:39:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:32.995 17:39:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:32.995 17:39:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:23:32.995 17:39:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:23:32.995 17:39:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:23:32.995 17:39:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:23:33.253 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe2048 0 00:23:33.253 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:23:33.253 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:23:33.253 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:23:33.253 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:23:33.253 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:23:33.253 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:23:33.253 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:33.253 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:33.253 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:33.253 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:23:33.253 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:23:33.253 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:23:33.511 00:23:33.768 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:23:33.768 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:23:33.768 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:23:34.025 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:34.025 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:23:34.025 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:34.025 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:34.025 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:34.025 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:23:34.025 { 00:23:34.025 "cntlid": 57, 00:23:34.025 "qid": 0, 00:23:34.025 "state": "enabled", 00:23:34.025 "thread": "nvmf_tgt_poll_group_000", 00:23:34.025 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:23:34.025 "listen_address": { 00:23:34.025 "trtype": "TCP", 00:23:34.025 "adrfam": "IPv4", 00:23:34.025 "traddr": "10.0.0.2", 00:23:34.025 "trsvcid": "4420" 00:23:34.025 }, 00:23:34.025 "peer_address": { 00:23:34.025 "trtype": "TCP", 00:23:34.025 "adrfam": "IPv4", 00:23:34.025 "traddr": "10.0.0.1", 00:23:34.025 "trsvcid": "40036" 00:23:34.025 }, 00:23:34.025 "auth": { 00:23:34.025 "state": "completed", 00:23:34.025 "digest": "sha384", 00:23:34.025 "dhgroup": "ffdhe2048" 00:23:34.025 } 00:23:34.025 } 00:23:34.025 ]' 00:23:34.025 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:23:34.025 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:23:34.025 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:23:34.025 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:23:34.025 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:23:34.025 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:23:34.025 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:23:34.025 17:39:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:23:34.282 17:39:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:23:34.282 17:39:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:23:35.212 17:39:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:23:35.212 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:23:35.212 17:39:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:23:35.212 17:39:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:35.212 17:39:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:35.212 17:39:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:35.212 17:39:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:23:35.212 17:39:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:23:35.212 17:39:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:23:35.469 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe2048 1 00:23:35.469 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:23:35.469 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:23:35.469 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:23:35.469 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:23:35.469 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:23:35.469 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:23:35.469 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:35.469 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:35.469 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:35.469 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:23:35.469 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:23:35.470 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:23:35.726 00:23:35.726 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:23:35.726 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:23:35.726 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:23:36.305 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:36.305 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:23:36.305 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:36.305 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:36.305 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:36.305 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:23:36.305 { 00:23:36.305 "cntlid": 59, 00:23:36.305 "qid": 0, 00:23:36.305 "state": "enabled", 00:23:36.305 "thread": "nvmf_tgt_poll_group_000", 00:23:36.305 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:23:36.305 "listen_address": { 00:23:36.305 "trtype": "TCP", 00:23:36.305 "adrfam": "IPv4", 00:23:36.305 "traddr": "10.0.0.2", 00:23:36.305 "trsvcid": "4420" 00:23:36.305 }, 00:23:36.305 "peer_address": { 00:23:36.305 "trtype": "TCP", 00:23:36.305 "adrfam": "IPv4", 00:23:36.305 "traddr": "10.0.0.1", 00:23:36.305 "trsvcid": "40064" 00:23:36.305 }, 00:23:36.305 "auth": { 00:23:36.305 "state": "completed", 00:23:36.305 "digest": "sha384", 00:23:36.305 "dhgroup": "ffdhe2048" 00:23:36.305 } 00:23:36.305 } 00:23:36.305 ]' 00:23:36.305 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:23:36.305 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:23:36.305 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:23:36.305 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:23:36.305 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:23:36.305 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:23:36.305 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:23:36.305 17:39:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:23:36.563 17:39:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:23:36.563 17:39:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:23:37.495 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:23:37.495 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:23:37.495 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:23:37.495 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:37.495 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:37.495 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:37.495 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:23:37.495 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:23:37.495 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:23:37.752 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe2048 2 00:23:37.752 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:23:37.752 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:23:37.752 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:23:37.752 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:23:37.753 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:23:37.753 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:23:37.753 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:37.753 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:37.753 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:37.753 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:23:37.753 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:23:37.753 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:23:38.010 00:23:38.010 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:23:38.010 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:23:38.010 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:23:38.268 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:38.268 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:23:38.268 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:38.268 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:38.268 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:38.268 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:23:38.268 { 00:23:38.268 "cntlid": 61, 00:23:38.268 "qid": 0, 00:23:38.268 "state": "enabled", 00:23:38.268 "thread": "nvmf_tgt_poll_group_000", 00:23:38.268 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:23:38.268 "listen_address": { 00:23:38.268 "trtype": "TCP", 00:23:38.268 "adrfam": "IPv4", 00:23:38.268 "traddr": "10.0.0.2", 00:23:38.268 "trsvcid": "4420" 00:23:38.268 }, 00:23:38.268 "peer_address": { 00:23:38.268 "trtype": "TCP", 00:23:38.268 "adrfam": "IPv4", 00:23:38.268 "traddr": "10.0.0.1", 00:23:38.268 "trsvcid": "40098" 00:23:38.268 }, 00:23:38.268 "auth": { 00:23:38.268 "state": "completed", 00:23:38.268 "digest": "sha384", 00:23:38.268 "dhgroup": "ffdhe2048" 00:23:38.268 } 00:23:38.268 } 00:23:38.268 ]' 00:23:38.268 17:39:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:23:38.268 17:39:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:23:38.268 17:39:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:23:38.268 17:39:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:23:38.268 17:39:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:23:38.268 17:39:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:23:38.268 17:39:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:23:38.268 17:39:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:23:38.833 17:39:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:23:38.833 17:39:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:23:39.422 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:23:39.679 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:23:39.679 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:23:39.679 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:39.679 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:39.679 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:39.679 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:23:39.679 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:23:39.679 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:23:39.937 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe2048 3 00:23:39.937 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:23:39.937 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:23:39.937 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:23:39.937 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:23:39.937 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:23:39.937 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key3 00:23:39.937 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:39.937 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:39.937 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:39.937 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:23:39.937 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:23:39.937 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:23:40.194 00:23:40.194 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:23:40.194 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:23:40.194 17:39:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:23:40.450 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:40.450 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:23:40.450 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:40.450 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:40.450 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:40.450 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:23:40.450 { 00:23:40.450 "cntlid": 63, 00:23:40.450 "qid": 0, 00:23:40.450 "state": "enabled", 00:23:40.450 "thread": "nvmf_tgt_poll_group_000", 00:23:40.450 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:23:40.450 "listen_address": { 00:23:40.450 "trtype": "TCP", 00:23:40.450 "adrfam": "IPv4", 00:23:40.450 "traddr": "10.0.0.2", 00:23:40.451 "trsvcid": "4420" 00:23:40.451 }, 00:23:40.451 "peer_address": { 00:23:40.451 "trtype": "TCP", 00:23:40.451 "adrfam": "IPv4", 00:23:40.451 "traddr": "10.0.0.1", 00:23:40.451 "trsvcid": "40134" 00:23:40.451 }, 00:23:40.451 "auth": { 00:23:40.451 "state": "completed", 00:23:40.451 "digest": "sha384", 00:23:40.451 "dhgroup": "ffdhe2048" 00:23:40.451 } 00:23:40.451 } 00:23:40.451 ]' 00:23:40.451 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:23:40.451 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:23:40.451 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:23:40.451 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:23:40.451 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:23:40.708 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:23:40.708 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:23:40.708 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:23:40.965 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:23:40.965 17:39:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:23:41.897 17:39:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:23:41.897 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:23:41.897 17:39:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:23:41.897 17:39:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:41.897 17:39:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:41.897 17:39:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:41.897 17:39:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:23:41.897 17:39:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:23:41.897 17:39:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:23:41.897 17:39:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:23:42.155 17:39:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe3072 0 00:23:42.155 17:39:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:23:42.155 17:39:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:23:42.155 17:39:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:23:42.155 17:39:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:23:42.155 17:39:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:23:42.155 17:39:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:23:42.155 17:39:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:42.155 17:39:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:42.155 17:39:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:42.155 17:39:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:23:42.155 17:39:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:23:42.155 17:39:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:23:42.412 00:23:42.412 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:23:42.412 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:23:42.412 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:23:42.671 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:42.671 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:23:42.671 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:42.671 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:42.671 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:42.671 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:23:42.671 { 00:23:42.671 "cntlid": 65, 00:23:42.671 "qid": 0, 00:23:42.671 "state": "enabled", 00:23:42.671 "thread": "nvmf_tgt_poll_group_000", 00:23:42.671 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:23:42.671 "listen_address": { 00:23:42.671 "trtype": "TCP", 00:23:42.671 "adrfam": "IPv4", 00:23:42.671 "traddr": "10.0.0.2", 00:23:42.671 "trsvcid": "4420" 00:23:42.671 }, 00:23:42.671 "peer_address": { 00:23:42.671 "trtype": "TCP", 00:23:42.671 "adrfam": "IPv4", 00:23:42.671 "traddr": "10.0.0.1", 00:23:42.671 "trsvcid": "40158" 00:23:42.671 }, 00:23:42.671 "auth": { 00:23:42.671 "state": "completed", 00:23:42.671 "digest": "sha384", 00:23:42.671 "dhgroup": "ffdhe3072" 00:23:42.671 } 00:23:42.671 } 00:23:42.671 ]' 00:23:42.671 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:23:42.671 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:23:42.671 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:23:42.671 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:23:42.929 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:23:42.929 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:23:42.929 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:23:42.929 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:23:43.186 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:23:43.187 17:39:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:23:44.117 17:39:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:23:44.117 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:23:44.117 17:39:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:23:44.117 17:39:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:44.117 17:39:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:44.117 17:39:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:44.117 17:39:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:23:44.117 17:39:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:23:44.117 17:39:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:23:44.374 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe3072 1 00:23:44.374 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:23:44.374 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:23:44.374 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:23:44.374 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:23:44.374 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:23:44.374 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:23:44.374 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:44.374 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:44.374 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:44.374 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:23:44.374 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:23:44.374 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:23:44.631 00:23:44.631 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:23:44.631 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:23:44.631 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:23:45.196 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:45.196 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:23:45.196 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:45.196 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:45.196 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:45.196 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:23:45.196 { 00:23:45.196 "cntlid": 67, 00:23:45.196 "qid": 0, 00:23:45.196 "state": "enabled", 00:23:45.196 "thread": "nvmf_tgt_poll_group_000", 00:23:45.196 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:23:45.196 "listen_address": { 00:23:45.196 "trtype": "TCP", 00:23:45.196 "adrfam": "IPv4", 00:23:45.196 "traddr": "10.0.0.2", 00:23:45.196 "trsvcid": "4420" 00:23:45.196 }, 00:23:45.196 "peer_address": { 00:23:45.196 "trtype": "TCP", 00:23:45.196 "adrfam": "IPv4", 00:23:45.196 "traddr": "10.0.0.1", 00:23:45.196 "trsvcid": "42994" 00:23:45.196 }, 00:23:45.196 "auth": { 00:23:45.196 "state": "completed", 00:23:45.196 "digest": "sha384", 00:23:45.196 "dhgroup": "ffdhe3072" 00:23:45.196 } 00:23:45.196 } 00:23:45.196 ]' 00:23:45.196 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:23:45.196 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:23:45.196 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:23:45.196 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:23:45.196 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:23:45.196 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:23:45.196 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:23:45.196 17:39:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:23:45.453 17:39:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:23:45.453 17:39:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:23:46.386 17:39:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:23:46.386 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:23:46.386 17:39:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:23:46.386 17:39:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:46.386 17:39:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:46.386 17:39:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:46.386 17:39:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:23:46.386 17:39:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:23:46.386 17:39:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:23:46.644 17:39:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe3072 2 00:23:46.644 17:39:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:23:46.644 17:39:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:23:46.644 17:39:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:23:46.644 17:39:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:23:46.644 17:39:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:23:46.644 17:39:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:23:46.644 17:39:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:46.644 17:39:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:46.644 17:39:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:46.644 17:39:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:23:46.644 17:39:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:23:46.644 17:39:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:23:46.902 00:23:47.160 17:39:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:23:47.160 17:39:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:23:47.160 17:39:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:23:47.418 17:39:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:47.418 17:39:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:23:47.418 17:39:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:47.418 17:39:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:47.418 17:39:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:47.418 17:39:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:23:47.418 { 00:23:47.418 "cntlid": 69, 00:23:47.418 "qid": 0, 00:23:47.418 "state": "enabled", 00:23:47.418 "thread": "nvmf_tgt_poll_group_000", 00:23:47.418 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:23:47.418 "listen_address": { 00:23:47.418 "trtype": "TCP", 00:23:47.418 "adrfam": "IPv4", 00:23:47.418 "traddr": "10.0.0.2", 00:23:47.418 "trsvcid": "4420" 00:23:47.418 }, 00:23:47.418 "peer_address": { 00:23:47.418 "trtype": "TCP", 00:23:47.418 "adrfam": "IPv4", 00:23:47.418 "traddr": "10.0.0.1", 00:23:47.418 "trsvcid": "43014" 00:23:47.418 }, 00:23:47.418 "auth": { 00:23:47.418 "state": "completed", 00:23:47.418 "digest": "sha384", 00:23:47.418 "dhgroup": "ffdhe3072" 00:23:47.418 } 00:23:47.418 } 00:23:47.418 ]' 00:23:47.418 17:39:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:23:47.418 17:39:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:23:47.418 17:39:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:23:47.418 17:39:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:23:47.418 17:39:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:23:47.418 17:39:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:23:47.418 17:39:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:23:47.418 17:39:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:23:47.676 17:39:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:23:47.676 17:39:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:23:48.608 17:39:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:23:48.608 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:23:48.608 17:39:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:23:48.608 17:39:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:48.608 17:39:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:48.608 17:39:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:48.608 17:39:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:23:48.608 17:39:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:23:48.608 17:39:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:23:48.865 17:39:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe3072 3 00:23:48.865 17:39:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:23:48.865 17:39:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:23:48.865 17:39:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:23:48.865 17:39:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:23:48.865 17:39:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:23:48.865 17:39:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key3 00:23:48.865 17:39:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:48.865 17:39:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:48.865 17:39:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:48.865 17:39:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:23:48.865 17:39:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:23:48.865 17:39:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:23:49.429 00:23:49.429 17:39:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:23:49.429 17:39:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:23:49.429 17:39:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:23:49.686 17:39:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:49.686 17:39:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:23:49.686 17:39:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:49.686 17:39:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:49.686 17:39:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:49.686 17:39:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:23:49.686 { 00:23:49.686 "cntlid": 71, 00:23:49.686 "qid": 0, 00:23:49.686 "state": "enabled", 00:23:49.686 "thread": "nvmf_tgt_poll_group_000", 00:23:49.686 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:23:49.686 "listen_address": { 00:23:49.686 "trtype": "TCP", 00:23:49.686 "adrfam": "IPv4", 00:23:49.686 "traddr": "10.0.0.2", 00:23:49.686 "trsvcid": "4420" 00:23:49.686 }, 00:23:49.686 "peer_address": { 00:23:49.686 "trtype": "TCP", 00:23:49.686 "adrfam": "IPv4", 00:23:49.686 "traddr": "10.0.0.1", 00:23:49.686 "trsvcid": "43048" 00:23:49.686 }, 00:23:49.686 "auth": { 00:23:49.686 "state": "completed", 00:23:49.686 "digest": "sha384", 00:23:49.686 "dhgroup": "ffdhe3072" 00:23:49.686 } 00:23:49.686 } 00:23:49.686 ]' 00:23:49.686 17:39:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:23:49.686 17:39:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:23:49.686 17:39:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:23:49.686 17:39:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:23:49.686 17:39:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:23:49.686 17:39:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:23:49.686 17:39:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:23:49.686 17:39:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:23:50.248 17:39:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:23:50.248 17:39:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:23:50.835 17:39:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:23:50.835 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:23:50.835 17:39:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:23:50.835 17:39:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:50.835 17:39:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:50.835 17:39:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:50.835 17:39:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:23:50.835 17:39:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:23:50.835 17:39:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:23:50.836 17:39:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:23:51.093 17:39:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe4096 0 00:23:51.093 17:39:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:23:51.093 17:39:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:23:51.093 17:39:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:23:51.093 17:39:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:23:51.093 17:39:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:23:51.093 17:39:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:23:51.093 17:39:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:51.093 17:39:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:51.093 17:39:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:51.093 17:39:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:23:51.093 17:39:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:23:51.093 17:39:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:23:51.657 00:23:51.657 17:39:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:23:51.657 17:39:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:23:51.657 17:39:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:23:51.915 17:39:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:51.915 17:39:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:23:51.915 17:39:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:51.915 17:39:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:51.915 17:39:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:51.915 17:39:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:23:51.915 { 00:23:51.915 "cntlid": 73, 00:23:51.915 "qid": 0, 00:23:51.915 "state": "enabled", 00:23:51.915 "thread": "nvmf_tgt_poll_group_000", 00:23:51.915 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:23:51.915 "listen_address": { 00:23:51.915 "trtype": "TCP", 00:23:51.915 "adrfam": "IPv4", 00:23:51.915 "traddr": "10.0.0.2", 00:23:51.915 "trsvcid": "4420" 00:23:51.915 }, 00:23:51.915 "peer_address": { 00:23:51.915 "trtype": "TCP", 00:23:51.915 "adrfam": "IPv4", 00:23:51.915 "traddr": "10.0.0.1", 00:23:51.915 "trsvcid": "43082" 00:23:51.915 }, 00:23:51.915 "auth": { 00:23:51.915 "state": "completed", 00:23:51.915 "digest": "sha384", 00:23:51.915 "dhgroup": "ffdhe4096" 00:23:51.915 } 00:23:51.915 } 00:23:51.915 ]' 00:23:51.915 17:39:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:23:51.915 17:39:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:23:51.915 17:39:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:23:51.915 17:39:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:23:51.915 17:39:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:23:51.915 17:39:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:23:51.915 17:39:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:23:51.915 17:39:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:23:52.479 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:23:52.479 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:23:53.410 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:23:53.410 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:23:53.410 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:23:53.410 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:53.410 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:53.410 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:53.410 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:23:53.410 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:23:53.410 17:39:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:23:53.410 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe4096 1 00:23:53.410 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:23:53.410 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:23:53.410 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:23:53.410 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:23:53.410 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:23:53.410 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:23:53.410 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:53.411 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:53.411 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:53.411 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:23:53.411 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:23:53.411 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:23:53.975 00:23:53.975 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:23:53.975 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:23:53.975 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:23:54.231 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:54.231 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:23:54.231 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:54.231 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:54.231 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:54.231 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:23:54.231 { 00:23:54.231 "cntlid": 75, 00:23:54.231 "qid": 0, 00:23:54.231 "state": "enabled", 00:23:54.231 "thread": "nvmf_tgt_poll_group_000", 00:23:54.231 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:23:54.231 "listen_address": { 00:23:54.231 "trtype": "TCP", 00:23:54.231 "adrfam": "IPv4", 00:23:54.231 "traddr": "10.0.0.2", 00:23:54.231 "trsvcid": "4420" 00:23:54.231 }, 00:23:54.231 "peer_address": { 00:23:54.231 "trtype": "TCP", 00:23:54.231 "adrfam": "IPv4", 00:23:54.231 "traddr": "10.0.0.1", 00:23:54.231 "trsvcid": "50422" 00:23:54.231 }, 00:23:54.231 "auth": { 00:23:54.231 "state": "completed", 00:23:54.231 "digest": "sha384", 00:23:54.231 "dhgroup": "ffdhe4096" 00:23:54.231 } 00:23:54.231 } 00:23:54.231 ]' 00:23:54.231 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:23:54.231 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:23:54.231 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:23:54.231 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:23:54.231 17:39:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:23:54.231 17:39:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:23:54.231 17:39:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:23:54.231 17:39:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:23:54.488 17:39:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:23:54.488 17:39:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:23:55.420 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:23:55.420 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:23:55.420 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:23:55.420 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:55.420 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:55.420 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:55.420 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:23:55.420 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:23:55.420 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:23:55.677 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe4096 2 00:23:55.677 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:23:55.677 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:23:55.677 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:23:55.677 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:23:55.677 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:23:55.677 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:23:55.677 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:55.677 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:55.677 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:55.677 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:23:55.677 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:23:55.677 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:23:56.242 00:23:56.242 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:23:56.242 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:23:56.242 17:39:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:23:56.499 17:39:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:56.499 17:39:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:23:56.499 17:39:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:56.499 17:39:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:56.499 17:39:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:56.499 17:39:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:23:56.499 { 00:23:56.499 "cntlid": 77, 00:23:56.499 "qid": 0, 00:23:56.499 "state": "enabled", 00:23:56.499 "thread": "nvmf_tgt_poll_group_000", 00:23:56.499 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:23:56.499 "listen_address": { 00:23:56.499 "trtype": "TCP", 00:23:56.499 "adrfam": "IPv4", 00:23:56.499 "traddr": "10.0.0.2", 00:23:56.499 "trsvcid": "4420" 00:23:56.499 }, 00:23:56.499 "peer_address": { 00:23:56.499 "trtype": "TCP", 00:23:56.499 "adrfam": "IPv4", 00:23:56.499 "traddr": "10.0.0.1", 00:23:56.499 "trsvcid": "50442" 00:23:56.499 }, 00:23:56.499 "auth": { 00:23:56.499 "state": "completed", 00:23:56.499 "digest": "sha384", 00:23:56.499 "dhgroup": "ffdhe4096" 00:23:56.499 } 00:23:56.499 } 00:23:56.499 ]' 00:23:56.499 17:39:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:23:56.499 17:39:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:23:56.499 17:39:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:23:56.499 17:39:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:23:56.499 17:39:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:23:56.499 17:39:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:23:56.499 17:39:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:23:56.499 17:39:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:23:56.757 17:39:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:23:56.757 17:39:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:23:57.689 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:23:57.689 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:23:57.689 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:23:57.689 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:57.689 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:57.689 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:57.689 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:23:57.689 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:23:57.689 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:23:57.947 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe4096 3 00:23:57.947 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:23:57.947 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:23:57.947 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:23:57.947 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:23:57.947 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:23:57.947 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key3 00:23:57.947 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:57.947 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:57.947 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:57.947 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:23:57.947 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:23:57.947 17:39:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:23:58.512 00:23:58.512 17:39:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:23:58.512 17:39:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:23:58.512 17:39:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:23:58.769 17:39:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:58.769 17:39:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:23:58.769 17:39:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:58.769 17:39:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:58.769 17:39:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:58.769 17:39:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:23:58.769 { 00:23:58.769 "cntlid": 79, 00:23:58.769 "qid": 0, 00:23:58.769 "state": "enabled", 00:23:58.769 "thread": "nvmf_tgt_poll_group_000", 00:23:58.769 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:23:58.769 "listen_address": { 00:23:58.769 "trtype": "TCP", 00:23:58.769 "adrfam": "IPv4", 00:23:58.769 "traddr": "10.0.0.2", 00:23:58.769 "trsvcid": "4420" 00:23:58.769 }, 00:23:58.769 "peer_address": { 00:23:58.769 "trtype": "TCP", 00:23:58.769 "adrfam": "IPv4", 00:23:58.769 "traddr": "10.0.0.1", 00:23:58.769 "trsvcid": "50456" 00:23:58.769 }, 00:23:58.769 "auth": { 00:23:58.769 "state": "completed", 00:23:58.769 "digest": "sha384", 00:23:58.769 "dhgroup": "ffdhe4096" 00:23:58.769 } 00:23:58.769 } 00:23:58.769 ]' 00:23:58.769 17:39:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:23:58.769 17:39:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:23:58.769 17:39:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:23:58.769 17:39:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:23:58.769 17:39:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:23:58.769 17:39:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:23:58.770 17:39:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:23:58.770 17:39:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:23:59.027 17:39:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:23:59.027 17:39:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:23:59.960 17:39:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:23:59.960 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:23:59.960 17:39:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:23:59.960 17:39:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:23:59.960 17:39:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:23:59.960 17:39:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:23:59.960 17:39:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:23:59.960 17:39:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:23:59.960 17:39:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:23:59.960 17:39:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:24:00.218 17:39:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe6144 0 00:24:00.218 17:39:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:24:00.218 17:39:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:24:00.218 17:39:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:24:00.218 17:39:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:24:00.218 17:39:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:24:00.218 17:39:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:24:00.218 17:39:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:00.218 17:39:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:00.218 17:39:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:00.218 17:39:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:24:00.218 17:39:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:24:00.218 17:39:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:24:01.151 00:24:01.151 17:39:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:24:01.151 17:39:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:24:01.151 17:39:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:24:01.151 17:39:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:01.151 17:39:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:24:01.151 17:39:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:01.151 17:39:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:01.151 17:39:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:01.151 17:39:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:24:01.151 { 00:24:01.151 "cntlid": 81, 00:24:01.151 "qid": 0, 00:24:01.151 "state": "enabled", 00:24:01.151 "thread": "nvmf_tgt_poll_group_000", 00:24:01.151 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:24:01.151 "listen_address": { 00:24:01.151 "trtype": "TCP", 00:24:01.151 "adrfam": "IPv4", 00:24:01.151 "traddr": "10.0.0.2", 00:24:01.151 "trsvcid": "4420" 00:24:01.151 }, 00:24:01.151 "peer_address": { 00:24:01.151 "trtype": "TCP", 00:24:01.151 "adrfam": "IPv4", 00:24:01.151 "traddr": "10.0.0.1", 00:24:01.151 "trsvcid": "50496" 00:24:01.151 }, 00:24:01.151 "auth": { 00:24:01.151 "state": "completed", 00:24:01.151 "digest": "sha384", 00:24:01.151 "dhgroup": "ffdhe6144" 00:24:01.151 } 00:24:01.151 } 00:24:01.151 ]' 00:24:01.151 17:39:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:24:01.151 17:39:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:24:01.151 17:39:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:24:01.408 17:39:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:24:01.408 17:39:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:24:01.408 17:39:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:24:01.408 17:39:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:24:01.408 17:39:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:24:01.666 17:39:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:24:01.666 17:39:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:24:02.600 17:39:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:24:02.600 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:24:02.600 17:39:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:24:02.600 17:39:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:02.600 17:39:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:02.600 17:39:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:02.600 17:39:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:24:02.600 17:39:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:24:02.600 17:39:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:24:02.857 17:39:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe6144 1 00:24:02.857 17:39:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:24:02.857 17:39:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:24:02.857 17:39:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:24:02.857 17:39:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:24:02.857 17:39:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:24:02.857 17:39:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:24:02.857 17:39:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:02.857 17:39:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:02.857 17:39:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:02.857 17:39:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:24:02.857 17:39:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:24:02.857 17:39:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:24:03.423 00:24:03.423 17:39:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:24:03.423 17:39:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:24:03.423 17:39:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:24:03.679 17:39:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:03.679 17:39:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:24:03.679 17:39:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:03.679 17:39:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:03.679 17:39:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:03.679 17:39:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:24:03.679 { 00:24:03.679 "cntlid": 83, 00:24:03.679 "qid": 0, 00:24:03.679 "state": "enabled", 00:24:03.679 "thread": "nvmf_tgt_poll_group_000", 00:24:03.679 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:24:03.679 "listen_address": { 00:24:03.679 "trtype": "TCP", 00:24:03.679 "adrfam": "IPv4", 00:24:03.679 "traddr": "10.0.0.2", 00:24:03.679 "trsvcid": "4420" 00:24:03.679 }, 00:24:03.679 "peer_address": { 00:24:03.679 "trtype": "TCP", 00:24:03.679 "adrfam": "IPv4", 00:24:03.679 "traddr": "10.0.0.1", 00:24:03.679 "trsvcid": "50522" 00:24:03.679 }, 00:24:03.679 "auth": { 00:24:03.679 "state": "completed", 00:24:03.679 "digest": "sha384", 00:24:03.679 "dhgroup": "ffdhe6144" 00:24:03.679 } 00:24:03.679 } 00:24:03.679 ]' 00:24:03.679 17:39:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:24:03.679 17:39:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:24:03.679 17:39:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:24:03.679 17:39:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:24:03.680 17:39:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:24:03.680 17:39:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:24:03.680 17:39:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:24:03.680 17:39:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:24:03.935 17:39:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:24:03.936 17:39:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:24:04.865 17:39:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:24:04.865 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:24:04.865 17:39:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:24:04.865 17:39:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:04.865 17:39:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:04.865 17:39:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:04.865 17:39:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:24:04.865 17:39:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:24:04.865 17:39:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:24:05.121 17:39:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe6144 2 00:24:05.121 17:39:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:24:05.121 17:39:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:24:05.121 17:39:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:24:05.121 17:39:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:24:05.121 17:39:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:24:05.121 17:39:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:24:05.121 17:39:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:05.121 17:39:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:05.121 17:39:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:05.121 17:39:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:24:05.121 17:39:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:24:05.121 17:39:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:24:05.683 00:24:05.683 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:24:05.683 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:24:05.683 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:24:05.940 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:05.940 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:24:05.940 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:05.940 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:05.940 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:05.940 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:24:05.940 { 00:24:05.940 "cntlid": 85, 00:24:05.940 "qid": 0, 00:24:05.940 "state": "enabled", 00:24:05.940 "thread": "nvmf_tgt_poll_group_000", 00:24:05.940 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:24:05.940 "listen_address": { 00:24:05.940 "trtype": "TCP", 00:24:05.940 "adrfam": "IPv4", 00:24:05.940 "traddr": "10.0.0.2", 00:24:05.940 "trsvcid": "4420" 00:24:05.940 }, 00:24:05.940 "peer_address": { 00:24:05.940 "trtype": "TCP", 00:24:05.940 "adrfam": "IPv4", 00:24:05.940 "traddr": "10.0.0.1", 00:24:05.940 "trsvcid": "33774" 00:24:05.940 }, 00:24:05.940 "auth": { 00:24:05.940 "state": "completed", 00:24:05.940 "digest": "sha384", 00:24:05.940 "dhgroup": "ffdhe6144" 00:24:05.940 } 00:24:05.940 } 00:24:05.940 ]' 00:24:05.940 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:24:05.940 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:24:05.940 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:24:06.197 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:24:06.197 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:24:06.197 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:24:06.197 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:24:06.197 17:39:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:24:06.454 17:39:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:24:06.454 17:39:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:24:07.385 17:39:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:24:07.385 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:24:07.385 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:24:07.386 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:07.386 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:07.386 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:07.386 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:24:07.386 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:24:07.386 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:24:07.643 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe6144 3 00:24:07.643 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:24:07.643 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:24:07.643 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:24:07.643 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:24:07.643 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:24:07.643 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key3 00:24:07.643 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:07.643 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:07.643 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:07.643 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:24:07.643 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:24:07.643 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:24:08.208 00:24:08.208 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:24:08.208 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:24:08.208 17:39:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:24:08.467 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:08.467 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:24:08.467 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:08.467 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:08.467 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:08.467 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:24:08.467 { 00:24:08.467 "cntlid": 87, 00:24:08.467 "qid": 0, 00:24:08.467 "state": "enabled", 00:24:08.467 "thread": "nvmf_tgt_poll_group_000", 00:24:08.467 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:24:08.467 "listen_address": { 00:24:08.467 "trtype": "TCP", 00:24:08.467 "adrfam": "IPv4", 00:24:08.467 "traddr": "10.0.0.2", 00:24:08.467 "trsvcid": "4420" 00:24:08.467 }, 00:24:08.467 "peer_address": { 00:24:08.467 "trtype": "TCP", 00:24:08.467 "adrfam": "IPv4", 00:24:08.467 "traddr": "10.0.0.1", 00:24:08.467 "trsvcid": "33796" 00:24:08.467 }, 00:24:08.467 "auth": { 00:24:08.467 "state": "completed", 00:24:08.467 "digest": "sha384", 00:24:08.467 "dhgroup": "ffdhe6144" 00:24:08.467 } 00:24:08.467 } 00:24:08.467 ]' 00:24:08.467 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:24:08.467 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:24:08.467 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:24:08.467 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:24:08.467 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:24:08.467 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:24:08.467 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:24:08.467 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:24:08.725 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:24:08.725 17:39:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:24:09.657 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:24:09.657 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:24:09.657 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:24:09.657 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:09.657 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:09.657 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:09.657 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:24:09.657 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:24:09.657 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:24:09.657 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:24:09.915 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe8192 0 00:24:09.915 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:24:09.915 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:24:09.915 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:24:09.915 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:24:09.915 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:24:09.915 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:24:09.915 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:09.915 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:09.915 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:09.915 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:24:09.915 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:24:09.915 17:39:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:24:10.848 00:24:10.848 17:39:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:24:10.848 17:39:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:24:10.848 17:39:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:24:11.106 17:39:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:11.106 17:39:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:24:11.106 17:39:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:11.106 17:39:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:11.106 17:39:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:11.106 17:39:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:24:11.106 { 00:24:11.106 "cntlid": 89, 00:24:11.106 "qid": 0, 00:24:11.106 "state": "enabled", 00:24:11.106 "thread": "nvmf_tgt_poll_group_000", 00:24:11.106 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:24:11.106 "listen_address": { 00:24:11.106 "trtype": "TCP", 00:24:11.106 "adrfam": "IPv4", 00:24:11.106 "traddr": "10.0.0.2", 00:24:11.106 "trsvcid": "4420" 00:24:11.106 }, 00:24:11.106 "peer_address": { 00:24:11.106 "trtype": "TCP", 00:24:11.106 "adrfam": "IPv4", 00:24:11.106 "traddr": "10.0.0.1", 00:24:11.106 "trsvcid": "33830" 00:24:11.106 }, 00:24:11.106 "auth": { 00:24:11.106 "state": "completed", 00:24:11.106 "digest": "sha384", 00:24:11.106 "dhgroup": "ffdhe8192" 00:24:11.106 } 00:24:11.106 } 00:24:11.106 ]' 00:24:11.106 17:39:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:24:11.106 17:39:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:24:11.106 17:39:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:24:11.106 17:39:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:24:11.106 17:39:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:24:11.106 17:39:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:24:11.106 17:39:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:24:11.106 17:39:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:24:11.364 17:39:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:24:11.364 17:39:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:24:12.298 17:39:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:24:12.298 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:24:12.298 17:39:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:24:12.298 17:39:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:12.298 17:39:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:12.298 17:39:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:12.298 17:39:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:24:12.298 17:39:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:24:12.298 17:39:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:24:12.864 17:39:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe8192 1 00:24:12.864 17:39:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:24:12.864 17:39:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:24:12.864 17:39:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:24:12.864 17:39:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:24:12.864 17:39:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:24:12.864 17:39:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:24:12.864 17:39:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:12.864 17:39:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:12.864 17:39:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:12.864 17:39:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:24:12.864 17:39:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:24:12.864 17:39:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:24:13.428 00:24:13.428 17:39:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:24:13.428 17:39:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:24:13.428 17:39:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:24:13.684 17:39:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:13.684 17:39:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:24:13.941 17:39:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:13.941 17:39:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:13.941 17:39:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:13.941 17:39:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:24:13.941 { 00:24:13.941 "cntlid": 91, 00:24:13.941 "qid": 0, 00:24:13.941 "state": "enabled", 00:24:13.941 "thread": "nvmf_tgt_poll_group_000", 00:24:13.941 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:24:13.941 "listen_address": { 00:24:13.941 "trtype": "TCP", 00:24:13.941 "adrfam": "IPv4", 00:24:13.941 "traddr": "10.0.0.2", 00:24:13.941 "trsvcid": "4420" 00:24:13.941 }, 00:24:13.941 "peer_address": { 00:24:13.941 "trtype": "TCP", 00:24:13.941 "adrfam": "IPv4", 00:24:13.941 "traddr": "10.0.0.1", 00:24:13.941 "trsvcid": "33856" 00:24:13.941 }, 00:24:13.941 "auth": { 00:24:13.941 "state": "completed", 00:24:13.941 "digest": "sha384", 00:24:13.941 "dhgroup": "ffdhe8192" 00:24:13.941 } 00:24:13.941 } 00:24:13.941 ]' 00:24:13.941 17:39:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:24:13.941 17:39:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:24:13.941 17:39:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:24:13.941 17:39:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:24:13.941 17:39:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:24:13.941 17:39:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:24:13.941 17:39:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:24:13.941 17:39:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:24:14.198 17:39:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:24:14.198 17:39:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:24:15.130 17:39:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:24:15.130 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:24:15.131 17:39:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:24:15.131 17:39:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:15.131 17:39:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:15.131 17:39:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:15.131 17:39:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:24:15.131 17:39:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:24:15.131 17:39:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:24:15.389 17:39:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe8192 2 00:24:15.389 17:39:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:24:15.389 17:39:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:24:15.389 17:39:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:24:15.389 17:39:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:24:15.389 17:39:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:24:15.389 17:39:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:24:15.389 17:39:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:15.389 17:39:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:15.389 17:39:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:15.389 17:39:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:24:15.389 17:39:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:24:15.389 17:39:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:24:16.323 00:24:16.323 17:39:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:24:16.323 17:39:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:24:16.323 17:39:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:24:16.581 17:39:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:16.581 17:39:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:24:16.581 17:39:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:16.581 17:39:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:16.581 17:39:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:16.581 17:39:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:24:16.581 { 00:24:16.581 "cntlid": 93, 00:24:16.581 "qid": 0, 00:24:16.581 "state": "enabled", 00:24:16.581 "thread": "nvmf_tgt_poll_group_000", 00:24:16.581 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:24:16.581 "listen_address": { 00:24:16.582 "trtype": "TCP", 00:24:16.582 "adrfam": "IPv4", 00:24:16.582 "traddr": "10.0.0.2", 00:24:16.582 "trsvcid": "4420" 00:24:16.582 }, 00:24:16.582 "peer_address": { 00:24:16.582 "trtype": "TCP", 00:24:16.582 "adrfam": "IPv4", 00:24:16.582 "traddr": "10.0.0.1", 00:24:16.582 "trsvcid": "48090" 00:24:16.582 }, 00:24:16.582 "auth": { 00:24:16.582 "state": "completed", 00:24:16.582 "digest": "sha384", 00:24:16.582 "dhgroup": "ffdhe8192" 00:24:16.582 } 00:24:16.582 } 00:24:16.582 ]' 00:24:16.582 17:39:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:24:16.582 17:39:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:24:16.582 17:39:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:24:16.582 17:39:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:24:16.582 17:39:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:24:16.582 17:39:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:24:16.582 17:39:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:24:16.582 17:39:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:24:16.840 17:39:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:24:16.840 17:39:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:24:18.217 17:39:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:24:18.217 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:24:18.217 17:39:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:24:18.217 17:39:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:18.217 17:39:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:18.217 17:39:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:18.217 17:39:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:24:18.217 17:39:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:24:18.217 17:39:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:24:18.217 17:39:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe8192 3 00:24:18.217 17:39:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:24:18.217 17:39:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:24:18.217 17:39:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:24:18.218 17:39:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:24:18.218 17:39:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:24:18.218 17:39:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key3 00:24:18.218 17:39:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:18.218 17:39:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:18.218 17:39:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:18.218 17:39:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:24:18.218 17:39:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:24:18.218 17:39:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:24:19.152 00:24:19.152 17:40:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:24:19.152 17:40:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:24:19.152 17:40:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:24:19.408 17:40:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:19.408 17:40:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:24:19.408 17:40:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:19.408 17:40:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:19.408 17:40:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:19.408 17:40:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:24:19.408 { 00:24:19.408 "cntlid": 95, 00:24:19.408 "qid": 0, 00:24:19.408 "state": "enabled", 00:24:19.408 "thread": "nvmf_tgt_poll_group_000", 00:24:19.408 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:24:19.408 "listen_address": { 00:24:19.408 "trtype": "TCP", 00:24:19.408 "adrfam": "IPv4", 00:24:19.408 "traddr": "10.0.0.2", 00:24:19.408 "trsvcid": "4420" 00:24:19.408 }, 00:24:19.408 "peer_address": { 00:24:19.408 "trtype": "TCP", 00:24:19.408 "adrfam": "IPv4", 00:24:19.408 "traddr": "10.0.0.1", 00:24:19.408 "trsvcid": "48114" 00:24:19.408 }, 00:24:19.408 "auth": { 00:24:19.408 "state": "completed", 00:24:19.408 "digest": "sha384", 00:24:19.408 "dhgroup": "ffdhe8192" 00:24:19.408 } 00:24:19.408 } 00:24:19.408 ]' 00:24:19.408 17:40:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:24:19.408 17:40:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:24:19.408 17:40:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:24:19.408 17:40:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:24:19.408 17:40:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:24:19.408 17:40:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:24:19.408 17:40:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:24:19.408 17:40:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:24:19.665 17:40:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:24:19.665 17:40:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:24:20.596 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:24:20.596 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:24:20.596 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:24:20.596 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:20.596 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:20.596 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:20.596 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@118 -- # for digest in "${digests[@]}" 00:24:20.596 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:24:20.596 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:24:20.597 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:24:20.597 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:24:20.853 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 null 0 00:24:20.853 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:24:20.853 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:24:20.854 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:24:20.854 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:24:20.854 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:24:20.854 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:24:20.854 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:20.854 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:21.111 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:21.111 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:24:21.111 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:24:21.111 17:40:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:24:21.368 00:24:21.368 17:40:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:24:21.368 17:40:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:24:21.368 17:40:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:24:21.625 17:40:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:21.625 17:40:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:24:21.625 17:40:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:21.687 17:40:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:21.687 17:40:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:21.687 17:40:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:24:21.687 { 00:24:21.687 "cntlid": 97, 00:24:21.687 "qid": 0, 00:24:21.687 "state": "enabled", 00:24:21.687 "thread": "nvmf_tgt_poll_group_000", 00:24:21.687 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:24:21.687 "listen_address": { 00:24:21.687 "trtype": "TCP", 00:24:21.687 "adrfam": "IPv4", 00:24:21.687 "traddr": "10.0.0.2", 00:24:21.687 "trsvcid": "4420" 00:24:21.687 }, 00:24:21.687 "peer_address": { 00:24:21.687 "trtype": "TCP", 00:24:21.687 "adrfam": "IPv4", 00:24:21.687 "traddr": "10.0.0.1", 00:24:21.687 "trsvcid": "48128" 00:24:21.687 }, 00:24:21.687 "auth": { 00:24:21.687 "state": "completed", 00:24:21.687 "digest": "sha512", 00:24:21.687 "dhgroup": "null" 00:24:21.687 } 00:24:21.687 } 00:24:21.687 ]' 00:24:21.687 17:40:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:24:21.687 17:40:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:24:21.687 17:40:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:24:21.687 17:40:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:24:21.687 17:40:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:24:21.687 17:40:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:24:21.687 17:40:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:24:21.687 17:40:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:24:21.944 17:40:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:24:21.944 17:40:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:24:22.875 17:40:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:24:22.875 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:24:22.875 17:40:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:24:22.875 17:40:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:22.875 17:40:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:22.876 17:40:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:22.876 17:40:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:24:22.876 17:40:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:24:22.876 17:40:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:24:23.132 17:40:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 null 1 00:24:23.132 17:40:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:24:23.132 17:40:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:24:23.132 17:40:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:24:23.132 17:40:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:24:23.132 17:40:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:24:23.132 17:40:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:24:23.132 17:40:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:23.132 17:40:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:23.132 17:40:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:23.132 17:40:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:24:23.132 17:40:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:24:23.132 17:40:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:24:23.696 00:24:23.696 17:40:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:24:23.696 17:40:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:24:23.696 17:40:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:24:23.696 17:40:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:23.696 17:40:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:24:23.696 17:40:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:23.696 17:40:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:23.696 17:40:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:23.696 17:40:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:24:23.696 { 00:24:23.696 "cntlid": 99, 00:24:23.696 "qid": 0, 00:24:23.696 "state": "enabled", 00:24:23.696 "thread": "nvmf_tgt_poll_group_000", 00:24:23.696 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:24:23.696 "listen_address": { 00:24:23.696 "trtype": "TCP", 00:24:23.696 "adrfam": "IPv4", 00:24:23.696 "traddr": "10.0.0.2", 00:24:23.696 "trsvcid": "4420" 00:24:23.696 }, 00:24:23.696 "peer_address": { 00:24:23.696 "trtype": "TCP", 00:24:23.696 "adrfam": "IPv4", 00:24:23.696 "traddr": "10.0.0.1", 00:24:23.696 "trsvcid": "44890" 00:24:23.696 }, 00:24:23.696 "auth": { 00:24:23.696 "state": "completed", 00:24:23.696 "digest": "sha512", 00:24:23.696 "dhgroup": "null" 00:24:23.696 } 00:24:23.696 } 00:24:23.696 ]' 00:24:23.696 17:40:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:24:23.953 17:40:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:24:23.953 17:40:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:24:23.953 17:40:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:24:23.953 17:40:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:24:23.953 17:40:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:24:23.953 17:40:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:24:23.953 17:40:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:24:24.210 17:40:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:24:24.210 17:40:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:24:25.141 17:40:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:24:25.141 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:24:25.141 17:40:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:24:25.141 17:40:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:25.141 17:40:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:25.141 17:40:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:25.141 17:40:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:24:25.141 17:40:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:24:25.141 17:40:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:24:25.399 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 null 2 00:24:25.399 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:24:25.399 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:24:25.399 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:24:25.399 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:24:25.399 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:24:25.399 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:24:25.399 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:25.399 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:25.399 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:25.399 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:24:25.399 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:24:25.399 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:24:25.657 00:24:25.657 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:24:25.657 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:24:25.657 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:24:25.914 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:25.915 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:24:25.915 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:25.915 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:25.915 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:25.915 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:24:25.915 { 00:24:25.915 "cntlid": 101, 00:24:25.915 "qid": 0, 00:24:25.915 "state": "enabled", 00:24:25.915 "thread": "nvmf_tgt_poll_group_000", 00:24:25.915 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:24:25.915 "listen_address": { 00:24:25.915 "trtype": "TCP", 00:24:25.915 "adrfam": "IPv4", 00:24:25.915 "traddr": "10.0.0.2", 00:24:25.915 "trsvcid": "4420" 00:24:25.915 }, 00:24:25.915 "peer_address": { 00:24:25.915 "trtype": "TCP", 00:24:25.915 "adrfam": "IPv4", 00:24:25.915 "traddr": "10.0.0.1", 00:24:25.915 "trsvcid": "44910" 00:24:25.915 }, 00:24:25.915 "auth": { 00:24:25.915 "state": "completed", 00:24:25.915 "digest": "sha512", 00:24:25.915 "dhgroup": "null" 00:24:25.915 } 00:24:25.915 } 00:24:25.915 ]' 00:24:25.915 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:24:25.915 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:24:25.915 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:24:26.173 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:24:26.173 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:24:26.173 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:24:26.173 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:24:26.173 17:40:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:24:26.431 17:40:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:24:26.431 17:40:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:24:27.365 17:40:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:24:27.365 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:24:27.365 17:40:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:24:27.365 17:40:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:27.365 17:40:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:27.365 17:40:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:27.365 17:40:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:24:27.365 17:40:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:24:27.365 17:40:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:24:27.624 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 null 3 00:24:27.624 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:24:27.624 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:24:27.624 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:24:27.624 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:24:27.624 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:24:27.624 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key3 00:24:27.624 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:27.624 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:27.624 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:27.624 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:24:27.624 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:24:27.624 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:24:27.883 00:24:27.883 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:24:27.883 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:24:27.883 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:24:28.140 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:28.140 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:24:28.140 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:28.140 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:28.140 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:28.140 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:24:28.140 { 00:24:28.140 "cntlid": 103, 00:24:28.140 "qid": 0, 00:24:28.140 "state": "enabled", 00:24:28.140 "thread": "nvmf_tgt_poll_group_000", 00:24:28.140 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:24:28.140 "listen_address": { 00:24:28.140 "trtype": "TCP", 00:24:28.140 "adrfam": "IPv4", 00:24:28.140 "traddr": "10.0.0.2", 00:24:28.140 "trsvcid": "4420" 00:24:28.140 }, 00:24:28.140 "peer_address": { 00:24:28.140 "trtype": "TCP", 00:24:28.140 "adrfam": "IPv4", 00:24:28.140 "traddr": "10.0.0.1", 00:24:28.140 "trsvcid": "44934" 00:24:28.140 }, 00:24:28.140 "auth": { 00:24:28.140 "state": "completed", 00:24:28.140 "digest": "sha512", 00:24:28.140 "dhgroup": "null" 00:24:28.140 } 00:24:28.140 } 00:24:28.140 ]' 00:24:28.141 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:24:28.141 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:24:28.141 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:24:28.141 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:24:28.141 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:24:28.398 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:24:28.398 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:24:28.398 17:40:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:24:28.656 17:40:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:24:28.656 17:40:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:24:29.590 17:40:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:24:29.590 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:24:29.590 17:40:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:24:29.590 17:40:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:29.590 17:40:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:29.590 17:40:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:29.590 17:40:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:24:29.590 17:40:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:24:29.590 17:40:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:24:29.590 17:40:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:24:29.590 17:40:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe2048 0 00:24:29.590 17:40:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:24:29.590 17:40:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:24:29.590 17:40:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:24:29.590 17:40:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:24:29.590 17:40:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:24:29.590 17:40:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:24:29.590 17:40:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:29.590 17:40:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:29.590 17:40:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:29.590 17:40:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:24:29.590 17:40:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:24:29.590 17:40:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:24:30.155 00:24:30.155 17:40:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:24:30.155 17:40:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:24:30.155 17:40:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:24:30.413 17:40:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:30.413 17:40:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:24:30.413 17:40:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:30.413 17:40:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:30.413 17:40:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:30.413 17:40:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:24:30.413 { 00:24:30.413 "cntlid": 105, 00:24:30.413 "qid": 0, 00:24:30.413 "state": "enabled", 00:24:30.413 "thread": "nvmf_tgt_poll_group_000", 00:24:30.413 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:24:30.413 "listen_address": { 00:24:30.413 "trtype": "TCP", 00:24:30.413 "adrfam": "IPv4", 00:24:30.413 "traddr": "10.0.0.2", 00:24:30.413 "trsvcid": "4420" 00:24:30.413 }, 00:24:30.413 "peer_address": { 00:24:30.413 "trtype": "TCP", 00:24:30.413 "adrfam": "IPv4", 00:24:30.413 "traddr": "10.0.0.1", 00:24:30.413 "trsvcid": "44946" 00:24:30.413 }, 00:24:30.413 "auth": { 00:24:30.413 "state": "completed", 00:24:30.413 "digest": "sha512", 00:24:30.413 "dhgroup": "ffdhe2048" 00:24:30.413 } 00:24:30.413 } 00:24:30.413 ]' 00:24:30.413 17:40:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:24:30.413 17:40:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:24:30.413 17:40:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:24:30.413 17:40:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:24:30.413 17:40:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:24:30.413 17:40:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:24:30.413 17:40:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:24:30.413 17:40:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:24:30.671 17:40:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:24:30.671 17:40:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:24:31.621 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:24:31.621 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:24:31.621 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:24:31.621 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:31.621 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:31.621 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:31.621 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:24:31.621 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:24:31.621 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:24:31.878 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe2048 1 00:24:31.878 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:24:31.878 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:24:31.878 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:24:31.878 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:24:31.878 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:24:31.878 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:24:31.878 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:31.878 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:31.878 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:31.878 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:24:31.878 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:24:31.878 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:24:32.135 00:24:32.135 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:24:32.135 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:24:32.135 17:40:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:24:32.393 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:32.393 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:24:32.393 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:32.393 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:32.393 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:32.393 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:24:32.393 { 00:24:32.393 "cntlid": 107, 00:24:32.393 "qid": 0, 00:24:32.393 "state": "enabled", 00:24:32.393 "thread": "nvmf_tgt_poll_group_000", 00:24:32.393 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:24:32.393 "listen_address": { 00:24:32.393 "trtype": "TCP", 00:24:32.393 "adrfam": "IPv4", 00:24:32.393 "traddr": "10.0.0.2", 00:24:32.393 "trsvcid": "4420" 00:24:32.393 }, 00:24:32.393 "peer_address": { 00:24:32.393 "trtype": "TCP", 00:24:32.393 "adrfam": "IPv4", 00:24:32.393 "traddr": "10.0.0.1", 00:24:32.393 "trsvcid": "44960" 00:24:32.393 }, 00:24:32.393 "auth": { 00:24:32.393 "state": "completed", 00:24:32.393 "digest": "sha512", 00:24:32.393 "dhgroup": "ffdhe2048" 00:24:32.393 } 00:24:32.393 } 00:24:32.393 ]' 00:24:32.393 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:24:32.651 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:24:32.651 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:24:32.651 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:24:32.651 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:24:32.651 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:24:32.651 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:24:32.651 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:24:32.909 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:24:32.909 17:40:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:24:33.841 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:24:33.841 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:24:33.841 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:24:33.841 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:33.841 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:33.841 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:33.841 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:24:33.841 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:24:33.841 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:24:34.098 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe2048 2 00:24:34.098 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:24:34.098 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:24:34.098 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:24:34.098 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:24:34.098 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:24:34.098 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:24:34.098 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:34.098 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:34.098 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:34.098 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:24:34.098 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:24:34.098 17:40:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:24:34.355 00:24:34.355 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:24:34.355 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:24:34.355 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:24:34.613 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:34.613 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:24:34.614 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:34.614 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:34.614 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:34.614 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:24:34.614 { 00:24:34.614 "cntlid": 109, 00:24:34.614 "qid": 0, 00:24:34.614 "state": "enabled", 00:24:34.614 "thread": "nvmf_tgt_poll_group_000", 00:24:34.614 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:24:34.614 "listen_address": { 00:24:34.614 "trtype": "TCP", 00:24:34.614 "adrfam": "IPv4", 00:24:34.614 "traddr": "10.0.0.2", 00:24:34.614 "trsvcid": "4420" 00:24:34.614 }, 00:24:34.614 "peer_address": { 00:24:34.614 "trtype": "TCP", 00:24:34.614 "adrfam": "IPv4", 00:24:34.614 "traddr": "10.0.0.1", 00:24:34.614 "trsvcid": "51190" 00:24:34.614 }, 00:24:34.614 "auth": { 00:24:34.614 "state": "completed", 00:24:34.614 "digest": "sha512", 00:24:34.614 "dhgroup": "ffdhe2048" 00:24:34.614 } 00:24:34.614 } 00:24:34.614 ]' 00:24:34.614 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:24:34.871 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:24:34.871 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:24:34.871 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:24:34.871 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:24:34.871 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:24:34.871 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:24:34.871 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:24:35.128 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:24:35.128 17:40:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:24:36.101 17:40:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:24:36.101 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:24:36.101 17:40:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:24:36.101 17:40:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:36.101 17:40:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:36.101 17:40:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:36.101 17:40:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:24:36.101 17:40:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:24:36.101 17:40:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:24:36.358 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe2048 3 00:24:36.358 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:24:36.358 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:24:36.358 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:24:36.358 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:24:36.358 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:24:36.358 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key3 00:24:36.358 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:36.358 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:36.358 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:36.358 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:24:36.358 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:24:36.358 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:24:36.615 00:24:36.615 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:24:36.615 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:24:36.615 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:24:36.872 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:36.872 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:24:36.872 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:36.872 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:36.872 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:36.872 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:24:36.872 { 00:24:36.872 "cntlid": 111, 00:24:36.872 "qid": 0, 00:24:36.872 "state": "enabled", 00:24:36.872 "thread": "nvmf_tgt_poll_group_000", 00:24:36.872 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:24:36.872 "listen_address": { 00:24:36.872 "trtype": "TCP", 00:24:36.872 "adrfam": "IPv4", 00:24:36.872 "traddr": "10.0.0.2", 00:24:36.872 "trsvcid": "4420" 00:24:36.872 }, 00:24:36.872 "peer_address": { 00:24:36.872 "trtype": "TCP", 00:24:36.872 "adrfam": "IPv4", 00:24:36.872 "traddr": "10.0.0.1", 00:24:36.872 "trsvcid": "51230" 00:24:36.872 }, 00:24:36.872 "auth": { 00:24:36.872 "state": "completed", 00:24:36.872 "digest": "sha512", 00:24:36.872 "dhgroup": "ffdhe2048" 00:24:36.872 } 00:24:36.872 } 00:24:36.872 ]' 00:24:36.872 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:24:36.872 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:24:36.872 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:24:37.129 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:24:37.129 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:24:37.129 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:24:37.129 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:24:37.129 17:40:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:24:37.386 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:24:37.386 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:24:38.317 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:24:38.317 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:24:38.317 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:24:38.317 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:38.317 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:38.317 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:38.317 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:24:38.317 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:24:38.317 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:24:38.317 17:40:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:24:38.575 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe3072 0 00:24:38.575 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:24:38.575 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:24:38.575 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:24:38.575 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:24:38.575 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:24:38.575 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:24:38.575 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:38.575 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:38.575 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:38.575 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:24:38.575 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:24:38.575 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:24:38.833 00:24:38.833 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:24:38.833 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:24:38.833 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:24:39.091 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:39.091 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:24:39.091 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:39.091 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:39.091 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:39.091 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:24:39.091 { 00:24:39.091 "cntlid": 113, 00:24:39.091 "qid": 0, 00:24:39.091 "state": "enabled", 00:24:39.091 "thread": "nvmf_tgt_poll_group_000", 00:24:39.091 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:24:39.091 "listen_address": { 00:24:39.091 "trtype": "TCP", 00:24:39.091 "adrfam": "IPv4", 00:24:39.091 "traddr": "10.0.0.2", 00:24:39.091 "trsvcid": "4420" 00:24:39.091 }, 00:24:39.091 "peer_address": { 00:24:39.091 "trtype": "TCP", 00:24:39.091 "adrfam": "IPv4", 00:24:39.091 "traddr": "10.0.0.1", 00:24:39.091 "trsvcid": "51252" 00:24:39.091 }, 00:24:39.091 "auth": { 00:24:39.091 "state": "completed", 00:24:39.091 "digest": "sha512", 00:24:39.091 "dhgroup": "ffdhe3072" 00:24:39.091 } 00:24:39.091 } 00:24:39.091 ]' 00:24:39.091 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:24:39.091 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:24:39.091 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:24:39.091 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:24:39.091 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:24:39.348 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:24:39.348 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:24:39.348 17:40:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:24:39.606 17:40:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:24:39.606 17:40:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:24:40.538 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:24:40.538 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:24:40.538 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:24:40.538 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:40.538 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:40.538 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:40.538 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:24:40.538 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:24:40.538 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:24:40.796 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe3072 1 00:24:40.796 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:24:40.796 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:24:40.796 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:24:40.796 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:24:40.796 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:24:40.796 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:24:40.796 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:40.796 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:40.796 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:40.796 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:24:40.796 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:24:40.796 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:24:41.055 00:24:41.055 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:24:41.055 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:24:41.055 17:40:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:24:41.312 17:40:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:41.312 17:40:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:24:41.312 17:40:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:41.312 17:40:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:41.312 17:40:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:41.312 17:40:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:24:41.312 { 00:24:41.312 "cntlid": 115, 00:24:41.312 "qid": 0, 00:24:41.312 "state": "enabled", 00:24:41.312 "thread": "nvmf_tgt_poll_group_000", 00:24:41.312 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:24:41.312 "listen_address": { 00:24:41.312 "trtype": "TCP", 00:24:41.312 "adrfam": "IPv4", 00:24:41.312 "traddr": "10.0.0.2", 00:24:41.312 "trsvcid": "4420" 00:24:41.312 }, 00:24:41.312 "peer_address": { 00:24:41.312 "trtype": "TCP", 00:24:41.312 "adrfam": "IPv4", 00:24:41.312 "traddr": "10.0.0.1", 00:24:41.312 "trsvcid": "51270" 00:24:41.312 }, 00:24:41.312 "auth": { 00:24:41.312 "state": "completed", 00:24:41.312 "digest": "sha512", 00:24:41.312 "dhgroup": "ffdhe3072" 00:24:41.312 } 00:24:41.312 } 00:24:41.312 ]' 00:24:41.312 17:40:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:24:41.312 17:40:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:24:41.312 17:40:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:24:41.313 17:40:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:24:41.313 17:40:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:24:41.570 17:40:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:24:41.570 17:40:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:24:41.570 17:40:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:24:41.828 17:40:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:24:41.828 17:40:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:24:42.762 17:40:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:24:42.762 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:24:42.763 17:40:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:24:42.763 17:40:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:42.763 17:40:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:42.763 17:40:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:42.763 17:40:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:24:42.763 17:40:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:24:42.763 17:40:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:24:43.021 17:40:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe3072 2 00:24:43.021 17:40:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:24:43.021 17:40:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:24:43.021 17:40:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:24:43.021 17:40:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:24:43.021 17:40:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:24:43.021 17:40:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:24:43.021 17:40:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:43.021 17:40:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:43.021 17:40:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:43.021 17:40:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:24:43.021 17:40:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:24:43.021 17:40:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:24:43.279 00:24:43.279 17:40:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:24:43.279 17:40:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:24:43.279 17:40:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:24:43.843 17:40:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:43.843 17:40:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:24:43.843 17:40:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:43.843 17:40:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:43.843 17:40:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:43.843 17:40:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:24:43.843 { 00:24:43.843 "cntlid": 117, 00:24:43.843 "qid": 0, 00:24:43.843 "state": "enabled", 00:24:43.843 "thread": "nvmf_tgt_poll_group_000", 00:24:43.843 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:24:43.843 "listen_address": { 00:24:43.843 "trtype": "TCP", 00:24:43.843 "adrfam": "IPv4", 00:24:43.843 "traddr": "10.0.0.2", 00:24:43.843 "trsvcid": "4420" 00:24:43.843 }, 00:24:43.843 "peer_address": { 00:24:43.843 "trtype": "TCP", 00:24:43.843 "adrfam": "IPv4", 00:24:43.843 "traddr": "10.0.0.1", 00:24:43.843 "trsvcid": "51298" 00:24:43.843 }, 00:24:43.843 "auth": { 00:24:43.843 "state": "completed", 00:24:43.843 "digest": "sha512", 00:24:43.843 "dhgroup": "ffdhe3072" 00:24:43.843 } 00:24:43.843 } 00:24:43.843 ]' 00:24:43.843 17:40:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:24:43.843 17:40:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:24:43.843 17:40:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:24:43.843 17:40:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:24:43.843 17:40:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:24:43.843 17:40:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:24:43.843 17:40:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:24:43.843 17:40:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:24:44.101 17:40:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:24:44.101 17:40:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:24:45.033 17:40:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:24:45.033 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:24:45.033 17:40:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:24:45.033 17:40:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:45.033 17:40:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:45.033 17:40:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:45.033 17:40:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:24:45.033 17:40:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:24:45.033 17:40:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:24:45.292 17:40:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe3072 3 00:24:45.292 17:40:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:24:45.292 17:40:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:24:45.292 17:40:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:24:45.292 17:40:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:24:45.292 17:40:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:24:45.292 17:40:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key3 00:24:45.292 17:40:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:45.292 17:40:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:45.292 17:40:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:45.292 17:40:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:24:45.292 17:40:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:24:45.292 17:40:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:24:45.550 00:24:45.550 17:40:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:24:45.550 17:40:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:24:45.550 17:40:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:24:45.807 17:40:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:45.807 17:40:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:24:45.807 17:40:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:45.807 17:40:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:45.807 17:40:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:45.808 17:40:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:24:45.808 { 00:24:45.808 "cntlid": 119, 00:24:45.808 "qid": 0, 00:24:45.808 "state": "enabled", 00:24:45.808 "thread": "nvmf_tgt_poll_group_000", 00:24:45.808 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:24:45.808 "listen_address": { 00:24:45.808 "trtype": "TCP", 00:24:45.808 "adrfam": "IPv4", 00:24:45.808 "traddr": "10.0.0.2", 00:24:45.808 "trsvcid": "4420" 00:24:45.808 }, 00:24:45.808 "peer_address": { 00:24:45.808 "trtype": "TCP", 00:24:45.808 "adrfam": "IPv4", 00:24:45.808 "traddr": "10.0.0.1", 00:24:45.808 "trsvcid": "42176" 00:24:45.808 }, 00:24:45.808 "auth": { 00:24:45.808 "state": "completed", 00:24:45.808 "digest": "sha512", 00:24:45.808 "dhgroup": "ffdhe3072" 00:24:45.808 } 00:24:45.808 } 00:24:45.808 ]' 00:24:45.808 17:40:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:24:45.808 17:40:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:24:45.808 17:40:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:24:46.065 17:40:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:24:46.065 17:40:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:24:46.065 17:40:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:24:46.065 17:40:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:24:46.065 17:40:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:24:46.323 17:40:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:24:46.323 17:40:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:24:47.254 17:40:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:24:47.254 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:24:47.254 17:40:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:24:47.254 17:40:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:47.254 17:40:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:47.254 17:40:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:47.254 17:40:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:24:47.254 17:40:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:24:47.254 17:40:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:24:47.254 17:40:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:24:47.510 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe4096 0 00:24:47.510 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:24:47.510 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:24:47.510 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:24:47.510 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:24:47.510 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:24:47.510 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:24:47.510 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:47.510 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:47.510 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:47.510 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:24:47.510 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:24:47.510 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:24:47.766 00:24:47.766 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:24:47.766 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:24:47.766 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:24:48.023 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:48.023 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:24:48.023 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:48.023 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:48.023 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:48.023 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:24:48.023 { 00:24:48.023 "cntlid": 121, 00:24:48.023 "qid": 0, 00:24:48.023 "state": "enabled", 00:24:48.023 "thread": "nvmf_tgt_poll_group_000", 00:24:48.023 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:24:48.023 "listen_address": { 00:24:48.023 "trtype": "TCP", 00:24:48.023 "adrfam": "IPv4", 00:24:48.023 "traddr": "10.0.0.2", 00:24:48.023 "trsvcid": "4420" 00:24:48.023 }, 00:24:48.023 "peer_address": { 00:24:48.023 "trtype": "TCP", 00:24:48.024 "adrfam": "IPv4", 00:24:48.024 "traddr": "10.0.0.1", 00:24:48.024 "trsvcid": "42198" 00:24:48.024 }, 00:24:48.024 "auth": { 00:24:48.024 "state": "completed", 00:24:48.024 "digest": "sha512", 00:24:48.024 "dhgroup": "ffdhe4096" 00:24:48.024 } 00:24:48.024 } 00:24:48.024 ]' 00:24:48.024 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:24:48.024 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:24:48.024 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:24:48.281 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:24:48.281 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:24:48.281 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:24:48.281 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:24:48.281 17:40:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:24:48.538 17:40:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:24:48.539 17:40:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:24:49.469 17:40:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:24:49.469 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:24:49.469 17:40:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:24:49.469 17:40:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:49.469 17:40:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:49.469 17:40:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:49.469 17:40:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:24:49.469 17:40:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:24:49.469 17:40:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:24:49.726 17:40:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe4096 1 00:24:49.726 17:40:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:24:49.726 17:40:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:24:49.726 17:40:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:24:49.726 17:40:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:24:49.726 17:40:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:24:49.726 17:40:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:24:49.726 17:40:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:49.726 17:40:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:49.726 17:40:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:49.726 17:40:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:24:49.726 17:40:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:24:49.726 17:40:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:24:50.290 00:24:50.290 17:40:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:24:50.290 17:40:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:24:50.290 17:40:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:24:50.547 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:50.547 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:24:50.547 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:50.547 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:50.547 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:50.547 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:24:50.547 { 00:24:50.547 "cntlid": 123, 00:24:50.547 "qid": 0, 00:24:50.547 "state": "enabled", 00:24:50.547 "thread": "nvmf_tgt_poll_group_000", 00:24:50.547 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:24:50.547 "listen_address": { 00:24:50.547 "trtype": "TCP", 00:24:50.547 "adrfam": "IPv4", 00:24:50.547 "traddr": "10.0.0.2", 00:24:50.547 "trsvcid": "4420" 00:24:50.547 }, 00:24:50.547 "peer_address": { 00:24:50.547 "trtype": "TCP", 00:24:50.547 "adrfam": "IPv4", 00:24:50.547 "traddr": "10.0.0.1", 00:24:50.547 "trsvcid": "42226" 00:24:50.547 }, 00:24:50.547 "auth": { 00:24:50.547 "state": "completed", 00:24:50.547 "digest": "sha512", 00:24:50.547 "dhgroup": "ffdhe4096" 00:24:50.547 } 00:24:50.547 } 00:24:50.547 ]' 00:24:50.547 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:24:50.547 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:24:50.547 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:24:50.547 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:24:50.547 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:24:50.547 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:24:50.547 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:24:50.547 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:24:50.804 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:24:50.804 17:40:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:24:51.734 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:24:51.734 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:24:51.734 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:24:51.734 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:51.734 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:51.734 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:51.734 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:24:51.734 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:24:51.734 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:24:51.992 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe4096 2 00:24:51.992 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:24:51.992 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:24:51.992 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:24:51.992 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:24:51.992 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:24:51.992 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:24:51.992 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:51.992 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:51.992 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:51.992 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:24:51.992 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:24:51.992 17:40:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:24:52.557 00:24:52.557 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:24:52.557 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:24:52.557 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:24:52.815 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:52.815 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:24:52.815 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:52.815 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:52.815 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:52.815 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:24:52.815 { 00:24:52.815 "cntlid": 125, 00:24:52.815 "qid": 0, 00:24:52.815 "state": "enabled", 00:24:52.815 "thread": "nvmf_tgt_poll_group_000", 00:24:52.815 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:24:52.815 "listen_address": { 00:24:52.815 "trtype": "TCP", 00:24:52.815 "adrfam": "IPv4", 00:24:52.815 "traddr": "10.0.0.2", 00:24:52.815 "trsvcid": "4420" 00:24:52.815 }, 00:24:52.815 "peer_address": { 00:24:52.815 "trtype": "TCP", 00:24:52.815 "adrfam": "IPv4", 00:24:52.815 "traddr": "10.0.0.1", 00:24:52.815 "trsvcid": "42242" 00:24:52.815 }, 00:24:52.815 "auth": { 00:24:52.815 "state": "completed", 00:24:52.815 "digest": "sha512", 00:24:52.815 "dhgroup": "ffdhe4096" 00:24:52.815 } 00:24:52.815 } 00:24:52.815 ]' 00:24:52.815 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:24:52.815 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:24:52.815 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:24:52.815 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:24:52.815 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:24:52.815 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:24:52.815 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:24:52.815 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:24:53.099 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:24:53.099 17:40:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:24:54.042 17:40:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:24:54.042 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:24:54.042 17:40:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:24:54.042 17:40:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:54.042 17:40:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:54.042 17:40:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:54.042 17:40:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:24:54.042 17:40:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:24:54.042 17:40:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:24:54.299 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe4096 3 00:24:54.299 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:24:54.299 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:24:54.299 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:24:54.299 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:24:54.299 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:24:54.299 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key3 00:24:54.299 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:54.299 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:54.299 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:54.299 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:24:54.299 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:24:54.300 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:24:54.558 00:24:54.558 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:24:54.558 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:24:54.558 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:24:55.124 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:55.124 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:24:55.124 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:55.124 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:55.124 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:55.124 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:24:55.124 { 00:24:55.124 "cntlid": 127, 00:24:55.124 "qid": 0, 00:24:55.124 "state": "enabled", 00:24:55.124 "thread": "nvmf_tgt_poll_group_000", 00:24:55.124 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:24:55.124 "listen_address": { 00:24:55.124 "trtype": "TCP", 00:24:55.124 "adrfam": "IPv4", 00:24:55.124 "traddr": "10.0.0.2", 00:24:55.124 "trsvcid": "4420" 00:24:55.124 }, 00:24:55.124 "peer_address": { 00:24:55.124 "trtype": "TCP", 00:24:55.124 "adrfam": "IPv4", 00:24:55.124 "traddr": "10.0.0.1", 00:24:55.124 "trsvcid": "43258" 00:24:55.124 }, 00:24:55.124 "auth": { 00:24:55.124 "state": "completed", 00:24:55.124 "digest": "sha512", 00:24:55.124 "dhgroup": "ffdhe4096" 00:24:55.124 } 00:24:55.124 } 00:24:55.124 ]' 00:24:55.124 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:24:55.124 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:24:55.124 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:24:55.124 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:24:55.124 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:24:55.124 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:24:55.124 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:24:55.124 17:40:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:24:55.382 17:40:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:24:55.382 17:40:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:24:56.430 17:40:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:24:56.430 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:24:56.430 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:24:56.430 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:56.430 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:56.430 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:56.430 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:24:56.430 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:24:56.430 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:24:56.430 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:24:56.743 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe6144 0 00:24:56.743 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:24:56.743 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:24:56.743 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:24:56.743 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:24:56.743 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:24:56.743 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:24:56.743 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:56.743 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:56.743 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:56.743 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:24:56.743 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:24:56.743 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:24:57.335 00:24:57.335 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:24:57.335 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:24:57.335 17:40:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:24:57.592 17:40:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:57.592 17:40:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:24:57.592 17:40:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:57.592 17:40:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:57.592 17:40:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:57.592 17:40:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:24:57.592 { 00:24:57.592 "cntlid": 129, 00:24:57.592 "qid": 0, 00:24:57.592 "state": "enabled", 00:24:57.592 "thread": "nvmf_tgt_poll_group_000", 00:24:57.592 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:24:57.592 "listen_address": { 00:24:57.592 "trtype": "TCP", 00:24:57.592 "adrfam": "IPv4", 00:24:57.592 "traddr": "10.0.0.2", 00:24:57.592 "trsvcid": "4420" 00:24:57.592 }, 00:24:57.592 "peer_address": { 00:24:57.592 "trtype": "TCP", 00:24:57.592 "adrfam": "IPv4", 00:24:57.592 "traddr": "10.0.0.1", 00:24:57.592 "trsvcid": "43274" 00:24:57.592 }, 00:24:57.592 "auth": { 00:24:57.592 "state": "completed", 00:24:57.592 "digest": "sha512", 00:24:57.592 "dhgroup": "ffdhe6144" 00:24:57.592 } 00:24:57.592 } 00:24:57.592 ]' 00:24:57.592 17:40:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:24:57.592 17:40:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:24:57.592 17:40:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:24:57.592 17:40:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:24:57.592 17:40:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:24:57.592 17:40:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:24:57.592 17:40:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:24:57.592 17:40:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:24:57.849 17:40:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:24:57.849 17:40:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:24:58.782 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:24:58.782 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:24:58.782 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:24:58.782 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:58.782 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:58.782 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:58.782 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:24:58.782 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:24:58.782 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:24:59.039 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe6144 1 00:24:59.039 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:24:59.039 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:24:59.039 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:24:59.039 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:24:59.039 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:24:59.039 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:24:59.039 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:59.040 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:59.040 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:59.040 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:24:59.040 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:24:59.040 17:40:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:24:59.604 00:24:59.604 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:24:59.604 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:24:59.604 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:24:59.862 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:59.862 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:24:59.862 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:24:59.862 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:24:59.862 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:24:59.862 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:24:59.862 { 00:24:59.862 "cntlid": 131, 00:24:59.862 "qid": 0, 00:24:59.862 "state": "enabled", 00:24:59.862 "thread": "nvmf_tgt_poll_group_000", 00:24:59.862 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:24:59.862 "listen_address": { 00:24:59.862 "trtype": "TCP", 00:24:59.862 "adrfam": "IPv4", 00:24:59.862 "traddr": "10.0.0.2", 00:24:59.862 "trsvcid": "4420" 00:24:59.862 }, 00:24:59.862 "peer_address": { 00:24:59.862 "trtype": "TCP", 00:24:59.862 "adrfam": "IPv4", 00:24:59.862 "traddr": "10.0.0.1", 00:24:59.862 "trsvcid": "43312" 00:24:59.862 }, 00:24:59.862 "auth": { 00:24:59.862 "state": "completed", 00:24:59.862 "digest": "sha512", 00:24:59.862 "dhgroup": "ffdhe6144" 00:24:59.862 } 00:24:59.862 } 00:24:59.862 ]' 00:24:59.862 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:25:00.119 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:25:00.119 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:25:00.119 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:25:00.119 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:25:00.119 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:25:00.119 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:25:00.119 17:40:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:25:00.377 17:40:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:25:00.377 17:40:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:25:01.310 17:40:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:25:01.310 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:25:01.310 17:40:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:25:01.310 17:40:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:01.310 17:40:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:01.310 17:40:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:01.310 17:40:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:25:01.310 17:40:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:25:01.310 17:40:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:25:01.568 17:40:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe6144 2 00:25:01.568 17:40:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:25:01.568 17:40:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:25:01.568 17:40:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:25:01.568 17:40:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:25:01.568 17:40:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:25:01.568 17:40:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:25:01.568 17:40:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:01.568 17:40:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:01.568 17:40:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:01.568 17:40:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:25:01.568 17:40:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:25:01.568 17:40:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:25:02.137 00:25:02.137 17:40:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:25:02.137 17:40:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:25:02.137 17:40:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:25:02.393 17:40:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:02.393 17:40:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:25:02.393 17:40:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:02.393 17:40:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:02.393 17:40:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:02.393 17:40:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:25:02.393 { 00:25:02.393 "cntlid": 133, 00:25:02.393 "qid": 0, 00:25:02.393 "state": "enabled", 00:25:02.393 "thread": "nvmf_tgt_poll_group_000", 00:25:02.393 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:25:02.393 "listen_address": { 00:25:02.393 "trtype": "TCP", 00:25:02.393 "adrfam": "IPv4", 00:25:02.393 "traddr": "10.0.0.2", 00:25:02.393 "trsvcid": "4420" 00:25:02.393 }, 00:25:02.393 "peer_address": { 00:25:02.393 "trtype": "TCP", 00:25:02.393 "adrfam": "IPv4", 00:25:02.393 "traddr": "10.0.0.1", 00:25:02.393 "trsvcid": "43348" 00:25:02.393 }, 00:25:02.393 "auth": { 00:25:02.394 "state": "completed", 00:25:02.394 "digest": "sha512", 00:25:02.394 "dhgroup": "ffdhe6144" 00:25:02.394 } 00:25:02.394 } 00:25:02.394 ]' 00:25:02.394 17:40:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:25:02.394 17:40:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:25:02.394 17:40:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:25:02.394 17:40:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:25:02.394 17:40:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:25:02.651 17:40:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:25:02.651 17:40:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:25:02.651 17:40:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:25:02.909 17:40:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:25:02.909 17:40:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:25:03.861 17:40:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:25:03.861 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:25:03.861 17:40:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:25:03.861 17:40:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:03.861 17:40:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:03.861 17:40:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:03.861 17:40:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:25:03.861 17:40:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:25:03.861 17:40:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:25:04.117 17:40:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe6144 3 00:25:04.117 17:40:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:25:04.117 17:40:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:25:04.117 17:40:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:25:04.117 17:40:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:25:04.117 17:40:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:25:04.117 17:40:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key3 00:25:04.117 17:40:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:04.117 17:40:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:04.117 17:40:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:04.117 17:40:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:25:04.117 17:40:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:25:04.117 17:40:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:25:04.680 00:25:04.680 17:40:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:25:04.680 17:40:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:25:04.680 17:40:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:25:04.936 17:40:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:04.936 17:40:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:25:04.936 17:40:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:04.936 17:40:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:04.936 17:40:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:04.936 17:40:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:25:04.936 { 00:25:04.936 "cntlid": 135, 00:25:04.936 "qid": 0, 00:25:04.936 "state": "enabled", 00:25:04.936 "thread": "nvmf_tgt_poll_group_000", 00:25:04.936 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:25:04.936 "listen_address": { 00:25:04.936 "trtype": "TCP", 00:25:04.936 "adrfam": "IPv4", 00:25:04.936 "traddr": "10.0.0.2", 00:25:04.936 "trsvcid": "4420" 00:25:04.936 }, 00:25:04.936 "peer_address": { 00:25:04.936 "trtype": "TCP", 00:25:04.936 "adrfam": "IPv4", 00:25:04.936 "traddr": "10.0.0.1", 00:25:04.936 "trsvcid": "56174" 00:25:04.936 }, 00:25:04.936 "auth": { 00:25:04.936 "state": "completed", 00:25:04.936 "digest": "sha512", 00:25:04.936 "dhgroup": "ffdhe6144" 00:25:04.936 } 00:25:04.936 } 00:25:04.936 ]' 00:25:04.936 17:40:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:25:04.936 17:40:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:25:04.936 17:40:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:25:04.936 17:40:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:25:04.936 17:40:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:25:04.936 17:40:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:25:04.936 17:40:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:25:04.936 17:40:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:25:05.193 17:40:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:25:05.193 17:40:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:25:06.131 17:40:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:25:06.131 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:25:06.131 17:40:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:25:06.131 17:40:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:06.131 17:40:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:06.131 17:40:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:06.131 17:40:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:25:06.131 17:40:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:25:06.131 17:40:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:25:06.131 17:40:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:25:06.388 17:40:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe8192 0 00:25:06.388 17:40:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:25:06.388 17:40:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:25:06.388 17:40:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:25:06.388 17:40:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:25:06.388 17:40:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:25:06.388 17:40:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:25:06.389 17:40:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:06.389 17:40:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:06.389 17:40:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:06.389 17:40:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:25:06.389 17:40:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:25:06.389 17:40:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:25:07.320 00:25:07.320 17:40:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:25:07.320 17:40:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:25:07.320 17:40:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:25:07.578 17:40:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:07.578 17:40:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:25:07.578 17:40:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:07.578 17:40:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:07.578 17:40:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:07.578 17:40:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:25:07.578 { 00:25:07.578 "cntlid": 137, 00:25:07.578 "qid": 0, 00:25:07.578 "state": "enabled", 00:25:07.578 "thread": "nvmf_tgt_poll_group_000", 00:25:07.578 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:25:07.578 "listen_address": { 00:25:07.579 "trtype": "TCP", 00:25:07.579 "adrfam": "IPv4", 00:25:07.579 "traddr": "10.0.0.2", 00:25:07.579 "trsvcid": "4420" 00:25:07.579 }, 00:25:07.579 "peer_address": { 00:25:07.579 "trtype": "TCP", 00:25:07.579 "adrfam": "IPv4", 00:25:07.579 "traddr": "10.0.0.1", 00:25:07.579 "trsvcid": "56212" 00:25:07.579 }, 00:25:07.579 "auth": { 00:25:07.579 "state": "completed", 00:25:07.579 "digest": "sha512", 00:25:07.579 "dhgroup": "ffdhe8192" 00:25:07.579 } 00:25:07.579 } 00:25:07.579 ]' 00:25:07.579 17:40:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:25:07.579 17:40:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:25:07.579 17:40:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:25:07.579 17:40:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:25:07.579 17:40:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:25:07.579 17:40:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:25:07.579 17:40:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:25:07.579 17:40:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:25:07.837 17:40:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:25:07.837 17:40:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:25:09.210 17:40:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:25:09.210 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:25:09.210 17:40:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:25:09.210 17:40:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:09.210 17:40:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:09.210 17:40:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:09.210 17:40:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:25:09.210 17:40:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:25:09.210 17:40:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:25:09.210 17:40:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe8192 1 00:25:09.210 17:40:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:25:09.210 17:40:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:25:09.210 17:40:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:25:09.210 17:40:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:25:09.210 17:40:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:25:09.210 17:40:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:25:09.210 17:40:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:09.210 17:40:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:09.210 17:40:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:09.210 17:40:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:25:09.210 17:40:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:25:09.210 17:40:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:25:10.144 00:25:10.144 17:40:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:25:10.144 17:40:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:25:10.144 17:40:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:25:10.403 17:40:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:10.403 17:40:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:25:10.403 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:10.403 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:10.403 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:10.403 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:25:10.403 { 00:25:10.403 "cntlid": 139, 00:25:10.403 "qid": 0, 00:25:10.403 "state": "enabled", 00:25:10.403 "thread": "nvmf_tgt_poll_group_000", 00:25:10.403 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:25:10.403 "listen_address": { 00:25:10.403 "trtype": "TCP", 00:25:10.403 "adrfam": "IPv4", 00:25:10.403 "traddr": "10.0.0.2", 00:25:10.403 "trsvcid": "4420" 00:25:10.403 }, 00:25:10.403 "peer_address": { 00:25:10.403 "trtype": "TCP", 00:25:10.403 "adrfam": "IPv4", 00:25:10.403 "traddr": "10.0.0.1", 00:25:10.403 "trsvcid": "56244" 00:25:10.403 }, 00:25:10.403 "auth": { 00:25:10.403 "state": "completed", 00:25:10.403 "digest": "sha512", 00:25:10.403 "dhgroup": "ffdhe8192" 00:25:10.403 } 00:25:10.403 } 00:25:10.403 ]' 00:25:10.403 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:25:10.403 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:25:10.403 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:25:10.403 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:25:10.403 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:25:10.403 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:25:10.403 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:25:10.403 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:25:10.660 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:25:10.660 17:40:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: --dhchap-ctrl-secret DHHC-1:02:NWZiY2UzNmMzOTZlN2ZhNWY5ZTRmNWYyZGI4MTRlYmExYTIwY2I2OTFhNjZlYmE5LOr5FA==: 00:25:11.591 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:25:11.591 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:25:11.591 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:25:11.591 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:11.591 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:11.591 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:11.591 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:25:11.591 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:25:11.591 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:25:11.849 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe8192 2 00:25:11.849 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:25:11.849 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:25:11.849 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:25:11.849 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:25:11.849 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:25:11.849 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:25:11.849 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:11.849 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:11.849 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:11.849 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:25:11.849 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:25:11.849 17:40:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:25:12.782 00:25:12.782 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:25:12.782 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:25:12.782 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:25:13.040 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:13.040 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:25:13.040 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:13.040 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:13.040 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:13.040 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:25:13.040 { 00:25:13.040 "cntlid": 141, 00:25:13.040 "qid": 0, 00:25:13.040 "state": "enabled", 00:25:13.040 "thread": "nvmf_tgt_poll_group_000", 00:25:13.040 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:25:13.040 "listen_address": { 00:25:13.040 "trtype": "TCP", 00:25:13.040 "adrfam": "IPv4", 00:25:13.040 "traddr": "10.0.0.2", 00:25:13.040 "trsvcid": "4420" 00:25:13.040 }, 00:25:13.040 "peer_address": { 00:25:13.040 "trtype": "TCP", 00:25:13.040 "adrfam": "IPv4", 00:25:13.040 "traddr": "10.0.0.1", 00:25:13.040 "trsvcid": "56258" 00:25:13.040 }, 00:25:13.040 "auth": { 00:25:13.040 "state": "completed", 00:25:13.040 "digest": "sha512", 00:25:13.040 "dhgroup": "ffdhe8192" 00:25:13.040 } 00:25:13.040 } 00:25:13.040 ]' 00:25:13.040 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:25:13.040 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:25:13.040 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:25:13.040 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:25:13.040 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:25:13.040 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:25:13.040 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:25:13.040 17:40:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:25:13.297 17:40:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:25:13.297 17:40:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:01:NGQyNjhkYWQzMmE3ZGJmYmJiYmI4YjA5YzM0MTI0OWR+Afzu: 00:25:14.228 17:40:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:25:14.228 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:25:14.228 17:40:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:25:14.228 17:40:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:14.228 17:40:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:14.228 17:40:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:14.228 17:40:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:25:14.228 17:40:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:25:14.228 17:40:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:25:14.485 17:40:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe8192 3 00:25:14.485 17:40:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:25:14.485 17:40:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:25:14.485 17:40:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:25:14.485 17:40:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:25:14.485 17:40:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:25:14.485 17:40:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key3 00:25:14.485 17:40:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:14.485 17:40:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:14.485 17:40:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:14.485 17:40:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:25:14.485 17:40:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:25:14.485 17:40:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:25:15.417 00:25:15.417 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:25:15.417 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:25:15.417 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:25:15.674 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:15.674 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:25:15.674 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:15.674 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:15.674 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:15.674 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:25:15.674 { 00:25:15.674 "cntlid": 143, 00:25:15.674 "qid": 0, 00:25:15.674 "state": "enabled", 00:25:15.674 "thread": "nvmf_tgt_poll_group_000", 00:25:15.674 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:25:15.674 "listen_address": { 00:25:15.674 "trtype": "TCP", 00:25:15.674 "adrfam": "IPv4", 00:25:15.674 "traddr": "10.0.0.2", 00:25:15.674 "trsvcid": "4420" 00:25:15.674 }, 00:25:15.674 "peer_address": { 00:25:15.674 "trtype": "TCP", 00:25:15.674 "adrfam": "IPv4", 00:25:15.674 "traddr": "10.0.0.1", 00:25:15.674 "trsvcid": "48334" 00:25:15.674 }, 00:25:15.674 "auth": { 00:25:15.674 "state": "completed", 00:25:15.674 "digest": "sha512", 00:25:15.674 "dhgroup": "ffdhe8192" 00:25:15.674 } 00:25:15.674 } 00:25:15.674 ]' 00:25:15.674 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:25:15.674 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:25:15.674 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:25:15.674 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:25:15.674 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:25:15.674 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:25:15.674 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:25:15.674 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:25:15.930 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:25:15.930 17:40:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:25:16.860 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:25:16.860 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:25:16.860 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:25:16.860 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:16.860 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:16.860 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:16.860 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@129 -- # IFS=, 00:25:16.860 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@130 -- # printf %s sha256,sha384,sha512 00:25:16.860 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@129 -- # IFS=, 00:25:16.860 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@130 -- # printf %s null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:25:16.860 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@129 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256,sha384,sha512 --dhchap-dhgroups null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:25:16.860 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256,sha384,sha512 --dhchap-dhgroups null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:25:17.118 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@141 -- # connect_authenticate sha512 ffdhe8192 0 00:25:17.118 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:25:17.118 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:25:17.118 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:25:17.118 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:25:17.118 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:25:17.118 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:25:17.118 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:17.118 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:17.118 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:17.118 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:25:17.118 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:25:17.118 17:40:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:25:18.048 00:25:18.048 17:40:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:25:18.048 17:40:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:25:18.048 17:40:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:25:18.305 17:41:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:18.305 17:41:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:25:18.305 17:41:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:18.305 17:41:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:18.305 17:41:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:18.305 17:41:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:25:18.305 { 00:25:18.305 "cntlid": 145, 00:25:18.305 "qid": 0, 00:25:18.305 "state": "enabled", 00:25:18.305 "thread": "nvmf_tgt_poll_group_000", 00:25:18.305 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:25:18.305 "listen_address": { 00:25:18.305 "trtype": "TCP", 00:25:18.305 "adrfam": "IPv4", 00:25:18.305 "traddr": "10.0.0.2", 00:25:18.305 "trsvcid": "4420" 00:25:18.305 }, 00:25:18.305 "peer_address": { 00:25:18.305 "trtype": "TCP", 00:25:18.305 "adrfam": "IPv4", 00:25:18.305 "traddr": "10.0.0.1", 00:25:18.305 "trsvcid": "48360" 00:25:18.305 }, 00:25:18.305 "auth": { 00:25:18.305 "state": "completed", 00:25:18.305 "digest": "sha512", 00:25:18.305 "dhgroup": "ffdhe8192" 00:25:18.305 } 00:25:18.305 } 00:25:18.305 ]' 00:25:18.305 17:41:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:25:18.305 17:41:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:25:18.305 17:41:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:25:18.305 17:41:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:25:18.305 17:41:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:25:18.305 17:41:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:25:18.305 17:41:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:25:18.305 17:41:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:25:18.868 17:41:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:25:18.868 17:41:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:00:ZDRjNzVlYWYzMTQxMzg4MTAxYzg5MGFkNTM5ODZlNDI2ODdjZjAwNGQ2ZmMxYmY38UJlkQ==: --dhchap-ctrl-secret DHHC-1:03:MTU5ZWRkOGE5MDczNzU3OGJhMjk1Mjg3N2I2NzQ1ZjliZjVhZDZlZmVkOWE3YmVmYWU1YjUyM2YwYzQ0ZGI4Zc9+SiY=: 00:25:19.799 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:25:19.799 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:25:19.800 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:25:19.800 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:19.800 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:19.800 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:19.800 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@144 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key1 00:25:19.800 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:19.800 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:19.800 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:19.800 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@145 -- # NOT bdev_connect -b nvme0 --dhchap-key key2 00:25:19.800 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # local es=0 00:25:19.800 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@654 -- # valid_exec_arg bdev_connect -b nvme0 --dhchap-key key2 00:25:19.800 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@640 -- # local arg=bdev_connect 00:25:19.800 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:25:19.800 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # type -t bdev_connect 00:25:19.800 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:25:19.800 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # bdev_connect -b nvme0 --dhchap-key key2 00:25:19.800 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 00:25:19.800 17:41:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 00:25:20.365 request: 00:25:20.365 { 00:25:20.365 "name": "nvme0", 00:25:20.365 "trtype": "tcp", 00:25:20.365 "traddr": "10.0.0.2", 00:25:20.365 "adrfam": "ipv4", 00:25:20.365 "trsvcid": "4420", 00:25:20.365 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:25:20.365 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:25:20.365 "prchk_reftag": false, 00:25:20.365 "prchk_guard": false, 00:25:20.365 "hdgst": false, 00:25:20.365 "ddgst": false, 00:25:20.365 "dhchap_key": "key2", 00:25:20.365 "allow_unrecognized_csi": false, 00:25:20.365 "method": "bdev_nvme_attach_controller", 00:25:20.365 "req_id": 1 00:25:20.365 } 00:25:20.365 Got JSON-RPC error response 00:25:20.365 response: 00:25:20.365 { 00:25:20.365 "code": -5, 00:25:20.365 "message": "Input/output error" 00:25:20.365 } 00:25:20.365 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # es=1 00:25:20.365 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:25:20.365 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:25:20.365 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:25:20.365 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@146 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:25:20.365 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:20.365 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:20.365 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:20.365 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@149 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:25:20.365 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:20.365 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:20.365 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:20.365 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@150 -- # NOT bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:25:20.365 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # local es=0 00:25:20.365 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@654 -- # valid_exec_arg bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:25:20.365 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@640 -- # local arg=bdev_connect 00:25:20.365 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:25:20.365 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # type -t bdev_connect 00:25:20.365 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:25:20.365 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:25:20.365 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:25:20.365 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:25:21.298 request: 00:25:21.298 { 00:25:21.298 "name": "nvme0", 00:25:21.298 "trtype": "tcp", 00:25:21.298 "traddr": "10.0.0.2", 00:25:21.298 "adrfam": "ipv4", 00:25:21.298 "trsvcid": "4420", 00:25:21.298 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:25:21.298 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:25:21.298 "prchk_reftag": false, 00:25:21.298 "prchk_guard": false, 00:25:21.298 "hdgst": false, 00:25:21.298 "ddgst": false, 00:25:21.298 "dhchap_key": "key1", 00:25:21.298 "dhchap_ctrlr_key": "ckey2", 00:25:21.298 "allow_unrecognized_csi": false, 00:25:21.298 "method": "bdev_nvme_attach_controller", 00:25:21.298 "req_id": 1 00:25:21.298 } 00:25:21.298 Got JSON-RPC error response 00:25:21.298 response: 00:25:21.298 { 00:25:21.298 "code": -5, 00:25:21.298 "message": "Input/output error" 00:25:21.298 } 00:25:21.298 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # es=1 00:25:21.298 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:25:21.298 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:25:21.298 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:25:21.298 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@151 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:25:21.298 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:21.298 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:21.298 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:21.298 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@154 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key1 00:25:21.298 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:21.298 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:21.298 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:21.298 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@155 -- # NOT bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:25:21.298 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # local es=0 00:25:21.299 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@654 -- # valid_exec_arg bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:25:21.299 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@640 -- # local arg=bdev_connect 00:25:21.299 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:25:21.299 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # type -t bdev_connect 00:25:21.299 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:25:21.299 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:25:21.299 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:25:21.299 17:41:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:25:22.232 request: 00:25:22.232 { 00:25:22.232 "name": "nvme0", 00:25:22.232 "trtype": "tcp", 00:25:22.232 "traddr": "10.0.0.2", 00:25:22.232 "adrfam": "ipv4", 00:25:22.232 "trsvcid": "4420", 00:25:22.232 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:25:22.232 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:25:22.232 "prchk_reftag": false, 00:25:22.232 "prchk_guard": false, 00:25:22.232 "hdgst": false, 00:25:22.232 "ddgst": false, 00:25:22.232 "dhchap_key": "key1", 00:25:22.232 "dhchap_ctrlr_key": "ckey1", 00:25:22.232 "allow_unrecognized_csi": false, 00:25:22.232 "method": "bdev_nvme_attach_controller", 00:25:22.232 "req_id": 1 00:25:22.232 } 00:25:22.232 Got JSON-RPC error response 00:25:22.232 response: 00:25:22.232 { 00:25:22.232 "code": -5, 00:25:22.232 "message": "Input/output error" 00:25:22.232 } 00:25:22.232 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # es=1 00:25:22.232 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:25:22.232 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:25:22.232 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:25:22.232 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@156 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:25:22.232 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:22.232 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:22.232 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:22.232 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@159 -- # killprocess 251695 00:25:22.232 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@954 -- # '[' -z 251695 ']' 00:25:22.232 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@958 -- # kill -0 251695 00:25:22.232 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@959 -- # uname 00:25:22.232 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:25:22.232 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 251695 00:25:22.232 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:25:22.232 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:25:22.232 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@972 -- # echo 'killing process with pid 251695' 00:25:22.232 killing process with pid 251695 00:25:22.232 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@973 -- # kill 251695 00:25:22.232 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@978 -- # wait 251695 00:25:22.232 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@160 -- # nvmfappstart --wait-for-rpc -L nvmf_auth 00:25:22.232 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:25:22.232 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@726 -- # xtrace_disable 00:25:22.233 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:22.233 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@509 -- # nvmfpid=274677 00:25:22.233 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --wait-for-rpc -L nvmf_auth 00:25:22.233 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@510 -- # waitforlisten 274677 00:25:22.233 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@835 -- # '[' -z 274677 ']' 00:25:22.233 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:22.233 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@840 -- # local max_retries=100 00:25:22.233 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:22.233 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@844 -- # xtrace_disable 00:25:22.233 17:41:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:22.491 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:25:22.491 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@868 -- # return 0 00:25:22.491 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:25:22.491 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@732 -- # xtrace_disable 00:25:22.491 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:22.491 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:25:22.491 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@161 -- # trap 'dumplogs; cleanup' SIGINT SIGTERM EXIT 00:25:22.491 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@163 -- # waitforlisten 274677 00:25:22.491 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@835 -- # '[' -z 274677 ']' 00:25:22.491 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:22.491 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@840 -- # local max_retries=100 00:25:22.491 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:22.491 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:22.491 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@844 -- # xtrace_disable 00:25:22.491 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:22.749 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:25:22.749 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@868 -- # return 0 00:25:22.749 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@164 -- # rpc_cmd 00:25:22.749 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:22.749 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:23.007 null0 00:25:23.007 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:23.007 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@174 -- # for i in "${!keys[@]}" 00:25:23.007 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@175 -- # rpc_cmd keyring_file_add_key key0 /tmp/spdk.key-null.b1A 00:25:23.007 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:23.007 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:23.007 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:23.007 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@176 -- # [[ -n /tmp/spdk.key-sha512.7q2 ]] 00:25:23.007 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@176 -- # rpc_cmd keyring_file_add_key ckey0 /tmp/spdk.key-sha512.7q2 00:25:23.007 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@174 -- # for i in "${!keys[@]}" 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@175 -- # rpc_cmd keyring_file_add_key key1 /tmp/spdk.key-sha256.zDo 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@176 -- # [[ -n /tmp/spdk.key-sha384.OEt ]] 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@176 -- # rpc_cmd keyring_file_add_key ckey1 /tmp/spdk.key-sha384.OEt 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@174 -- # for i in "${!keys[@]}" 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@175 -- # rpc_cmd keyring_file_add_key key2 /tmp/spdk.key-sha384.TnN 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@176 -- # [[ -n /tmp/spdk.key-sha256.Avu ]] 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@176 -- # rpc_cmd keyring_file_add_key ckey2 /tmp/spdk.key-sha256.Avu 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@174 -- # for i in "${!keys[@]}" 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@175 -- # rpc_cmd keyring_file_add_key key3 /tmp/spdk.key-sha512.BLn 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@176 -- # [[ -n '' ]] 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@179 -- # connect_authenticate sha512 ffdhe8192 3 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key3 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:25:23.008 17:41:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:25:24.380 nvme0n1 00:25:24.380 17:41:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:25:24.380 17:41:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:25:24.380 17:41:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:25:24.638 17:41:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:24.638 17:41:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:25:24.638 17:41:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:24.638 17:41:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:24.638 17:41:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:24.638 17:41:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:25:24.638 { 00:25:24.638 "cntlid": 1, 00:25:24.638 "qid": 0, 00:25:24.638 "state": "enabled", 00:25:24.638 "thread": "nvmf_tgt_poll_group_000", 00:25:24.638 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:25:24.638 "listen_address": { 00:25:24.638 "trtype": "TCP", 00:25:24.638 "adrfam": "IPv4", 00:25:24.638 "traddr": "10.0.0.2", 00:25:24.638 "trsvcid": "4420" 00:25:24.638 }, 00:25:24.638 "peer_address": { 00:25:24.638 "trtype": "TCP", 00:25:24.638 "adrfam": "IPv4", 00:25:24.638 "traddr": "10.0.0.1", 00:25:24.638 "trsvcid": "48414" 00:25:24.638 }, 00:25:24.638 "auth": { 00:25:24.638 "state": "completed", 00:25:24.638 "digest": "sha512", 00:25:24.638 "dhgroup": "ffdhe8192" 00:25:24.638 } 00:25:24.638 } 00:25:24.638 ]' 00:25:24.638 17:41:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:25:24.638 17:41:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:25:24.638 17:41:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:25:24.895 17:41:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:25:24.895 17:41:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:25:24.895 17:41:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:25:24.895 17:41:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:25:24.895 17:41:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:25:25.153 17:41:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:25:25.153 17:41:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:25:26.087 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:25:26.087 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:25:26.087 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:25:26.087 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:26.087 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:26.087 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:26.087 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@182 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key3 00:25:26.087 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:26.087 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:26.087 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:26.087 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@183 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 00:25:26.087 17:41:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 00:25:26.345 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@184 -- # NOT bdev_connect -b nvme0 --dhchap-key key3 00:25:26.345 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # local es=0 00:25:26.345 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@654 -- # valid_exec_arg bdev_connect -b nvme0 --dhchap-key key3 00:25:26.345 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@640 -- # local arg=bdev_connect 00:25:26.345 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:25:26.345 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # type -t bdev_connect 00:25:26.345 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:25:26.345 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # bdev_connect -b nvme0 --dhchap-key key3 00:25:26.345 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:25:26.345 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:25:26.603 request: 00:25:26.603 { 00:25:26.603 "name": "nvme0", 00:25:26.603 "trtype": "tcp", 00:25:26.603 "traddr": "10.0.0.2", 00:25:26.603 "adrfam": "ipv4", 00:25:26.603 "trsvcid": "4420", 00:25:26.603 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:25:26.603 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:25:26.603 "prchk_reftag": false, 00:25:26.603 "prchk_guard": false, 00:25:26.603 "hdgst": false, 00:25:26.603 "ddgst": false, 00:25:26.603 "dhchap_key": "key3", 00:25:26.603 "allow_unrecognized_csi": false, 00:25:26.603 "method": "bdev_nvme_attach_controller", 00:25:26.603 "req_id": 1 00:25:26.603 } 00:25:26.603 Got JSON-RPC error response 00:25:26.603 response: 00:25:26.603 { 00:25:26.603 "code": -5, 00:25:26.603 "message": "Input/output error" 00:25:26.603 } 00:25:26.603 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # es=1 00:25:26.603 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:25:26.603 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:25:26.603 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:25:26.603 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@187 -- # IFS=, 00:25:26.603 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@188 -- # printf %s sha256,sha384,sha512 00:25:26.603 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@187 -- # hostrpc bdev_nvme_set_options --dhchap-dhgroups ffdhe2048 --dhchap-digests sha256,sha384,sha512 00:25:26.603 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-dhgroups ffdhe2048 --dhchap-digests sha256,sha384,sha512 00:25:26.860 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@193 -- # NOT bdev_connect -b nvme0 --dhchap-key key3 00:25:26.860 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # local es=0 00:25:26.860 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@654 -- # valid_exec_arg bdev_connect -b nvme0 --dhchap-key key3 00:25:26.860 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@640 -- # local arg=bdev_connect 00:25:26.860 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:25:26.860 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # type -t bdev_connect 00:25:26.860 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:25:26.860 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # bdev_connect -b nvme0 --dhchap-key key3 00:25:26.860 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:25:26.860 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:25:27.118 request: 00:25:27.118 { 00:25:27.118 "name": "nvme0", 00:25:27.118 "trtype": "tcp", 00:25:27.118 "traddr": "10.0.0.2", 00:25:27.118 "adrfam": "ipv4", 00:25:27.118 "trsvcid": "4420", 00:25:27.118 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:25:27.118 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:25:27.118 "prchk_reftag": false, 00:25:27.118 "prchk_guard": false, 00:25:27.118 "hdgst": false, 00:25:27.118 "ddgst": false, 00:25:27.118 "dhchap_key": "key3", 00:25:27.118 "allow_unrecognized_csi": false, 00:25:27.118 "method": "bdev_nvme_attach_controller", 00:25:27.118 "req_id": 1 00:25:27.118 } 00:25:27.118 Got JSON-RPC error response 00:25:27.118 response: 00:25:27.118 { 00:25:27.118 "code": -5, 00:25:27.118 "message": "Input/output error" 00:25:27.118 } 00:25:27.118 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # es=1 00:25:27.118 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:25:27.118 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:25:27.118 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:25:27.118 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@197 -- # IFS=, 00:25:27.118 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@198 -- # printf %s sha256,sha384,sha512 00:25:27.118 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@197 -- # IFS=, 00:25:27.118 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@198 -- # printf %s null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:25:27.118 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@197 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256,sha384,sha512 --dhchap-dhgroups null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:25:27.118 17:41:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256,sha384,sha512 --dhchap-dhgroups null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:25:27.376 17:41:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@208 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:25:27.376 17:41:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:27.376 17:41:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:27.376 17:41:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:27.376 17:41:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@209 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:25:27.376 17:41:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:27.376 17:41:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:27.634 17:41:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:27.634 17:41:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@210 -- # NOT bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 00:25:27.634 17:41:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # local es=0 00:25:27.634 17:41:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@654 -- # valid_exec_arg bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 00:25:27.634 17:41:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@640 -- # local arg=bdev_connect 00:25:27.634 17:41:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:25:27.634 17:41:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # type -t bdev_connect 00:25:27.634 17:41:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:25:27.634 17:41:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 00:25:27.634 17:41:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 00:25:27.634 17:41:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 00:25:28.198 request: 00:25:28.198 { 00:25:28.198 "name": "nvme0", 00:25:28.198 "trtype": "tcp", 00:25:28.198 "traddr": "10.0.0.2", 00:25:28.198 "adrfam": "ipv4", 00:25:28.198 "trsvcid": "4420", 00:25:28.198 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:25:28.198 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:25:28.198 "prchk_reftag": false, 00:25:28.198 "prchk_guard": false, 00:25:28.198 "hdgst": false, 00:25:28.198 "ddgst": false, 00:25:28.198 "dhchap_key": "key0", 00:25:28.198 "dhchap_ctrlr_key": "key1", 00:25:28.198 "allow_unrecognized_csi": false, 00:25:28.198 "method": "bdev_nvme_attach_controller", 00:25:28.198 "req_id": 1 00:25:28.198 } 00:25:28.198 Got JSON-RPC error response 00:25:28.198 response: 00:25:28.198 { 00:25:28.198 "code": -5, 00:25:28.198 "message": "Input/output error" 00:25:28.198 } 00:25:28.198 17:41:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # es=1 00:25:28.198 17:41:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:25:28.198 17:41:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:25:28.198 17:41:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:25:28.199 17:41:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@213 -- # bdev_connect -b nvme0 --dhchap-key key0 00:25:28.199 17:41:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 00:25:28.199 17:41:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 00:25:28.456 nvme0n1 00:25:28.456 17:41:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@214 -- # hostrpc bdev_nvme_get_controllers 00:25:28.456 17:41:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@214 -- # jq -r '.[].name' 00:25:28.456 17:41:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:25:28.715 17:41:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@214 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:28.715 17:41:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@215 -- # hostrpc bdev_nvme_detach_controller nvme0 00:25:28.715 17:41:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:25:28.973 17:41:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@218 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key1 00:25:28.973 17:41:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:28.973 17:41:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:28.973 17:41:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:28.973 17:41:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@219 -- # bdev_connect -b nvme0 --dhchap-key key1 00:25:28.973 17:41:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 00:25:28.973 17:41:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 00:25:30.342 nvme0n1 00:25:30.342 17:41:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@220 -- # hostrpc bdev_nvme_get_controllers 00:25:30.342 17:41:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@220 -- # jq -r '.[].name' 00:25:30.342 17:41:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:25:30.600 17:41:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@220 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:30.600 17:41:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@222 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key2 --dhchap-ctrlr-key key3 00:25:30.600 17:41:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:30.600 17:41:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:30.600 17:41:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:30.600 17:41:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@223 -- # hostrpc bdev_nvme_get_controllers 00:25:30.600 17:41:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@223 -- # jq -r '.[].name' 00:25:30.600 17:41:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:25:30.858 17:41:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@223 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:30.858 17:41:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@225 -- # nvme_connect --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:25:30.858 17:41:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid 5b23e107-7094-e311-b1cb-001e67a97d55 -l 0 --dhchap-secret DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: --dhchap-ctrl-secret DHHC-1:03:NmZlMmQzYTFlNzA2MDAyNzg4MzllYzY4MWIxYTRmNGNkYjc4YTBjOWVkOTMxYjMxYjc5ZDg2ZWNjZGYxZTM2Ygwiw2U=: 00:25:31.788 17:41:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@226 -- # nvme_get_ctrlr 00:25:31.788 17:41:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@41 -- # local dev 00:25:31.788 17:41:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@43 -- # for dev in /sys/devices/virtual/nvme-fabrics/ctl/nvme* 00:25:31.788 17:41:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@44 -- # [[ nqn.2024-03.io.spdk:cnode0 == \n\q\n\.\2\0\2\4\-\0\3\.\i\o\.\s\p\d\k\:\c\n\o\d\e\0 ]] 00:25:31.788 17:41:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@44 -- # echo nvme0 00:25:31.788 17:41:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@44 -- # break 00:25:31.788 17:41:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@226 -- # nctrlr=nvme0 00:25:31.788 17:41:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@227 -- # hostrpc bdev_nvme_detach_controller nvme0 00:25:31.788 17:41:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:25:32.045 17:41:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@228 -- # NOT bdev_connect -b nvme0 --dhchap-key key1 00:25:32.045 17:41:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # local es=0 00:25:32.045 17:41:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@654 -- # valid_exec_arg bdev_connect -b nvme0 --dhchap-key key1 00:25:32.045 17:41:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@640 -- # local arg=bdev_connect 00:25:32.045 17:41:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:25:32.045 17:41:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # type -t bdev_connect 00:25:32.045 17:41:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:25:32.045 17:41:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # bdev_connect -b nvme0 --dhchap-key key1 00:25:32.045 17:41:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 00:25:32.045 17:41:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 00:25:32.978 request: 00:25:32.978 { 00:25:32.978 "name": "nvme0", 00:25:32.978 "trtype": "tcp", 00:25:32.978 "traddr": "10.0.0.2", 00:25:32.978 "adrfam": "ipv4", 00:25:32.978 "trsvcid": "4420", 00:25:32.978 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:25:32.978 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55", 00:25:32.978 "prchk_reftag": false, 00:25:32.978 "prchk_guard": false, 00:25:32.978 "hdgst": false, 00:25:32.978 "ddgst": false, 00:25:32.978 "dhchap_key": "key1", 00:25:32.978 "allow_unrecognized_csi": false, 00:25:32.978 "method": "bdev_nvme_attach_controller", 00:25:32.978 "req_id": 1 00:25:32.978 } 00:25:32.978 Got JSON-RPC error response 00:25:32.978 response: 00:25:32.978 { 00:25:32.978 "code": -5, 00:25:32.978 "message": "Input/output error" 00:25:32.978 } 00:25:32.978 17:41:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # es=1 00:25:32.978 17:41:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:25:32.978 17:41:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:25:32.978 17:41:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:25:32.978 17:41:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@229 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key key3 00:25:32.978 17:41:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key key3 00:25:32.978 17:41:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key key3 00:25:34.347 nvme0n1 00:25:34.347 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@230 -- # hostrpc bdev_nvme_get_controllers 00:25:34.347 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@230 -- # jq -r '.[].name' 00:25:34.347 17:41:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:25:34.347 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@230 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:34.347 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@231 -- # hostrpc bdev_nvme_detach_controller nvme0 00:25:34.347 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:25:34.910 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@233 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:25:34.910 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:34.910 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:34.910 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:34.910 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@234 -- # bdev_connect -b nvme0 00:25:34.910 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 00:25:34.910 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 00:25:35.167 nvme0n1 00:25:35.167 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@235 -- # hostrpc bdev_nvme_get_controllers 00:25:35.167 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@235 -- # jq -r '.[].name' 00:25:35.167 17:41:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:25:35.424 17:41:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@235 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:35.424 17:41:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@236 -- # hostrpc bdev_nvme_detach_controller nvme0 00:25:35.424 17:41:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:25:35.681 17:41:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@239 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key1 --dhchap-ctrlr-key key3 00:25:35.681 17:41:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:35.681 17:41:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:35.681 17:41:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:35.681 17:41:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@240 -- # nvme_set_keys nvme0 DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: '' 2s 00:25:35.682 17:41:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@49 -- # local ctl key ckey dev timeout 00:25:35.682 17:41:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # ctl=nvme0 00:25:35.682 17:41:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # key=DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: 00:25:35.682 17:41:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # ckey= 00:25:35.682 17:41:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # timeout=2s 00:25:35.682 17:41:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@52 -- # dev=/sys/devices/virtual/nvme-fabrics/ctl/nvme0 00:25:35.682 17:41:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@54 -- # [[ -z DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: ]] 00:25:35.682 17:41:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@54 -- # echo DHHC-1:01:NTJiZjEzZDZlMWJlODJmMDU3YWM1Y2QyMjVlZmMxMWHP+YDZ: 00:25:35.682 17:41:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@55 -- # [[ -z '' ]] 00:25:35.682 17:41:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@56 -- # [[ -z 2s ]] 00:25:35.682 17:41:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@56 -- # sleep 2s 00:25:37.577 17:41:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@241 -- # waitforblk nvme0n1 00:25:37.577 17:41:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1239 -- # local i=0 00:25:37.577 17:41:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1240 -- # lsblk -l -o NAME 00:25:37.577 17:41:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1240 -- # grep -q -w nvme0n1 00:25:37.577 17:41:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1246 -- # lsblk -l -o NAME 00:25:37.577 17:41:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1246 -- # grep -q -w nvme0n1 00:25:37.577 17:41:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1250 -- # return 0 00:25:37.578 17:41:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@243 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key1 --dhchap-ctrlr-key key2 00:25:37.578 17:41:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:37.578 17:41:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:37.578 17:41:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:37.578 17:41:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@244 -- # nvme_set_keys nvme0 '' DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: 2s 00:25:37.578 17:41:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@49 -- # local ctl key ckey dev timeout 00:25:37.578 17:41:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # ctl=nvme0 00:25:37.578 17:41:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # key= 00:25:37.578 17:41:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # ckey=DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: 00:25:37.578 17:41:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # timeout=2s 00:25:37.578 17:41:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@52 -- # dev=/sys/devices/virtual/nvme-fabrics/ctl/nvme0 00:25:37.578 17:41:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@54 -- # [[ -z '' ]] 00:25:37.578 17:41:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@55 -- # [[ -z DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: ]] 00:25:37.578 17:41:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@55 -- # echo DHHC-1:02:YTgwOWYyYzA3MjFlMDc2N2I5NWU3MTdhYmIzOTkyMDYwNzdiYzM5Yzg5NDg3Nzdj8Ee9Yw==: 00:25:37.578 17:41:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@56 -- # [[ -z 2s ]] 00:25:37.578 17:41:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@56 -- # sleep 2s 00:25:40.103 17:41:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@245 -- # waitforblk nvme0n1 00:25:40.103 17:41:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1239 -- # local i=0 00:25:40.103 17:41:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1240 -- # lsblk -l -o NAME 00:25:40.103 17:41:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1240 -- # grep -q -w nvme0n1 00:25:40.103 17:41:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1246 -- # lsblk -l -o NAME 00:25:40.103 17:41:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1246 -- # grep -q -w nvme0n1 00:25:40.103 17:41:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1250 -- # return 0 00:25:40.103 17:41:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@246 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:25:40.103 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:25:40.103 17:41:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@249 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key0 --dhchap-ctrlr-key key1 00:25:40.103 17:41:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:40.103 17:41:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:40.103 17:41:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:40.103 17:41:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@250 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:25:40.103 17:41:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:25:40.103 17:41:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:25:41.036 nvme0n1 00:25:41.036 17:41:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@252 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key2 --dhchap-ctrlr-key key3 00:25:41.036 17:41:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:41.036 17:41:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:41.036 17:41:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:41.036 17:41:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@253 -- # hostrpc bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key key3 00:25:41.036 17:41:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key key3 00:25:41.968 17:41:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@254 -- # hostrpc bdev_nvme_get_controllers 00:25:41.968 17:41:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@254 -- # jq -r '.[].name' 00:25:41.968 17:41:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:25:42.226 17:41:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@254 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:42.226 17:41:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@256 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:25:42.226 17:41:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:42.226 17:41:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:42.226 17:41:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:42.226 17:41:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@257 -- # hostrpc bdev_nvme_set_keys nvme0 00:25:42.226 17:41:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_keys nvme0 00:25:42.483 17:41:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@258 -- # hostrpc bdev_nvme_get_controllers 00:25:42.483 17:41:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@258 -- # jq -r '.[].name' 00:25:42.483 17:41:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:25:42.755 17:41:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@258 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:42.755 17:41:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@260 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key2 --dhchap-ctrlr-key key3 00:25:42.755 17:41:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:42.755 17:41:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:42.755 17:41:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:42.755 17:41:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@261 -- # NOT hostrpc bdev_nvme_set_keys nvme0 --dhchap-key key1 --dhchap-ctrlr-key key3 00:25:42.755 17:41:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # local es=0 00:25:42.755 17:41:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@654 -- # valid_exec_arg hostrpc bdev_nvme_set_keys nvme0 --dhchap-key key1 --dhchap-ctrlr-key key3 00:25:42.755 17:41:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@640 -- # local arg=hostrpc 00:25:42.755 17:41:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:25:42.755 17:41:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # type -t hostrpc 00:25:42.755 17:41:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:25:42.755 17:41:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # hostrpc bdev_nvme_set_keys nvme0 --dhchap-key key1 --dhchap-ctrlr-key key3 00:25:42.755 17:41:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_keys nvme0 --dhchap-key key1 --dhchap-ctrlr-key key3 00:25:43.687 request: 00:25:43.687 { 00:25:43.687 "name": "nvme0", 00:25:43.687 "dhchap_key": "key1", 00:25:43.687 "dhchap_ctrlr_key": "key3", 00:25:43.687 "method": "bdev_nvme_set_keys", 00:25:43.687 "req_id": 1 00:25:43.687 } 00:25:43.687 Got JSON-RPC error response 00:25:43.687 response: 00:25:43.687 { 00:25:43.687 "code": -13, 00:25:43.687 "message": "Permission denied" 00:25:43.687 } 00:25:43.687 17:41:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # es=1 00:25:43.687 17:41:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:25:43.687 17:41:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:25:43.687 17:41:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:25:43.687 17:41:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@262 -- # hostrpc bdev_nvme_get_controllers 00:25:43.687 17:41:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@262 -- # jq length 00:25:43.687 17:41:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:25:43.945 17:41:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@262 -- # (( 1 != 0 )) 00:25:43.945 17:41:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@263 -- # sleep 1s 00:25:44.874 17:41:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@262 -- # hostrpc bdev_nvme_get_controllers 00:25:44.874 17:41:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@262 -- # jq length 00:25:44.874 17:41:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:25:45.130 17:41:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@262 -- # (( 0 != 0 )) 00:25:45.131 17:41:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@267 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key0 --dhchap-ctrlr-key key1 00:25:45.131 17:41:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:45.131 17:41:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:45.131 17:41:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:45.131 17:41:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@268 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:25:45.131 17:41:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:25:45.131 17:41:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:25:46.499 nvme0n1 00:25:46.499 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@270 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --dhchap-key key2 --dhchap-ctrlr-key key3 00:25:46.499 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:46.499 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:46.499 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:46.499 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@271 -- # NOT hostrpc bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key key0 00:25:46.499 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # local es=0 00:25:46.499 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@654 -- # valid_exec_arg hostrpc bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key key0 00:25:46.500 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@640 -- # local arg=hostrpc 00:25:46.500 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:25:46.500 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # type -t hostrpc 00:25:46.500 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:25:46.500 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # hostrpc bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key key0 00:25:46.500 17:41:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key key0 00:25:47.432 request: 00:25:47.432 { 00:25:47.432 "name": "nvme0", 00:25:47.432 "dhchap_key": "key2", 00:25:47.432 "dhchap_ctrlr_key": "key0", 00:25:47.432 "method": "bdev_nvme_set_keys", 00:25:47.432 "req_id": 1 00:25:47.432 } 00:25:47.432 Got JSON-RPC error response 00:25:47.432 response: 00:25:47.432 { 00:25:47.432 "code": -13, 00:25:47.432 "message": "Permission denied" 00:25:47.432 } 00:25:47.432 17:41:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@655 -- # es=1 00:25:47.432 17:41:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:25:47.432 17:41:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:25:47.432 17:41:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:25:47.432 17:41:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@272 -- # hostrpc bdev_nvme_get_controllers 00:25:47.432 17:41:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@272 -- # jq length 00:25:47.432 17:41:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:25:47.689 17:41:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@272 -- # (( 1 != 0 )) 00:25:47.689 17:41:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@273 -- # sleep 1s 00:25:48.621 17:41:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@272 -- # hostrpc bdev_nvme_get_controllers 00:25:48.621 17:41:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@272 -- # jq length 00:25:48.621 17:41:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:25:48.879 17:41:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@272 -- # (( 0 != 0 )) 00:25:48.879 17:41:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@276 -- # trap - SIGINT SIGTERM EXIT 00:25:48.879 17:41:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@277 -- # cleanup 00:25:48.879 17:41:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@21 -- # killprocess 251771 00:25:48.879 17:41:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@954 -- # '[' -z 251771 ']' 00:25:48.879 17:41:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@958 -- # kill -0 251771 00:25:48.879 17:41:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@959 -- # uname 00:25:48.879 17:41:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:25:48.879 17:41:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 251771 00:25:48.879 17:41:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:25:48.879 17:41:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:25:48.879 17:41:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@972 -- # echo 'killing process with pid 251771' 00:25:48.879 killing process with pid 251771 00:25:48.879 17:41:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@973 -- # kill 251771 00:25:48.879 17:41:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@978 -- # wait 251771 00:25:49.444 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@22 -- # nvmftestfini 00:25:49.444 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@516 -- # nvmfcleanup 00:25:49.444 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@121 -- # sync 00:25:49.444 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:25:49.444 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@124 -- # set +e 00:25:49.444 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@125 -- # for i in {1..20} 00:25:49.444 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:25:49.444 rmmod nvme_tcp 00:25:49.444 rmmod nvme_fabrics 00:25:49.444 rmmod nvme_keyring 00:25:49.444 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:25:49.444 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@128 -- # set -e 00:25:49.444 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@129 -- # return 0 00:25:49.444 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@517 -- # '[' -n 274677 ']' 00:25:49.444 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@518 -- # killprocess 274677 00:25:49.444 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@954 -- # '[' -z 274677 ']' 00:25:49.444 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@958 -- # kill -0 274677 00:25:49.444 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@959 -- # uname 00:25:49.444 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:25:49.444 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 274677 00:25:49.444 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:25:49.444 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:25:49.444 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@972 -- # echo 'killing process with pid 274677' 00:25:49.444 killing process with pid 274677 00:25:49.444 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@973 -- # kill 274677 00:25:49.444 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@978 -- # wait 274677 00:25:49.703 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:25:49.703 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:25:49.703 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:25:49.703 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@297 -- # iptr 00:25:49.703 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@791 -- # iptables-save 00:25:49.703 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:25:49.703 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@791 -- # iptables-restore 00:25:49.704 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:25:49.704 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@302 -- # remove_spdk_ns 00:25:49.704 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:25:49.704 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:25:49.704 17:41:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:25:51.612 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:25:51.612 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@23 -- # rm -f /tmp/spdk.key-null.b1A /tmp/spdk.key-sha256.zDo /tmp/spdk.key-sha384.TnN /tmp/spdk.key-sha512.BLn /tmp/spdk.key-sha512.7q2 /tmp/spdk.key-sha384.OEt /tmp/spdk.key-sha256.Avu '' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvme-auth.log /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf-auth.log 00:25:51.612 00:25:51.612 real 3m32.492s 00:25:51.612 user 8m16.433s 00:25:51.612 sys 0m28.196s 00:25:51.612 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1130 -- # xtrace_disable 00:25:51.612 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:25:51.612 ************************************ 00:25:51.612 END TEST nvmf_auth_target 00:25:51.612 ************************************ 00:25:51.612 17:41:33 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@39 -- # '[' tcp = tcp ']' 00:25:51.612 17:41:33 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@40 -- # run_test nvmf_bdevio_no_huge /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp --no-hugepages 00:25:51.612 17:41:33 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:25:51.612 17:41:33 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:25:51.612 17:41:33 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:25:51.871 ************************************ 00:25:51.871 START TEST nvmf_bdevio_no_huge 00:25:51.871 ************************************ 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp --no-hugepages 00:25:51.871 * Looking for test storage... 00:25:51.871 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1711 -- # lcov --version 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@333 -- # local ver1 ver1_l 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@334 -- # local ver2 ver2_l 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@336 -- # IFS=.-: 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@336 -- # read -ra ver1 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@337 -- # IFS=.-: 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@337 -- # read -ra ver2 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@338 -- # local 'op=<' 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@340 -- # ver1_l=2 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@341 -- # ver2_l=1 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@344 -- # case "$op" in 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@345 -- # : 1 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@364 -- # (( v = 0 )) 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@365 -- # decimal 1 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@353 -- # local d=1 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@355 -- # echo 1 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@365 -- # ver1[v]=1 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@366 -- # decimal 2 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@353 -- # local d=2 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@355 -- # echo 2 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@366 -- # ver2[v]=2 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@368 -- # return 0 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:25:51.871 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:51.871 --rc genhtml_branch_coverage=1 00:25:51.871 --rc genhtml_function_coverage=1 00:25:51.871 --rc genhtml_legend=1 00:25:51.871 --rc geninfo_all_blocks=1 00:25:51.871 --rc geninfo_unexecuted_blocks=1 00:25:51.871 00:25:51.871 ' 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:25:51.871 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:51.871 --rc genhtml_branch_coverage=1 00:25:51.871 --rc genhtml_function_coverage=1 00:25:51.871 --rc genhtml_legend=1 00:25:51.871 --rc geninfo_all_blocks=1 00:25:51.871 --rc geninfo_unexecuted_blocks=1 00:25:51.871 00:25:51.871 ' 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:25:51.871 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:51.871 --rc genhtml_branch_coverage=1 00:25:51.871 --rc genhtml_function_coverage=1 00:25:51.871 --rc genhtml_legend=1 00:25:51.871 --rc geninfo_all_blocks=1 00:25:51.871 --rc geninfo_unexecuted_blocks=1 00:25:51.871 00:25:51.871 ' 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:25:51.871 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:51.871 --rc genhtml_branch_coverage=1 00:25:51.871 --rc genhtml_function_coverage=1 00:25:51.871 --rc genhtml_legend=1 00:25:51.871 --rc geninfo_all_blocks=1 00:25:51.871 --rc geninfo_unexecuted_blocks=1 00:25:51.871 00:25:51.871 ' 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@7 -- # uname -s 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:25:51.871 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@15 -- # shopt -s extglob 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- paths/export.sh@5 -- # export PATH 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@51 -- # : 0 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:25:51.872 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@55 -- # have_pci_nics=0 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@11 -- # MALLOC_BDEV_SIZE=64 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@14 -- # nvmftestinit 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@476 -- # prepare_net_devs 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@438 -- # local -g is_hw=no 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@440 -- # remove_spdk_ns 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@309 -- # xtrace_disable 00:25:51.872 17:41:33 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:25:54.403 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:25:54.403 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@315 -- # pci_devs=() 00:25:54.403 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@315 -- # local -a pci_devs 00:25:54.403 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@316 -- # pci_net_devs=() 00:25:54.403 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:25:54.403 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@317 -- # pci_drivers=() 00:25:54.403 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@317 -- # local -A pci_drivers 00:25:54.403 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@319 -- # net_devs=() 00:25:54.403 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@319 -- # local -ga net_devs 00:25:54.403 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@320 -- # e810=() 00:25:54.403 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@320 -- # local -ga e810 00:25:54.403 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@321 -- # x722=() 00:25:54.403 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@321 -- # local -ga x722 00:25:54.403 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@322 -- # mlx=() 00:25:54.403 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@322 -- # local -ga mlx 00:25:54.403 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:25:54.403 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:25:54.403 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:25:54.403 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:25:54.403 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:25:54.403 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:25:54.403 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:25:54.403 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:25:54.403 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:25:54.404 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:25:54.404 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@418 -- # [[ up == up ]] 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:25:54.404 Found net devices under 0000:0a:00.0: cvl_0_0 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@418 -- # [[ up == up ]] 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:25:54.404 Found net devices under 0000:0a:00.1: cvl_0_1 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@442 -- # is_hw=yes 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:25:54.404 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:25:54.404 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.240 ms 00:25:54.404 00:25:54.404 --- 10.0.0.2 ping statistics --- 00:25:54.404 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:25:54.404 rtt min/avg/max/mdev = 0.240/0.240/0.240/0.000 ms 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:25:54.404 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:25:54.404 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.241 ms 00:25:54.404 00:25:54.404 --- 10.0.0.1 ping statistics --- 00:25:54.404 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:25:54.404 rtt min/avg/max/mdev = 0.241/0.241/0.241/0.000 ms 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@450 -- # return 0 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@16 -- # nvmfappstart -m 0x78 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@726 -- # xtrace_disable 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@509 -- # nvmfpid=279933 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@510 -- # waitforlisten 279933 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --no-huge -s 1024 -m 0x78 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@835 -- # '[' -z 279933 ']' 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@840 -- # local max_retries=100 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:54.404 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@844 -- # xtrace_disable 00:25:54.404 17:41:35 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:25:54.404 [2024-12-06 17:41:35.961591] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:25:54.404 [2024-12-06 17:41:35.961716] [ DPDK EAL parameters: nvmf -c 0x78 -m 1024 --no-huge --iova-mode=va --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --file-prefix=spdk0 --proc-type=auto ] 00:25:54.404 [2024-12-06 17:41:36.037262] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:25:54.404 [2024-12-06 17:41:36.083680] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:25:54.404 [2024-12-06 17:41:36.083750] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:25:54.404 [2024-12-06 17:41:36.083763] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:25:54.405 [2024-12-06 17:41:36.083774] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:25:54.405 [2024-12-06 17:41:36.083784] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:25:54.405 [2024-12-06 17:41:36.084809] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:25:54.405 [2024-12-06 17:41:36.084872] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 5 00:25:54.405 [2024-12-06 17:41:36.084939] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 6 00:25:54.405 [2024-12-06 17:41:36.084942] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:25:54.405 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:25:54.405 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@868 -- # return 0 00:25:54.405 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:25:54.405 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@732 -- # xtrace_disable 00:25:54.405 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:25:54.405 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:25:54.405 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:25:54.405 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:54.405 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:25:54.663 [2024-12-06 17:41:36.241646] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:25:54.663 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:54.663 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:25:54.663 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:54.663 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:25:54.663 Malloc0 00:25:54.663 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:54.663 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@20 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:25:54.663 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:54.663 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:25:54.663 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:54.663 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@21 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:25:54.663 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:54.663 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:25:54.663 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:54.663 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@22 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:25:54.663 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:54.663 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:25:54.663 [2024-12-06 17:41:36.280372] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:25:54.663 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:54.663 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/bdev/bdevio/bdevio --json /dev/fd/62 --no-huge -s 1024 00:25:54.663 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@24 -- # gen_nvmf_target_json 00:25:54.663 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@560 -- # config=() 00:25:54.663 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@560 -- # local subsystem config 00:25:54.663 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:25:54.663 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:25:54.663 { 00:25:54.663 "params": { 00:25:54.663 "name": "Nvme$subsystem", 00:25:54.663 "trtype": "$TEST_TRANSPORT", 00:25:54.663 "traddr": "$NVMF_FIRST_TARGET_IP", 00:25:54.663 "adrfam": "ipv4", 00:25:54.663 "trsvcid": "$NVMF_PORT", 00:25:54.663 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:25:54.663 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:25:54.663 "hdgst": ${hdgst:-false}, 00:25:54.663 "ddgst": ${ddgst:-false} 00:25:54.663 }, 00:25:54.663 "method": "bdev_nvme_attach_controller" 00:25:54.663 } 00:25:54.663 EOF 00:25:54.663 )") 00:25:54.663 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@582 -- # cat 00:25:54.663 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@584 -- # jq . 00:25:54.663 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@585 -- # IFS=, 00:25:54.663 17:41:36 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:25:54.663 "params": { 00:25:54.663 "name": "Nvme1", 00:25:54.663 "trtype": "tcp", 00:25:54.663 "traddr": "10.0.0.2", 00:25:54.663 "adrfam": "ipv4", 00:25:54.663 "trsvcid": "4420", 00:25:54.663 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:25:54.663 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:25:54.663 "hdgst": false, 00:25:54.663 "ddgst": false 00:25:54.663 }, 00:25:54.663 "method": "bdev_nvme_attach_controller" 00:25:54.663 }' 00:25:54.663 [2024-12-06 17:41:36.332327] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:25:54.663 [2024-12-06 17:41:36.332397] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 -m 1024 --no-huge --iova-mode=va --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --file-prefix=spdk_pid279959 ] 00:25:54.663 [2024-12-06 17:41:36.402266] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:25:54.663 [2024-12-06 17:41:36.453473] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:25:54.663 [2024-12-06 17:41:36.453526] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:25:54.663 [2024-12-06 17:41:36.453529] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:25:55.229 I/O targets: 00:25:55.230 Nvme1n1: 131072 blocks of 512 bytes (64 MiB) 00:25:55.230 00:25:55.230 00:25:55.230 CUnit - A unit testing framework for C - Version 2.1-3 00:25:55.230 http://cunit.sourceforge.net/ 00:25:55.230 00:25:55.230 00:25:55.230 Suite: bdevio tests on: Nvme1n1 00:25:55.230 Test: blockdev write read block ...passed 00:25:55.230 Test: blockdev write zeroes read block ...passed 00:25:55.230 Test: blockdev write zeroes read no split ...passed 00:25:55.230 Test: blockdev write zeroes read split ...passed 00:25:55.230 Test: blockdev write zeroes read split partial ...passed 00:25:55.230 Test: blockdev reset ...[2024-12-06 17:41:37.005301] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 1] resetting controller 00:25:55.230 [2024-12-06 17:41:37.005414] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1c29ef0 (9): Bad file descriptor 00:25:55.487 [2024-12-06 17:41:37.155149] bdev_nvme.c:2286:bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller successful. 00:25:55.487 passed 00:25:55.487 Test: blockdev write read 8 blocks ...passed 00:25:55.487 Test: blockdev write read size > 128k ...passed 00:25:55.487 Test: blockdev write read invalid size ...passed 00:25:55.487 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:25:55.487 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:25:55.487 Test: blockdev write read max offset ...passed 00:25:55.487 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:25:55.745 Test: blockdev writev readv 8 blocks ...passed 00:25:55.745 Test: blockdev writev readv 30 x 1block ...passed 00:25:55.745 Test: blockdev writev readv block ...passed 00:25:55.745 Test: blockdev writev readv size > 128k ...passed 00:25:55.745 Test: blockdev writev readv size > 128k in two iovs ...passed 00:25:55.745 Test: blockdev comparev and writev ...[2024-12-06 17:41:37.406723] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:25:55.745 [2024-12-06 17:41:37.406759] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:25:55.745 [2024-12-06 17:41:37.406794] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:25:55.745 [2024-12-06 17:41:37.406813] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:25:55.745 [2024-12-06 17:41:37.407140] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:25:55.745 [2024-12-06 17:41:37.407164] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:25:55.745 [2024-12-06 17:41:37.407192] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:25:55.745 [2024-12-06 17:41:37.407210] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:25:55.745 [2024-12-06 17:41:37.407546] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:25:55.745 [2024-12-06 17:41:37.407573] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:25:55.745 [2024-12-06 17:41:37.407595] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:25:55.745 [2024-12-06 17:41:37.407611] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:25:55.745 [2024-12-06 17:41:37.407959] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:25:55.745 [2024-12-06 17:41:37.407984] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:25:55.745 [2024-12-06 17:41:37.408006] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:25:55.745 [2024-12-06 17:41:37.408023] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:25:55.745 passed 00:25:55.745 Test: blockdev nvme passthru rw ...passed 00:25:55.745 Test: blockdev nvme passthru vendor specific ...[2024-12-06 17:41:37.491072] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:25:55.745 [2024-12-06 17:41:37.491135] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:25:55.745 [2024-12-06 17:41:37.491296] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:25:55.745 [2024-12-06 17:41:37.491320] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:25:55.745 [2024-12-06 17:41:37.491463] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:25:55.745 [2024-12-06 17:41:37.491485] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:25:55.745 [2024-12-06 17:41:37.491630] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:25:55.745 [2024-12-06 17:41:37.491653] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:25:55.745 passed 00:25:55.745 Test: blockdev nvme admin passthru ...passed 00:25:55.745 Test: blockdev copy ...passed 00:25:55.745 00:25:55.746 Run Summary: Type Total Ran Passed Failed Inactive 00:25:55.746 suites 1 1 n/a 0 0 00:25:55.746 tests 23 23 23 0 0 00:25:55.746 asserts 152 152 152 0 n/a 00:25:55.746 00:25:55.746 Elapsed time = 1.398 seconds 00:25:56.311 17:41:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@26 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:25:56.311 17:41:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@563 -- # xtrace_disable 00:25:56.311 17:41:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:25:56.311 17:41:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:25:56.311 17:41:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@28 -- # trap - SIGINT SIGTERM EXIT 00:25:56.311 17:41:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@30 -- # nvmftestfini 00:25:56.311 17:41:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@516 -- # nvmfcleanup 00:25:56.311 17:41:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@121 -- # sync 00:25:56.311 17:41:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:25:56.311 17:41:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@124 -- # set +e 00:25:56.311 17:41:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@125 -- # for i in {1..20} 00:25:56.311 17:41:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:25:56.311 rmmod nvme_tcp 00:25:56.311 rmmod nvme_fabrics 00:25:56.311 rmmod nvme_keyring 00:25:56.311 17:41:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:25:56.311 17:41:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@128 -- # set -e 00:25:56.311 17:41:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@129 -- # return 0 00:25:56.311 17:41:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@517 -- # '[' -n 279933 ']' 00:25:56.311 17:41:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@518 -- # killprocess 279933 00:25:56.311 17:41:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@954 -- # '[' -z 279933 ']' 00:25:56.311 17:41:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@958 -- # kill -0 279933 00:25:56.311 17:41:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@959 -- # uname 00:25:56.311 17:41:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:25:56.311 17:41:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 279933 00:25:56.311 17:41:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@960 -- # process_name=reactor_3 00:25:56.311 17:41:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@964 -- # '[' reactor_3 = sudo ']' 00:25:56.311 17:41:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@972 -- # echo 'killing process with pid 279933' 00:25:56.311 killing process with pid 279933 00:25:56.311 17:41:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@973 -- # kill 279933 00:25:56.311 17:41:37 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@978 -- # wait 279933 00:25:56.572 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:25:56.572 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:25:56.572 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:25:56.572 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@297 -- # iptr 00:25:56.572 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@791 -- # iptables-save 00:25:56.572 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:25:56.572 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@791 -- # iptables-restore 00:25:56.572 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:25:56.572 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@302 -- # remove_spdk_ns 00:25:56.572 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:25:56.572 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:25:56.572 17:41:38 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:25:59.108 00:25:59.108 real 0m6.912s 00:25:59.108 user 0m12.458s 00:25:59.108 sys 0m2.664s 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1130 -- # xtrace_disable 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:25:59.108 ************************************ 00:25:59.108 END TEST nvmf_bdevio_no_huge 00:25:59.108 ************************************ 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@41 -- # run_test nvmf_tls /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/tls.sh --transport=tcp 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:25:59.108 ************************************ 00:25:59.108 START TEST nvmf_tls 00:25:59.108 ************************************ 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/tls.sh --transport=tcp 00:25:59.108 * Looking for test storage... 00:25:59.108 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1711 -- # lcov --version 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@333 -- # local ver1 ver1_l 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@334 -- # local ver2 ver2_l 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@336 -- # IFS=.-: 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@336 -- # read -ra ver1 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@337 -- # IFS=.-: 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@337 -- # read -ra ver2 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@338 -- # local 'op=<' 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@340 -- # ver1_l=2 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@341 -- # ver2_l=1 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@344 -- # case "$op" in 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@345 -- # : 1 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@364 -- # (( v = 0 )) 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@365 -- # decimal 1 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@353 -- # local d=1 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@355 -- # echo 1 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@365 -- # ver1[v]=1 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@366 -- # decimal 2 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@353 -- # local d=2 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@355 -- # echo 2 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@366 -- # ver2[v]=2 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@368 -- # return 0 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:25:59.108 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:59.108 --rc genhtml_branch_coverage=1 00:25:59.108 --rc genhtml_function_coverage=1 00:25:59.108 --rc genhtml_legend=1 00:25:59.108 --rc geninfo_all_blocks=1 00:25:59.108 --rc geninfo_unexecuted_blocks=1 00:25:59.108 00:25:59.108 ' 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:25:59.108 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:59.108 --rc genhtml_branch_coverage=1 00:25:59.108 --rc genhtml_function_coverage=1 00:25:59.108 --rc genhtml_legend=1 00:25:59.108 --rc geninfo_all_blocks=1 00:25:59.108 --rc geninfo_unexecuted_blocks=1 00:25:59.108 00:25:59.108 ' 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:25:59.108 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:59.108 --rc genhtml_branch_coverage=1 00:25:59.108 --rc genhtml_function_coverage=1 00:25:59.108 --rc genhtml_legend=1 00:25:59.108 --rc geninfo_all_blocks=1 00:25:59.108 --rc geninfo_unexecuted_blocks=1 00:25:59.108 00:25:59.108 ' 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:25:59.108 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:59.108 --rc genhtml_branch_coverage=1 00:25:59.108 --rc genhtml_function_coverage=1 00:25:59.108 --rc genhtml_legend=1 00:25:59.108 --rc geninfo_all_blocks=1 00:25:59.108 --rc geninfo_unexecuted_blocks=1 00:25:59.108 00:25:59.108 ' 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@7 -- # uname -s 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@15 -- # shopt -s extglob 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:59.108 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:59.109 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:59.109 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- paths/export.sh@5 -- # export PATH 00:25:59.109 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:59.109 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@51 -- # : 0 00:25:59.109 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:25:59.109 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:25:59.109 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:25:59.109 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:25:59.109 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:25:59.109 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:25:59.109 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:25:59.109 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:25:59.109 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:25:59.109 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@55 -- # have_pci_nics=0 00:25:59.109 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@12 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:25:59.109 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@63 -- # nvmftestinit 00:25:59.109 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:25:59.109 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:25:59.109 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@476 -- # prepare_net_devs 00:25:59.109 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@438 -- # local -g is_hw=no 00:25:59.109 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@440 -- # remove_spdk_ns 00:25:59.109 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:25:59.109 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:25:59.109 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:25:59.109 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:25:59.109 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:25:59.109 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@309 -- # xtrace_disable 00:25:59.109 17:41:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@315 -- # pci_devs=() 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@315 -- # local -a pci_devs 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@316 -- # pci_net_devs=() 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@317 -- # pci_drivers=() 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@317 -- # local -A pci_drivers 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@319 -- # net_devs=() 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@319 -- # local -ga net_devs 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@320 -- # e810=() 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@320 -- # local -ga e810 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@321 -- # x722=() 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@321 -- # local -ga x722 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@322 -- # mlx=() 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@322 -- # local -ga mlx 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:26:01.011 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:26:01.012 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:26:01.012 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@418 -- # [[ up == up ]] 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:26:01.012 Found net devices under 0000:0a:00.0: cvl_0_0 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@418 -- # [[ up == up ]] 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:26:01.012 Found net devices under 0000:0a:00.1: cvl_0_1 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@442 -- # is_hw=yes 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:26:01.012 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:26:01.012 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.245 ms 00:26:01.012 00:26:01.012 --- 10.0.0.2 ping statistics --- 00:26:01.012 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:26:01.012 rtt min/avg/max/mdev = 0.245/0.245/0.245/0.000 ms 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:26:01.012 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:26:01.012 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.173 ms 00:26:01.012 00:26:01.012 --- 10.0.0.1 ping statistics --- 00:26:01.012 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:26:01.012 rtt min/avg/max/mdev = 0.173/0.173/0.173/0.000 ms 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@450 -- # return 0 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@64 -- # nvmfappstart -m 0x2 --wait-for-rpc 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@726 -- # xtrace_disable 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # nvmfpid=282169 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 --wait-for-rpc 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # waitforlisten 282169 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 282169 ']' 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:01.012 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:01.012 17:41:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:26:01.269 [2024-12-06 17:41:42.875910] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:26:01.269 [2024-12-06 17:41:42.876008] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:01.269 [2024-12-06 17:41:42.950938] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:01.269 [2024-12-06 17:41:42.998467] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:26:01.269 [2024-12-06 17:41:42.998531] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:26:01.269 [2024-12-06 17:41:42.998545] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:26:01.269 [2024-12-06 17:41:42.998556] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:26:01.269 [2024-12-06 17:41:42.998566] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:26:01.269 [2024-12-06 17:41:42.999225] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:26:01.269 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:01.269 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:26:01.269 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:26:01.269 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@732 -- # xtrace_disable 00:26:01.269 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:26:01.526 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:26:01.526 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@66 -- # '[' tcp '!=' tcp ']' 00:26:01.526 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_set_default_impl -i ssl 00:26:01.813 true 00:26:01.813 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:26:01.813 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@74 -- # jq -r .tls_version 00:26:02.069 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@74 -- # version=0 00:26:02.069 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@75 -- # [[ 0 != \0 ]] 00:26:02.069 17:41:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@81 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --tls-version 13 00:26:02.326 17:41:44 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:26:02.326 17:41:44 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@82 -- # jq -r .tls_version 00:26:02.583 17:41:44 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@82 -- # version=13 00:26:02.583 17:41:44 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@83 -- # [[ 13 != \1\3 ]] 00:26:02.583 17:41:44 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --tls-version 7 00:26:02.840 17:41:44 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@90 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:26:02.840 17:41:44 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@90 -- # jq -r .tls_version 00:26:03.097 17:41:44 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@90 -- # version=7 00:26:03.097 17:41:44 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@91 -- # [[ 7 != \7 ]] 00:26:03.097 17:41:44 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@97 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:26:03.097 17:41:44 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@97 -- # jq -r .enable_ktls 00:26:03.354 17:41:45 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@97 -- # ktls=false 00:26:03.354 17:41:45 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@98 -- # [[ false != \f\a\l\s\e ]] 00:26:03.354 17:41:45 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@104 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --enable-ktls 00:26:03.610 17:41:45 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@105 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:26:03.610 17:41:45 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@105 -- # jq -r .enable_ktls 00:26:03.867 17:41:45 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@105 -- # ktls=true 00:26:03.867 17:41:45 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@106 -- # [[ true != \t\r\u\e ]] 00:26:03.867 17:41:45 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@112 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --disable-ktls 00:26:04.430 17:41:45 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@113 -- # jq -r .enable_ktls 00:26:04.430 17:41:45 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@113 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:26:04.430 17:41:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@113 -- # ktls=false 00:26:04.430 17:41:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@114 -- # [[ false != \f\a\l\s\e ]] 00:26:04.430 17:41:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@119 -- # format_interchange_psk 00112233445566778899aabbccddeeff 1 00:26:04.430 17:41:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@743 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff 1 00:26:04.430 17:41:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@730 -- # local prefix key digest 00:26:04.430 17:41:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@732 -- # prefix=NVMeTLSkey-1 00:26:04.430 17:41:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@732 -- # key=00112233445566778899aabbccddeeff 00:26:04.430 17:41:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@732 -- # digest=1 00:26:04.430 17:41:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@733 -- # python - 00:26:04.686 17:41:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@119 -- # key=NVMeTLSkey-1:01:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmZwJEiQ: 00:26:04.686 17:41:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@120 -- # format_interchange_psk ffeeddccbbaa99887766554433221100 1 00:26:04.686 17:41:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@743 -- # format_key NVMeTLSkey-1 ffeeddccbbaa99887766554433221100 1 00:26:04.686 17:41:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@730 -- # local prefix key digest 00:26:04.686 17:41:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@732 -- # prefix=NVMeTLSkey-1 00:26:04.686 17:41:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@732 -- # key=ffeeddccbbaa99887766554433221100 00:26:04.686 17:41:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@732 -- # digest=1 00:26:04.686 17:41:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@733 -- # python - 00:26:04.686 17:41:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@120 -- # key_2=NVMeTLSkey-1:01:ZmZlZWRkY2NiYmFhOTk4ODc3NjY1NTQ0MzMyMjExMDBfBm/Y: 00:26:04.686 17:41:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@122 -- # mktemp 00:26:04.686 17:41:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@122 -- # key_path=/tmp/tmp.7KTTXF2gPL 00:26:04.686 17:41:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@123 -- # mktemp 00:26:04.686 17:41:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@123 -- # key_2_path=/tmp/tmp.WvD1hwJPAj 00:26:04.686 17:41:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@125 -- # echo -n NVMeTLSkey-1:01:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmZwJEiQ: 00:26:04.686 17:41:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@126 -- # echo -n NVMeTLSkey-1:01:ZmZlZWRkY2NiYmFhOTk4ODc3NjY1NTQ0MzMyMjExMDBfBm/Y: 00:26:04.686 17:41:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@128 -- # chmod 0600 /tmp/tmp.7KTTXF2gPL 00:26:04.686 17:41:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@129 -- # chmod 0600 /tmp/tmp.WvD1hwJPAj 00:26:04.686 17:41:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@131 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --tls-version 13 00:26:04.943 17:41:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@132 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py framework_start_init 00:26:05.201 17:41:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@134 -- # setup_nvmf_tgt /tmp/tmp.7KTTXF2gPL 00:26:05.201 17:41:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@50 -- # local key=/tmp/tmp.7KTTXF2gPL 00:26:05.201 17:41:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:26:05.766 [2024-12-06 17:41:47.306793] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:26:05.766 17:41:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:26:05.766 17:41:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:26:06.025 [2024-12-06 17:41:47.840127] tcp.c:1049:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:26:06.025 [2024-12-06 17:41:47.840353] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:26:06.025 17:41:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:26:06.283 malloc0 00:26:06.540 17:41:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:26:06.798 17:41:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py keyring_file_add_key key0 /tmp/tmp.7KTTXF2gPL 00:26:07.056 17:41:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk key0 00:26:07.314 17:41:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@138 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -S ssl -q 64 -o 4096 -w randrw -M 30 -t 10 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 hostnqn:nqn.2016-06.io.spdk:host1' --psk-path /tmp/tmp.7KTTXF2gPL 00:26:17.280 Initializing NVMe Controllers 00:26:17.280 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:26:17.280 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:26:17.280 Initialization complete. Launching workers. 00:26:17.280 ======================================================== 00:26:17.280 Latency(us) 00:26:17.280 Device Information : IOPS MiB/s Average min max 00:26:17.280 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 8732.79 34.11 7330.63 1210.89 8535.68 00:26:17.280 ======================================================== 00:26:17.280 Total : 8732.79 34.11 7330.63 1210.89 8535.68 00:26:17.280 00:26:17.280 17:41:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@144 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.7KTTXF2gPL 00:26:17.280 17:41:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:26:17.280 17:41:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:26:17.280 17:41:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:26:17.280 17:41:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # psk=/tmp/tmp.7KTTXF2gPL 00:26:17.280 17:41:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:26:17.280 17:41:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=284065 00:26:17.280 17:41:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:26:17.280 17:41:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:26:17.280 17:41:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 284065 /var/tmp/bdevperf.sock 00:26:17.280 17:41:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 284065 ']' 00:26:17.280 17:41:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:26:17.280 17:41:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:17.280 17:41:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:26:17.280 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:26:17.280 17:41:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:17.280 17:41:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:26:17.280 [2024-12-06 17:41:59.090262] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:26:17.280 [2024-12-06 17:41:59.090351] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid284065 ] 00:26:17.538 [2024-12-06 17:41:59.156540] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:17.538 [2024-12-06 17:41:59.202934] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:26:17.538 17:41:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:17.538 17:41:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:26:17.538 17:41:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.7KTTXF2gPL 00:26:17.796 17:41:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk key0 00:26:18.054 [2024-12-06 17:41:59.846480] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:26:18.311 TLSTESTn1 00:26:18.311 17:41:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -t 20 -s /var/tmp/bdevperf.sock perform_tests 00:26:18.311 Running I/O for 10 seconds... 00:26:20.618 3179.00 IOPS, 12.42 MiB/s [2024-12-06T16:42:03.403Z] 3304.50 IOPS, 12.91 MiB/s [2024-12-06T16:42:04.336Z] 3317.67 IOPS, 12.96 MiB/s [2024-12-06T16:42:05.277Z] 3302.75 IOPS, 12.90 MiB/s [2024-12-06T16:42:06.212Z] 3323.60 IOPS, 12.98 MiB/s [2024-12-06T16:42:07.146Z] 3349.83 IOPS, 13.09 MiB/s [2024-12-06T16:42:08.520Z] 3358.86 IOPS, 13.12 MiB/s [2024-12-06T16:42:09.085Z] 3365.00 IOPS, 13.14 MiB/s [2024-12-06T16:42:10.459Z] 3366.22 IOPS, 13.15 MiB/s [2024-12-06T16:42:10.459Z] 3359.00 IOPS, 13.12 MiB/s 00:26:28.620 Latency(us) 00:26:28.620 [2024-12-06T16:42:10.459Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:28.620 Job: TLSTESTn1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:26:28.620 Verification LBA range: start 0x0 length 0x2000 00:26:28.620 TLSTESTn1 : 10.02 3364.36 13.14 0.00 0.00 37982.70 7621.59 33981.63 00:26:28.620 [2024-12-06T16:42:10.459Z] =================================================================================================================== 00:26:28.620 [2024-12-06T16:42:10.459Z] Total : 3364.36 13.14 0.00 0.00 37982.70 7621.59 33981.63 00:26:28.620 { 00:26:28.620 "results": [ 00:26:28.620 { 00:26:28.620 "job": "TLSTESTn1", 00:26:28.620 "core_mask": "0x4", 00:26:28.620 "workload": "verify", 00:26:28.620 "status": "finished", 00:26:28.620 "verify_range": { 00:26:28.620 "start": 0, 00:26:28.620 "length": 8192 00:26:28.620 }, 00:26:28.620 "queue_depth": 128, 00:26:28.620 "io_size": 4096, 00:26:28.620 "runtime": 10.021813, 00:26:28.620 "iops": 3364.3613186556163, 00:26:28.620 "mibps": 13.142036400998501, 00:26:28.620 "io_failed": 0, 00:26:28.620 "io_timeout": 0, 00:26:28.620 "avg_latency_us": 37982.70415394366, 00:26:28.620 "min_latency_us": 7621.594074074074, 00:26:28.620 "max_latency_us": 33981.62962962963 00:26:28.620 } 00:26:28.620 ], 00:26:28.620 "core_count": 1 00:26:28.620 } 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@45 -- # trap 'nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@46 -- # killprocess 284065 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 284065 ']' 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 284065 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 284065 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 284065' 00:26:28.620 killing process with pid 284065 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 284065 00:26:28.620 Received shutdown signal, test time was about 10.000000 seconds 00:26:28.620 00:26:28.620 Latency(us) 00:26:28.620 [2024-12-06T16:42:10.459Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:28.620 [2024-12-06T16:42:10.459Z] =================================================================================================================== 00:26:28.620 [2024-12-06T16:42:10.459Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 284065 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@147 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.WvD1hwJPAj 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@652 -- # local es=0 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@654 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.WvD1hwJPAj 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@640 -- # local arg=run_bdevperf 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # type -t run_bdevperf 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@655 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.WvD1hwJPAj 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # psk=/tmp/tmp.WvD1hwJPAj 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=285999 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:26:28.620 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 285999 /var/tmp/bdevperf.sock 00:26:28.621 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 285999 ']' 00:26:28.621 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:26:28.621 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:28.621 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:26:28.621 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:26:28.621 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:28.621 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:26:28.621 [2024-12-06 17:42:10.381500] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:26:28.621 [2024-12-06 17:42:10.381593] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid285999 ] 00:26:28.621 [2024-12-06 17:42:10.450567] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:28.878 [2024-12-06 17:42:10.500296] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:26:28.878 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:28.878 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:26:28.878 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.WvD1hwJPAj 00:26:29.135 17:42:10 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk key0 00:26:29.392 [2024-12-06 17:42:11.155706] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:26:29.392 [2024-12-06 17:42:11.161545] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 421:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:26:29.393 [2024-12-06 17:42:11.162047] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x12e9610 (107): Transport endpoint is not connected 00:26:29.393 [2024-12-06 17:42:11.163036] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x12e9610 (9): Bad file descriptor 00:26:29.393 [2024-12-06 17:42:11.164035] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 0] Ctrlr is in error state 00:26:29.393 [2024-12-06 17:42:11.164062] nvme.c: 709:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 10.0.0.2 00:26:29.393 [2024-12-06 17:42:11.164076] nvme.c: 895:nvme_dummy_attach_fail_cb: *ERROR*: Failed to attach nvme ctrlr: trtype=TCP adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 subnqn=nqn.2016-06.io.spdk:cnode1, Operation not permitted 00:26:29.393 [2024-12-06 17:42:11.164093] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 0] in failed state. 00:26:29.393 request: 00:26:29.393 { 00:26:29.393 "name": "TLSTEST", 00:26:29.393 "trtype": "tcp", 00:26:29.393 "traddr": "10.0.0.2", 00:26:29.393 "adrfam": "ipv4", 00:26:29.393 "trsvcid": "4420", 00:26:29.393 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:26:29.393 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:26:29.393 "prchk_reftag": false, 00:26:29.393 "prchk_guard": false, 00:26:29.393 "hdgst": false, 00:26:29.393 "ddgst": false, 00:26:29.393 "psk": "key0", 00:26:29.393 "allow_unrecognized_csi": false, 00:26:29.393 "method": "bdev_nvme_attach_controller", 00:26:29.393 "req_id": 1 00:26:29.393 } 00:26:29.393 Got JSON-RPC error response 00:26:29.393 response: 00:26:29.393 { 00:26:29.393 "code": -5, 00:26:29.393 "message": "Input/output error" 00:26:29.393 } 00:26:29.393 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@37 -- # killprocess 285999 00:26:29.393 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 285999 ']' 00:26:29.393 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 285999 00:26:29.393 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:26:29.393 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:29.393 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 285999 00:26:29.393 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:26:29.393 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:26:29.393 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 285999' 00:26:29.393 killing process with pid 285999 00:26:29.393 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 285999 00:26:29.393 Received shutdown signal, test time was about 10.000000 seconds 00:26:29.393 00:26:29.393 Latency(us) 00:26:29.393 [2024-12-06T16:42:11.232Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:29.393 [2024-12-06T16:42:11.232Z] =================================================================================================================== 00:26:29.393 [2024-12-06T16:42:11.232Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:26:29.393 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 285999 00:26:29.650 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@38 -- # return 1 00:26:29.650 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@655 -- # es=1 00:26:29.650 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:26:29.650 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:26:29.650 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:26:29.650 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@150 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host2 /tmp/tmp.7KTTXF2gPL 00:26:29.650 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@652 -- # local es=0 00:26:29.650 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@654 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host2 /tmp/tmp.7KTTXF2gPL 00:26:29.650 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@640 -- # local arg=run_bdevperf 00:26:29.650 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:26:29.650 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # type -t run_bdevperf 00:26:29.650 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:26:29.650 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@655 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host2 /tmp/tmp.7KTTXF2gPL 00:26:29.650 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:26:29.650 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:26:29.650 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host2 00:26:29.650 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # psk=/tmp/tmp.7KTTXF2gPL 00:26:29.650 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:26:29.650 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=286137 00:26:29.650 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:26:29.650 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:26:29.650 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 286137 /var/tmp/bdevperf.sock 00:26:29.650 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 286137 ']' 00:26:29.650 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:26:29.650 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:29.650 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:26:29.650 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:26:29.651 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:29.651 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:26:29.651 [2024-12-06 17:42:11.471764] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:26:29.651 [2024-12-06 17:42:11.471842] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid286137 ] 00:26:29.908 [2024-12-06 17:42:11.538232] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:29.908 [2024-12-06 17:42:11.580773] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:26:29.908 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:29.908 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:26:29.908 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.7KTTXF2gPL 00:26:30.165 17:42:11 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host2 --psk key0 00:26:30.423 [2024-12-06 17:42:12.235154] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:26:30.423 [2024-12-06 17:42:12.241981] tcp.c: 987:tcp_sock_get_key: *ERROR*: Could not find PSK for identity: NVMe0R01 nqn.2016-06.io.spdk:host2 nqn.2016-06.io.spdk:cnode1 00:26:30.423 [2024-12-06 17:42:12.242010] posix.c: 573:posix_sock_psk_find_session_server_cb: *ERROR*: Unable to find PSK for identity: NVMe0R01 nqn.2016-06.io.spdk:host2 nqn.2016-06.io.spdk:cnode1 00:26:30.423 [2024-12-06 17:42:12.242067] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 421:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:26:30.423 [2024-12-06 17:42:12.242219] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xfa6610 (107): Transport endpoint is not connected 00:26:30.423 [2024-12-06 17:42:12.243209] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xfa6610 (9): Bad file descriptor 00:26:30.423 [2024-12-06 17:42:12.244209] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 0] Ctrlr is in error state 00:26:30.423 [2024-12-06 17:42:12.244233] nvme.c: 709:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 10.0.0.2 00:26:30.423 [2024-12-06 17:42:12.244255] nvme.c: 895:nvme_dummy_attach_fail_cb: *ERROR*: Failed to attach nvme ctrlr: trtype=TCP adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 subnqn=nqn.2016-06.io.spdk:cnode1, Operation not permitted 00:26:30.423 [2024-12-06 17:42:12.244274] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 0] in failed state. 00:26:30.423 request: 00:26:30.423 { 00:26:30.423 "name": "TLSTEST", 00:26:30.423 "trtype": "tcp", 00:26:30.423 "traddr": "10.0.0.2", 00:26:30.423 "adrfam": "ipv4", 00:26:30.423 "trsvcid": "4420", 00:26:30.423 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:26:30.423 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:26:30.423 "prchk_reftag": false, 00:26:30.423 "prchk_guard": false, 00:26:30.423 "hdgst": false, 00:26:30.423 "ddgst": false, 00:26:30.423 "psk": "key0", 00:26:30.423 "allow_unrecognized_csi": false, 00:26:30.423 "method": "bdev_nvme_attach_controller", 00:26:30.423 "req_id": 1 00:26:30.423 } 00:26:30.423 Got JSON-RPC error response 00:26:30.423 response: 00:26:30.423 { 00:26:30.423 "code": -5, 00:26:30.423 "message": "Input/output error" 00:26:30.423 } 00:26:30.681 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@37 -- # killprocess 286137 00:26:30.681 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 286137 ']' 00:26:30.681 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 286137 00:26:30.681 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 286137 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 286137' 00:26:30.682 killing process with pid 286137 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 286137 00:26:30.682 Received shutdown signal, test time was about 10.000000 seconds 00:26:30.682 00:26:30.682 Latency(us) 00:26:30.682 [2024-12-06T16:42:12.521Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:30.682 [2024-12-06T16:42:12.521Z] =================================================================================================================== 00:26:30.682 [2024-12-06T16:42:12.521Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 286137 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@38 -- # return 1 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@655 -- # es=1 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@153 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode2 nqn.2016-06.io.spdk:host1 /tmp/tmp.7KTTXF2gPL 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@652 -- # local es=0 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@654 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode2 nqn.2016-06.io.spdk:host1 /tmp/tmp.7KTTXF2gPL 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@640 -- # local arg=run_bdevperf 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # type -t run_bdevperf 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@655 -- # run_bdevperf nqn.2016-06.io.spdk:cnode2 nqn.2016-06.io.spdk:host1 /tmp/tmp.7KTTXF2gPL 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode2 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # psk=/tmp/tmp.7KTTXF2gPL 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=286280 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 286280 /var/tmp/bdevperf.sock 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 286280 ']' 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:26:30.682 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:30.682 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:26:30.939 [2024-12-06 17:42:12.525175] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:26:30.939 [2024-12-06 17:42:12.525268] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid286280 ] 00:26:30.939 [2024-12-06 17:42:12.592020] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:30.939 [2024-12-06 17:42:12.634201] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:26:30.939 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:30.939 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:26:30.939 17:42:12 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.7KTTXF2gPL 00:26:31.196 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode2 -q nqn.2016-06.io.spdk:host1 --psk key0 00:26:31.453 [2024-12-06 17:42:13.278857] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:26:31.453 [2024-12-06 17:42:13.284727] tcp.c: 987:tcp_sock_get_key: *ERROR*: Could not find PSK for identity: NVMe0R01 nqn.2016-06.io.spdk:host1 nqn.2016-06.io.spdk:cnode2 00:26:31.453 [2024-12-06 17:42:13.284771] posix.c: 573:posix_sock_psk_find_session_server_cb: *ERROR*: Unable to find PSK for identity: NVMe0R01 nqn.2016-06.io.spdk:host1 nqn.2016-06.io.spdk:cnode2 00:26:31.453 [2024-12-06 17:42:13.284817] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 421:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:26:31.453 [2024-12-06 17:42:13.285129] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x125b610 (107): Transport endpoint is not connected 00:26:31.453 [2024-12-06 17:42:13.286118] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x125b610 (9): Bad file descriptor 00:26:31.453 [2024-12-06 17:42:13.287116] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 0] Ctrlr is in error state 00:26:31.453 [2024-12-06 17:42:13.287137] nvme.c: 709:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 10.0.0.2 00:26:31.453 [2024-12-06 17:42:13.287151] nvme.c: 895:nvme_dummy_attach_fail_cb: *ERROR*: Failed to attach nvme ctrlr: trtype=TCP adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 subnqn=nqn.2016-06.io.spdk:cnode2, Operation not permitted 00:26:31.453 [2024-12-06 17:42:13.287170] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 0] in failed state. 00:26:31.711 request: 00:26:31.711 { 00:26:31.711 "name": "TLSTEST", 00:26:31.711 "trtype": "tcp", 00:26:31.711 "traddr": "10.0.0.2", 00:26:31.711 "adrfam": "ipv4", 00:26:31.711 "trsvcid": "4420", 00:26:31.711 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:26:31.711 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:26:31.711 "prchk_reftag": false, 00:26:31.711 "prchk_guard": false, 00:26:31.711 "hdgst": false, 00:26:31.711 "ddgst": false, 00:26:31.711 "psk": "key0", 00:26:31.711 "allow_unrecognized_csi": false, 00:26:31.711 "method": "bdev_nvme_attach_controller", 00:26:31.711 "req_id": 1 00:26:31.711 } 00:26:31.711 Got JSON-RPC error response 00:26:31.711 response: 00:26:31.711 { 00:26:31.711 "code": -5, 00:26:31.711 "message": "Input/output error" 00:26:31.711 } 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@37 -- # killprocess 286280 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 286280 ']' 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 286280 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 286280 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 286280' 00:26:31.711 killing process with pid 286280 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 286280 00:26:31.711 Received shutdown signal, test time was about 10.000000 seconds 00:26:31.711 00:26:31.711 Latency(us) 00:26:31.711 [2024-12-06T16:42:13.550Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:31.711 [2024-12-06T16:42:13.550Z] =================================================================================================================== 00:26:31.711 [2024-12-06T16:42:13.550Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 286280 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@38 -- # return 1 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@655 -- # es=1 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@156 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 '' 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@652 -- # local es=0 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@654 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 '' 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@640 -- # local arg=run_bdevperf 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # type -t run_bdevperf 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@655 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 '' 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # psk= 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=286420 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 286420 /var/tmp/bdevperf.sock 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 286420 ']' 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:26:31.711 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:31.711 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:26:31.969 [2024-12-06 17:42:13.591120] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:26:31.969 [2024-12-06 17:42:13.591212] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid286420 ] 00:26:31.969 [2024-12-06 17:42:13.657797] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:31.969 [2024-12-06 17:42:13.703264] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:26:32.226 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:32.226 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:26:32.226 17:42:13 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 '' 00:26:32.485 [2024-12-06 17:42:14.076964] keyring.c: 24:keyring_file_check_path: *ERROR*: Non-absolute paths are not allowed: 00:26:32.485 [2024-12-06 17:42:14.077026] keyring.c: 126:spdk_keyring_add_key: *ERROR*: Failed to add key 'key0' to the keyring 00:26:32.485 request: 00:26:32.485 { 00:26:32.485 "name": "key0", 00:26:32.485 "path": "", 00:26:32.485 "method": "keyring_file_add_key", 00:26:32.485 "req_id": 1 00:26:32.485 } 00:26:32.485 Got JSON-RPC error response 00:26:32.485 response: 00:26:32.485 { 00:26:32.485 "code": -1, 00:26:32.485 "message": "Operation not permitted" 00:26:32.485 } 00:26:32.485 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk key0 00:26:32.743 [2024-12-06 17:42:14.361842] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:26:32.743 [2024-12-06 17:42:14.361904] bdev_nvme.c:6749:spdk_bdev_nvme_create: *ERROR*: Could not load PSK: key0 00:26:32.743 request: 00:26:32.743 { 00:26:32.743 "name": "TLSTEST", 00:26:32.743 "trtype": "tcp", 00:26:32.743 "traddr": "10.0.0.2", 00:26:32.743 "adrfam": "ipv4", 00:26:32.743 "trsvcid": "4420", 00:26:32.743 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:26:32.743 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:26:32.743 "prchk_reftag": false, 00:26:32.743 "prchk_guard": false, 00:26:32.743 "hdgst": false, 00:26:32.743 "ddgst": false, 00:26:32.743 "psk": "key0", 00:26:32.743 "allow_unrecognized_csi": false, 00:26:32.743 "method": "bdev_nvme_attach_controller", 00:26:32.743 "req_id": 1 00:26:32.743 } 00:26:32.743 Got JSON-RPC error response 00:26:32.743 response: 00:26:32.743 { 00:26:32.743 "code": -126, 00:26:32.743 "message": "Required key not available" 00:26:32.743 } 00:26:32.743 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@37 -- # killprocess 286420 00:26:32.743 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 286420 ']' 00:26:32.743 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 286420 00:26:32.743 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:26:32.743 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:32.743 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 286420 00:26:32.743 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:26:32.743 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:26:32.743 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 286420' 00:26:32.743 killing process with pid 286420 00:26:32.743 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 286420 00:26:32.743 Received shutdown signal, test time was about 10.000000 seconds 00:26:32.743 00:26:32.743 Latency(us) 00:26:32.743 [2024-12-06T16:42:14.582Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:32.743 [2024-12-06T16:42:14.582Z] =================================================================================================================== 00:26:32.743 [2024-12-06T16:42:14.582Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:26:32.743 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 286420 00:26:33.001 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@38 -- # return 1 00:26:33.001 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@655 -- # es=1 00:26:33.001 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:26:33.001 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:26:33.001 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:26:33.001 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@159 -- # killprocess 282169 00:26:33.001 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 282169 ']' 00:26:33.001 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 282169 00:26:33.001 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:26:33.001 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:33.001 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 282169 00:26:33.001 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:26:33.001 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:26:33.001 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 282169' 00:26:33.001 killing process with pid 282169 00:26:33.001 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 282169 00:26:33.001 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 282169 00:26:33.259 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@160 -- # format_interchange_psk 00112233445566778899aabbccddeeff0011223344556677 2 00:26:33.259 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@743 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff0011223344556677 2 00:26:33.259 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@730 -- # local prefix key digest 00:26:33.259 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@732 -- # prefix=NVMeTLSkey-1 00:26:33.259 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@732 -- # key=00112233445566778899aabbccddeeff0011223344556677 00:26:33.259 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@732 -- # digest=2 00:26:33.259 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@733 -- # python - 00:26:33.259 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@160 -- # key_long=NVMeTLSkey-1:02:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmYwMDExMjIzMzQ0NTU2Njc3wWXNJw==: 00:26:33.259 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@161 -- # mktemp 00:26:33.259 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@161 -- # key_long_path=/tmp/tmp.OtECQfGJ9H 00:26:33.259 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@162 -- # echo -n NVMeTLSkey-1:02:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmYwMDExMjIzMzQ0NTU2Njc3wWXNJw==: 00:26:33.259 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@163 -- # chmod 0600 /tmp/tmp.OtECQfGJ9H 00:26:33.259 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@164 -- # nvmfappstart -m 0x2 00:26:33.259 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:26:33.259 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@726 -- # xtrace_disable 00:26:33.259 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:26:33.259 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # nvmfpid=286573 00:26:33.259 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:26:33.259 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # waitforlisten 286573 00:26:33.259 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 286573 ']' 00:26:33.259 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:33.259 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:33.259 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:33.259 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:33.259 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:33.259 17:42:14 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:26:33.259 [2024-12-06 17:42:14.967077] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:26:33.259 [2024-12-06 17:42:14.967184] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:33.259 [2024-12-06 17:42:15.039968] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:33.259 [2024-12-06 17:42:15.084406] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:26:33.259 [2024-12-06 17:42:15.084467] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:26:33.259 [2024-12-06 17:42:15.084489] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:26:33.259 [2024-12-06 17:42:15.084500] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:26:33.259 [2024-12-06 17:42:15.084510] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:26:33.259 [2024-12-06 17:42:15.085104] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:26:33.517 17:42:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:33.517 17:42:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:26:33.517 17:42:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:26:33.517 17:42:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@732 -- # xtrace_disable 00:26:33.517 17:42:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:26:33.517 17:42:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:26:33.517 17:42:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@166 -- # setup_nvmf_tgt /tmp/tmp.OtECQfGJ9H 00:26:33.517 17:42:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@50 -- # local key=/tmp/tmp.OtECQfGJ9H 00:26:33.517 17:42:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:26:33.774 [2024-12-06 17:42:15.484612] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:26:33.774 17:42:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:26:34.031 17:42:15 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:26:34.595 [2024-12-06 17:42:16.126357] tcp.c:1049:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:26:34.595 [2024-12-06 17:42:16.126641] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:26:34.595 17:42:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:26:34.853 malloc0 00:26:34.853 17:42:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:26:35.110 17:42:16 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py keyring_file_add_key key0 /tmp/tmp.OtECQfGJ9H 00:26:35.368 17:42:17 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk key0 00:26:35.626 17:42:17 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@168 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.OtECQfGJ9H 00:26:35.626 17:42:17 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:26:35.626 17:42:17 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:26:35.626 17:42:17 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:26:35.626 17:42:17 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # psk=/tmp/tmp.OtECQfGJ9H 00:26:35.626 17:42:17 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:26:35.626 17:42:17 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=286866 00:26:35.626 17:42:17 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:26:35.626 17:42:17 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:26:35.626 17:42:17 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 286866 /var/tmp/bdevperf.sock 00:26:35.626 17:42:17 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 286866 ']' 00:26:35.626 17:42:17 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:26:35.626 17:42:17 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:35.626 17:42:17 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:26:35.626 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:26:35.626 17:42:17 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:35.626 17:42:17 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:26:35.626 [2024-12-06 17:42:17.388404] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:26:35.626 [2024-12-06 17:42:17.388502] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid286866 ] 00:26:35.626 [2024-12-06 17:42:17.456392] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:35.883 [2024-12-06 17:42:17.503206] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:26:35.883 17:42:17 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:35.884 17:42:17 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:26:35.884 17:42:17 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.OtECQfGJ9H 00:26:36.141 17:42:17 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk key0 00:26:36.399 [2024-12-06 17:42:18.142047] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:26:36.399 TLSTESTn1 00:26:36.656 17:42:18 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -t 20 -s /var/tmp/bdevperf.sock perform_tests 00:26:36.656 Running I/O for 10 seconds... 00:26:38.962 3240.00 IOPS, 12.66 MiB/s [2024-12-06T16:42:21.732Z] 3337.50 IOPS, 13.04 MiB/s [2024-12-06T16:42:22.663Z] 3359.67 IOPS, 13.12 MiB/s [2024-12-06T16:42:23.596Z] 3373.25 IOPS, 13.18 MiB/s [2024-12-06T16:42:24.528Z] 3364.00 IOPS, 13.14 MiB/s [2024-12-06T16:42:25.460Z] 3356.33 IOPS, 13.11 MiB/s [2024-12-06T16:42:26.390Z] 3357.14 IOPS, 13.11 MiB/s [2024-12-06T16:42:27.761Z] 3373.25 IOPS, 13.18 MiB/s [2024-12-06T16:42:28.694Z] 3378.44 IOPS, 13.20 MiB/s [2024-12-06T16:42:28.694Z] 3379.50 IOPS, 13.20 MiB/s 00:26:46.855 Latency(us) 00:26:46.855 [2024-12-06T16:42:28.694Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:46.855 Job: TLSTESTn1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:26:46.855 Verification LBA range: start 0x0 length 0x2000 00:26:46.855 TLSTESTn1 : 10.04 3377.57 13.19 0.00 0.00 37800.93 9611.95 39807.05 00:26:46.855 [2024-12-06T16:42:28.694Z] =================================================================================================================== 00:26:46.855 [2024-12-06T16:42:28.694Z] Total : 3377.57 13.19 0.00 0.00 37800.93 9611.95 39807.05 00:26:46.855 { 00:26:46.855 "results": [ 00:26:46.855 { 00:26:46.855 "job": "TLSTESTn1", 00:26:46.855 "core_mask": "0x4", 00:26:46.855 "workload": "verify", 00:26:46.855 "status": "finished", 00:26:46.855 "verify_range": { 00:26:46.855 "start": 0, 00:26:46.855 "length": 8192 00:26:46.855 }, 00:26:46.855 "queue_depth": 128, 00:26:46.855 "io_size": 4096, 00:26:46.855 "runtime": 10.043028, 00:26:46.855 "iops": 3377.5670046922105, 00:26:46.855 "mibps": 13.193621112078947, 00:26:46.855 "io_failed": 0, 00:26:46.855 "io_timeout": 0, 00:26:46.855 "avg_latency_us": 37800.92826368894, 00:26:46.855 "min_latency_us": 9611.946666666667, 00:26:46.855 "max_latency_us": 39807.05185185185 00:26:46.855 } 00:26:46.855 ], 00:26:46.855 "core_count": 1 00:26:46.855 } 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@45 -- # trap 'nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@46 -- # killprocess 286866 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 286866 ']' 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 286866 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 286866 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 286866' 00:26:46.855 killing process with pid 286866 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 286866 00:26:46.855 Received shutdown signal, test time was about 10.000000 seconds 00:26:46.855 00:26:46.855 Latency(us) 00:26:46.855 [2024-12-06T16:42:28.694Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:46.855 [2024-12-06T16:42:28.694Z] =================================================================================================================== 00:26:46.855 [2024-12-06T16:42:28.694Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 286866 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@171 -- # chmod 0666 /tmp/tmp.OtECQfGJ9H 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@172 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.OtECQfGJ9H 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@652 -- # local es=0 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@654 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.OtECQfGJ9H 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@640 -- # local arg=run_bdevperf 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # type -t run_bdevperf 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@655 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.OtECQfGJ9H 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # psk=/tmp/tmp.OtECQfGJ9H 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=288179 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 288179 /var/tmp/bdevperf.sock 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 288179 ']' 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:26:46.855 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:46.855 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:26:47.112 [2024-12-06 17:42:28.704763] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:26:47.112 [2024-12-06 17:42:28.704856] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid288179 ] 00:26:47.112 [2024-12-06 17:42:28.771936] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:47.112 [2024-12-06 17:42:28.820544] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:26:47.112 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:47.112 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:26:47.112 17:42:28 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.OtECQfGJ9H 00:26:47.677 [2024-12-06 17:42:29.230740] keyring.c: 36:keyring_file_check_path: *ERROR*: Invalid permissions for key file '/tmp/tmp.OtECQfGJ9H': 0100666 00:26:47.678 [2024-12-06 17:42:29.230782] keyring.c: 126:spdk_keyring_add_key: *ERROR*: Failed to add key 'key0' to the keyring 00:26:47.678 request: 00:26:47.678 { 00:26:47.678 "name": "key0", 00:26:47.678 "path": "/tmp/tmp.OtECQfGJ9H", 00:26:47.678 "method": "keyring_file_add_key", 00:26:47.678 "req_id": 1 00:26:47.678 } 00:26:47.678 Got JSON-RPC error response 00:26:47.678 response: 00:26:47.678 { 00:26:47.678 "code": -1, 00:26:47.678 "message": "Operation not permitted" 00:26:47.678 } 00:26:47.678 17:42:29 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk key0 00:26:47.678 [2024-12-06 17:42:29.495544] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:26:47.678 [2024-12-06 17:42:29.495608] bdev_nvme.c:6749:spdk_bdev_nvme_create: *ERROR*: Could not load PSK: key0 00:26:47.678 request: 00:26:47.678 { 00:26:47.678 "name": "TLSTEST", 00:26:47.678 "trtype": "tcp", 00:26:47.678 "traddr": "10.0.0.2", 00:26:47.678 "adrfam": "ipv4", 00:26:47.678 "trsvcid": "4420", 00:26:47.678 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:26:47.678 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:26:47.678 "prchk_reftag": false, 00:26:47.678 "prchk_guard": false, 00:26:47.678 "hdgst": false, 00:26:47.678 "ddgst": false, 00:26:47.678 "psk": "key0", 00:26:47.678 "allow_unrecognized_csi": false, 00:26:47.678 "method": "bdev_nvme_attach_controller", 00:26:47.678 "req_id": 1 00:26:47.678 } 00:26:47.678 Got JSON-RPC error response 00:26:47.678 response: 00:26:47.678 { 00:26:47.678 "code": -126, 00:26:47.678 "message": "Required key not available" 00:26:47.678 } 00:26:47.935 17:42:29 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@37 -- # killprocess 288179 00:26:47.935 17:42:29 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 288179 ']' 00:26:47.935 17:42:29 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 288179 00:26:47.935 17:42:29 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:26:47.935 17:42:29 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:47.935 17:42:29 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 288179 00:26:47.935 17:42:29 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:26:47.935 17:42:29 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:26:47.935 17:42:29 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 288179' 00:26:47.935 killing process with pid 288179 00:26:47.935 17:42:29 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 288179 00:26:47.935 Received shutdown signal, test time was about 10.000000 seconds 00:26:47.935 00:26:47.935 Latency(us) 00:26:47.935 [2024-12-06T16:42:29.774Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:47.935 [2024-12-06T16:42:29.774Z] =================================================================================================================== 00:26:47.935 [2024-12-06T16:42:29.774Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:26:47.935 17:42:29 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 288179 00:26:47.935 17:42:29 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@38 -- # return 1 00:26:47.935 17:42:29 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@655 -- # es=1 00:26:47.935 17:42:29 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:26:47.935 17:42:29 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:26:47.935 17:42:29 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:26:47.935 17:42:29 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@175 -- # killprocess 286573 00:26:47.935 17:42:29 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 286573 ']' 00:26:47.935 17:42:29 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 286573 00:26:47.935 17:42:29 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:26:47.935 17:42:29 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:47.935 17:42:29 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 286573 00:26:48.193 17:42:29 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:26:48.193 17:42:29 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:26:48.193 17:42:29 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 286573' 00:26:48.193 killing process with pid 286573 00:26:48.193 17:42:29 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 286573 00:26:48.193 17:42:29 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 286573 00:26:48.193 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@176 -- # nvmfappstart -m 0x2 00:26:48.193 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:26:48.193 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@726 -- # xtrace_disable 00:26:48.193 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:26:48.193 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # nvmfpid=288328 00:26:48.193 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:26:48.193 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # waitforlisten 288328 00:26:48.193 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 288328 ']' 00:26:48.193 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:48.193 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:48.193 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:48.193 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:48.193 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:48.193 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:26:48.451 [2024-12-06 17:42:30.063727] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:26:48.451 [2024-12-06 17:42:30.063826] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:48.451 [2024-12-06 17:42:30.140370] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:48.451 [2024-12-06 17:42:30.184983] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:26:48.451 [2024-12-06 17:42:30.185056] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:26:48.451 [2024-12-06 17:42:30.185080] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:26:48.451 [2024-12-06 17:42:30.185091] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:26:48.451 [2024-12-06 17:42:30.185117] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:26:48.451 [2024-12-06 17:42:30.185705] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:26:48.710 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:48.710 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:26:48.710 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:26:48.710 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@732 -- # xtrace_disable 00:26:48.710 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:26:48.710 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:26:48.710 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@178 -- # NOT setup_nvmf_tgt /tmp/tmp.OtECQfGJ9H 00:26:48.710 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@652 -- # local es=0 00:26:48.710 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@654 -- # valid_exec_arg setup_nvmf_tgt /tmp/tmp.OtECQfGJ9H 00:26:48.710 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@640 -- # local arg=setup_nvmf_tgt 00:26:48.710 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:26:48.710 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # type -t setup_nvmf_tgt 00:26:48.710 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:26:48.710 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@655 -- # setup_nvmf_tgt /tmp/tmp.OtECQfGJ9H 00:26:48.710 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@50 -- # local key=/tmp/tmp.OtECQfGJ9H 00:26:48.710 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:26:48.967 [2024-12-06 17:42:30.578886] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:26:48.967 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:26:49.223 17:42:30 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:26:49.481 [2024-12-06 17:42:31.164427] tcp.c:1049:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:26:49.481 [2024-12-06 17:42:31.164697] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:26:49.481 17:42:31 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:26:49.738 malloc0 00:26:49.738 17:42:31 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:26:49.996 17:42:31 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py keyring_file_add_key key0 /tmp/tmp.OtECQfGJ9H 00:26:50.254 [2024-12-06 17:42:32.057236] keyring.c: 36:keyring_file_check_path: *ERROR*: Invalid permissions for key file '/tmp/tmp.OtECQfGJ9H': 0100666 00:26:50.254 [2024-12-06 17:42:32.057279] keyring.c: 126:spdk_keyring_add_key: *ERROR*: Failed to add key 'key0' to the keyring 00:26:50.254 request: 00:26:50.254 { 00:26:50.254 "name": "key0", 00:26:50.254 "path": "/tmp/tmp.OtECQfGJ9H", 00:26:50.254 "method": "keyring_file_add_key", 00:26:50.254 "req_id": 1 00:26:50.254 } 00:26:50.254 Got JSON-RPC error response 00:26:50.254 response: 00:26:50.254 { 00:26:50.254 "code": -1, 00:26:50.254 "message": "Operation not permitted" 00:26:50.254 } 00:26:50.254 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk key0 00:26:50.511 [2024-12-06 17:42:32.330034] tcp.c:3777:nvmf_tcp_subsystem_add_host: *ERROR*: Key 'key0' does not exist 00:26:50.511 [2024-12-06 17:42:32.330085] subsystem.c:1051:spdk_nvmf_subsystem_add_host_ext: *ERROR*: Unable to add host to TCP transport 00:26:50.511 request: 00:26:50.511 { 00:26:50.511 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:26:50.511 "host": "nqn.2016-06.io.spdk:host1", 00:26:50.511 "psk": "key0", 00:26:50.511 "method": "nvmf_subsystem_add_host", 00:26:50.511 "req_id": 1 00:26:50.511 } 00:26:50.511 Got JSON-RPC error response 00:26:50.511 response: 00:26:50.511 { 00:26:50.511 "code": -32603, 00:26:50.511 "message": "Internal error" 00:26:50.511 } 00:26:50.769 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@655 -- # es=1 00:26:50.769 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:26:50.769 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:26:50.769 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:26:50.769 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@181 -- # killprocess 288328 00:26:50.769 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 288328 ']' 00:26:50.769 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 288328 00:26:50.769 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:26:50.769 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:50.769 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 288328 00:26:50.769 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:26:50.769 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:26:50.769 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 288328' 00:26:50.769 killing process with pid 288328 00:26:50.769 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 288328 00:26:50.769 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 288328 00:26:50.769 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@182 -- # chmod 0600 /tmp/tmp.OtECQfGJ9H 00:26:50.769 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@185 -- # nvmfappstart -m 0x2 00:26:50.769 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:26:50.769 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@726 -- # xtrace_disable 00:26:50.769 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:26:50.769 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # nvmfpid=288710 00:26:50.769 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:26:50.769 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # waitforlisten 288710 00:26:50.769 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 288710 ']' 00:26:50.769 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:50.769 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:50.770 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:50.770 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:50.770 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:50.770 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:26:51.028 [2024-12-06 17:42:32.656866] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:26:51.028 [2024-12-06 17:42:32.656964] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:51.028 [2024-12-06 17:42:32.729158] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:51.028 [2024-12-06 17:42:32.769936] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:26:51.028 [2024-12-06 17:42:32.770015] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:26:51.028 [2024-12-06 17:42:32.770037] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:26:51.028 [2024-12-06 17:42:32.770048] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:26:51.028 [2024-12-06 17:42:32.770057] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:26:51.028 [2024-12-06 17:42:32.770593] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:26:51.285 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:51.285 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:26:51.285 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:26:51.285 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@732 -- # xtrace_disable 00:26:51.285 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:26:51.285 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:26:51.285 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@186 -- # setup_nvmf_tgt /tmp/tmp.OtECQfGJ9H 00:26:51.285 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@50 -- # local key=/tmp/tmp.OtECQfGJ9H 00:26:51.286 17:42:32 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:26:51.543 [2024-12-06 17:42:33.142931] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:26:51.543 17:42:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:26:51.801 17:42:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:26:52.081 [2024-12-06 17:42:33.684415] tcp.c:1049:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:26:52.081 [2024-12-06 17:42:33.684704] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:26:52.081 17:42:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:26:52.396 malloc0 00:26:52.396 17:42:33 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:26:52.674 17:42:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py keyring_file_add_key key0 /tmp/tmp.OtECQfGJ9H 00:26:52.966 17:42:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk key0 00:26:52.966 17:42:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@189 -- # bdevperf_pid=288938 00:26:52.966 17:42:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@188 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:26:52.966 17:42:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@191 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:26:52.966 17:42:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@192 -- # waitforlisten 288938 /var/tmp/bdevperf.sock 00:26:52.966 17:42:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 288938 ']' 00:26:52.966 17:42:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:26:52.966 17:42:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:52.966 17:42:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:26:52.966 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:26:52.966 17:42:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:52.966 17:42:34 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:26:53.243 [2024-12-06 17:42:34.837319] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:26:53.243 [2024-12-06 17:42:34.837433] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid288938 ] 00:26:53.243 [2024-12-06 17:42:34.906065] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:53.243 [2024-12-06 17:42:34.951008] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:26:53.243 17:42:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:53.243 17:42:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:26:53.243 17:42:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@193 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.OtECQfGJ9H 00:26:53.515 17:42:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@194 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk key0 00:26:53.796 [2024-12-06 17:42:35.583850] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:26:54.054 TLSTESTn1 00:26:54.054 17:42:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@198 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py save_config 00:26:54.312 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@198 -- # tgtconf='{ 00:26:54.312 "subsystems": [ 00:26:54.312 { 00:26:54.312 "subsystem": "keyring", 00:26:54.312 "config": [ 00:26:54.312 { 00:26:54.312 "method": "keyring_file_add_key", 00:26:54.312 "params": { 00:26:54.312 "name": "key0", 00:26:54.312 "path": "/tmp/tmp.OtECQfGJ9H" 00:26:54.312 } 00:26:54.312 } 00:26:54.312 ] 00:26:54.312 }, 00:26:54.312 { 00:26:54.312 "subsystem": "iobuf", 00:26:54.312 "config": [ 00:26:54.312 { 00:26:54.312 "method": "iobuf_set_options", 00:26:54.312 "params": { 00:26:54.312 "small_pool_count": 8192, 00:26:54.312 "large_pool_count": 1024, 00:26:54.312 "small_bufsize": 8192, 00:26:54.312 "large_bufsize": 135168, 00:26:54.312 "enable_numa": false 00:26:54.312 } 00:26:54.312 } 00:26:54.312 ] 00:26:54.312 }, 00:26:54.312 { 00:26:54.312 "subsystem": "sock", 00:26:54.312 "config": [ 00:26:54.312 { 00:26:54.312 "method": "sock_set_default_impl", 00:26:54.312 "params": { 00:26:54.312 "impl_name": "posix" 00:26:54.312 } 00:26:54.312 }, 00:26:54.312 { 00:26:54.312 "method": "sock_impl_set_options", 00:26:54.312 "params": { 00:26:54.312 "impl_name": "ssl", 00:26:54.312 "recv_buf_size": 4096, 00:26:54.312 "send_buf_size": 4096, 00:26:54.312 "enable_recv_pipe": true, 00:26:54.312 "enable_quickack": false, 00:26:54.312 "enable_placement_id": 0, 00:26:54.312 "enable_zerocopy_send_server": true, 00:26:54.312 "enable_zerocopy_send_client": false, 00:26:54.312 "zerocopy_threshold": 0, 00:26:54.312 "tls_version": 0, 00:26:54.312 "enable_ktls": false 00:26:54.312 } 00:26:54.312 }, 00:26:54.312 { 00:26:54.312 "method": "sock_impl_set_options", 00:26:54.312 "params": { 00:26:54.312 "impl_name": "posix", 00:26:54.312 "recv_buf_size": 2097152, 00:26:54.312 "send_buf_size": 2097152, 00:26:54.312 "enable_recv_pipe": true, 00:26:54.312 "enable_quickack": false, 00:26:54.312 "enable_placement_id": 0, 00:26:54.312 "enable_zerocopy_send_server": true, 00:26:54.312 "enable_zerocopy_send_client": false, 00:26:54.312 "zerocopy_threshold": 0, 00:26:54.312 "tls_version": 0, 00:26:54.312 "enable_ktls": false 00:26:54.312 } 00:26:54.312 } 00:26:54.312 ] 00:26:54.312 }, 00:26:54.312 { 00:26:54.312 "subsystem": "vmd", 00:26:54.312 "config": [] 00:26:54.312 }, 00:26:54.312 { 00:26:54.312 "subsystem": "accel", 00:26:54.312 "config": [ 00:26:54.312 { 00:26:54.312 "method": "accel_set_options", 00:26:54.312 "params": { 00:26:54.312 "small_cache_size": 128, 00:26:54.312 "large_cache_size": 16, 00:26:54.312 "task_count": 2048, 00:26:54.312 "sequence_count": 2048, 00:26:54.312 "buf_count": 2048 00:26:54.312 } 00:26:54.312 } 00:26:54.312 ] 00:26:54.312 }, 00:26:54.312 { 00:26:54.312 "subsystem": "bdev", 00:26:54.312 "config": [ 00:26:54.312 { 00:26:54.312 "method": "bdev_set_options", 00:26:54.312 "params": { 00:26:54.312 "bdev_io_pool_size": 65535, 00:26:54.312 "bdev_io_cache_size": 256, 00:26:54.312 "bdev_auto_examine": true, 00:26:54.312 "iobuf_small_cache_size": 128, 00:26:54.312 "iobuf_large_cache_size": 16 00:26:54.312 } 00:26:54.312 }, 00:26:54.312 { 00:26:54.312 "method": "bdev_raid_set_options", 00:26:54.312 "params": { 00:26:54.312 "process_window_size_kb": 1024, 00:26:54.312 "process_max_bandwidth_mb_sec": 0 00:26:54.312 } 00:26:54.312 }, 00:26:54.312 { 00:26:54.312 "method": "bdev_iscsi_set_options", 00:26:54.312 "params": { 00:26:54.313 "timeout_sec": 30 00:26:54.313 } 00:26:54.313 }, 00:26:54.313 { 00:26:54.313 "method": "bdev_nvme_set_options", 00:26:54.313 "params": { 00:26:54.313 "action_on_timeout": "none", 00:26:54.313 "timeout_us": 0, 00:26:54.313 "timeout_admin_us": 0, 00:26:54.313 "keep_alive_timeout_ms": 10000, 00:26:54.313 "arbitration_burst": 0, 00:26:54.313 "low_priority_weight": 0, 00:26:54.313 "medium_priority_weight": 0, 00:26:54.313 "high_priority_weight": 0, 00:26:54.313 "nvme_adminq_poll_period_us": 10000, 00:26:54.313 "nvme_ioq_poll_period_us": 0, 00:26:54.313 "io_queue_requests": 0, 00:26:54.313 "delay_cmd_submit": true, 00:26:54.313 "transport_retry_count": 4, 00:26:54.313 "bdev_retry_count": 3, 00:26:54.313 "transport_ack_timeout": 0, 00:26:54.313 "ctrlr_loss_timeout_sec": 0, 00:26:54.313 "reconnect_delay_sec": 0, 00:26:54.313 "fast_io_fail_timeout_sec": 0, 00:26:54.313 "disable_auto_failback": false, 00:26:54.313 "generate_uuids": false, 00:26:54.313 "transport_tos": 0, 00:26:54.313 "nvme_error_stat": false, 00:26:54.313 "rdma_srq_size": 0, 00:26:54.313 "io_path_stat": false, 00:26:54.313 "allow_accel_sequence": false, 00:26:54.313 "rdma_max_cq_size": 0, 00:26:54.313 "rdma_cm_event_timeout_ms": 0, 00:26:54.313 "dhchap_digests": [ 00:26:54.313 "sha256", 00:26:54.313 "sha384", 00:26:54.313 "sha512" 00:26:54.313 ], 00:26:54.313 "dhchap_dhgroups": [ 00:26:54.313 "null", 00:26:54.313 "ffdhe2048", 00:26:54.313 "ffdhe3072", 00:26:54.313 "ffdhe4096", 00:26:54.313 "ffdhe6144", 00:26:54.313 "ffdhe8192" 00:26:54.313 ] 00:26:54.313 } 00:26:54.313 }, 00:26:54.313 { 00:26:54.313 "method": "bdev_nvme_set_hotplug", 00:26:54.313 "params": { 00:26:54.313 "period_us": 100000, 00:26:54.313 "enable": false 00:26:54.313 } 00:26:54.313 }, 00:26:54.313 { 00:26:54.313 "method": "bdev_malloc_create", 00:26:54.313 "params": { 00:26:54.313 "name": "malloc0", 00:26:54.313 "num_blocks": 8192, 00:26:54.313 "block_size": 4096, 00:26:54.313 "physical_block_size": 4096, 00:26:54.313 "uuid": "8bfa5834-5ad3-40f1-8373-0cc1f0ddcc04", 00:26:54.313 "optimal_io_boundary": 0, 00:26:54.313 "md_size": 0, 00:26:54.313 "dif_type": 0, 00:26:54.313 "dif_is_head_of_md": false, 00:26:54.313 "dif_pi_format": 0 00:26:54.313 } 00:26:54.313 }, 00:26:54.313 { 00:26:54.313 "method": "bdev_wait_for_examine" 00:26:54.313 } 00:26:54.313 ] 00:26:54.313 }, 00:26:54.313 { 00:26:54.313 "subsystem": "nbd", 00:26:54.313 "config": [] 00:26:54.313 }, 00:26:54.313 { 00:26:54.313 "subsystem": "scheduler", 00:26:54.313 "config": [ 00:26:54.313 { 00:26:54.313 "method": "framework_set_scheduler", 00:26:54.313 "params": { 00:26:54.313 "name": "static" 00:26:54.313 } 00:26:54.313 } 00:26:54.313 ] 00:26:54.313 }, 00:26:54.313 { 00:26:54.313 "subsystem": "nvmf", 00:26:54.313 "config": [ 00:26:54.313 { 00:26:54.313 "method": "nvmf_set_config", 00:26:54.313 "params": { 00:26:54.313 "discovery_filter": "match_any", 00:26:54.313 "admin_cmd_passthru": { 00:26:54.313 "identify_ctrlr": false 00:26:54.313 }, 00:26:54.313 "dhchap_digests": [ 00:26:54.313 "sha256", 00:26:54.313 "sha384", 00:26:54.313 "sha512" 00:26:54.313 ], 00:26:54.313 "dhchap_dhgroups": [ 00:26:54.313 "null", 00:26:54.313 "ffdhe2048", 00:26:54.313 "ffdhe3072", 00:26:54.313 "ffdhe4096", 00:26:54.313 "ffdhe6144", 00:26:54.313 "ffdhe8192" 00:26:54.313 ] 00:26:54.313 } 00:26:54.313 }, 00:26:54.313 { 00:26:54.313 "method": "nvmf_set_max_subsystems", 00:26:54.313 "params": { 00:26:54.313 "max_subsystems": 1024 00:26:54.313 } 00:26:54.313 }, 00:26:54.313 { 00:26:54.313 "method": "nvmf_set_crdt", 00:26:54.313 "params": { 00:26:54.313 "crdt1": 0, 00:26:54.313 "crdt2": 0, 00:26:54.313 "crdt3": 0 00:26:54.313 } 00:26:54.313 }, 00:26:54.313 { 00:26:54.313 "method": "nvmf_create_transport", 00:26:54.313 "params": { 00:26:54.313 "trtype": "TCP", 00:26:54.313 "max_queue_depth": 128, 00:26:54.313 "max_io_qpairs_per_ctrlr": 127, 00:26:54.313 "in_capsule_data_size": 4096, 00:26:54.313 "max_io_size": 131072, 00:26:54.313 "io_unit_size": 131072, 00:26:54.313 "max_aq_depth": 128, 00:26:54.313 "num_shared_buffers": 511, 00:26:54.313 "buf_cache_size": 4294967295, 00:26:54.313 "dif_insert_or_strip": false, 00:26:54.313 "zcopy": false, 00:26:54.313 "c2h_success": false, 00:26:54.313 "sock_priority": 0, 00:26:54.313 "abort_timeout_sec": 1, 00:26:54.313 "ack_timeout": 0, 00:26:54.313 "data_wr_pool_size": 0 00:26:54.313 } 00:26:54.313 }, 00:26:54.313 { 00:26:54.313 "method": "nvmf_create_subsystem", 00:26:54.313 "params": { 00:26:54.313 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:26:54.313 "allow_any_host": false, 00:26:54.313 "serial_number": "SPDK00000000000001", 00:26:54.313 "model_number": "SPDK bdev Controller", 00:26:54.313 "max_namespaces": 10, 00:26:54.313 "min_cntlid": 1, 00:26:54.313 "max_cntlid": 65519, 00:26:54.313 "ana_reporting": false 00:26:54.313 } 00:26:54.313 }, 00:26:54.313 { 00:26:54.313 "method": "nvmf_subsystem_add_host", 00:26:54.313 "params": { 00:26:54.313 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:26:54.313 "host": "nqn.2016-06.io.spdk:host1", 00:26:54.313 "psk": "key0" 00:26:54.313 } 00:26:54.313 }, 00:26:54.313 { 00:26:54.313 "method": "nvmf_subsystem_add_ns", 00:26:54.313 "params": { 00:26:54.313 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:26:54.313 "namespace": { 00:26:54.313 "nsid": 1, 00:26:54.313 "bdev_name": "malloc0", 00:26:54.313 "nguid": "8BFA58345AD340F183730CC1F0DDCC04", 00:26:54.313 "uuid": "8bfa5834-5ad3-40f1-8373-0cc1f0ddcc04", 00:26:54.313 "no_auto_visible": false 00:26:54.313 } 00:26:54.313 } 00:26:54.313 }, 00:26:54.313 { 00:26:54.313 "method": "nvmf_subsystem_add_listener", 00:26:54.313 "params": { 00:26:54.313 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:26:54.313 "listen_address": { 00:26:54.313 "trtype": "TCP", 00:26:54.313 "adrfam": "IPv4", 00:26:54.313 "traddr": "10.0.0.2", 00:26:54.313 "trsvcid": "4420" 00:26:54.313 }, 00:26:54.313 "secure_channel": true 00:26:54.313 } 00:26:54.313 } 00:26:54.313 ] 00:26:54.313 } 00:26:54.313 ] 00:26:54.313 }' 00:26:54.313 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@199 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock save_config 00:26:54.572 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@199 -- # bdevperfconf='{ 00:26:54.572 "subsystems": [ 00:26:54.572 { 00:26:54.572 "subsystem": "keyring", 00:26:54.572 "config": [ 00:26:54.572 { 00:26:54.572 "method": "keyring_file_add_key", 00:26:54.572 "params": { 00:26:54.572 "name": "key0", 00:26:54.572 "path": "/tmp/tmp.OtECQfGJ9H" 00:26:54.572 } 00:26:54.572 } 00:26:54.572 ] 00:26:54.572 }, 00:26:54.572 { 00:26:54.572 "subsystem": "iobuf", 00:26:54.572 "config": [ 00:26:54.572 { 00:26:54.572 "method": "iobuf_set_options", 00:26:54.572 "params": { 00:26:54.572 "small_pool_count": 8192, 00:26:54.572 "large_pool_count": 1024, 00:26:54.572 "small_bufsize": 8192, 00:26:54.572 "large_bufsize": 135168, 00:26:54.572 "enable_numa": false 00:26:54.572 } 00:26:54.572 } 00:26:54.572 ] 00:26:54.572 }, 00:26:54.572 { 00:26:54.572 "subsystem": "sock", 00:26:54.572 "config": [ 00:26:54.572 { 00:26:54.572 "method": "sock_set_default_impl", 00:26:54.572 "params": { 00:26:54.572 "impl_name": "posix" 00:26:54.572 } 00:26:54.572 }, 00:26:54.572 { 00:26:54.572 "method": "sock_impl_set_options", 00:26:54.572 "params": { 00:26:54.572 "impl_name": "ssl", 00:26:54.572 "recv_buf_size": 4096, 00:26:54.572 "send_buf_size": 4096, 00:26:54.572 "enable_recv_pipe": true, 00:26:54.572 "enable_quickack": false, 00:26:54.572 "enable_placement_id": 0, 00:26:54.572 "enable_zerocopy_send_server": true, 00:26:54.572 "enable_zerocopy_send_client": false, 00:26:54.572 "zerocopy_threshold": 0, 00:26:54.572 "tls_version": 0, 00:26:54.572 "enable_ktls": false 00:26:54.572 } 00:26:54.572 }, 00:26:54.572 { 00:26:54.572 "method": "sock_impl_set_options", 00:26:54.572 "params": { 00:26:54.572 "impl_name": "posix", 00:26:54.572 "recv_buf_size": 2097152, 00:26:54.572 "send_buf_size": 2097152, 00:26:54.572 "enable_recv_pipe": true, 00:26:54.572 "enable_quickack": false, 00:26:54.572 "enable_placement_id": 0, 00:26:54.572 "enable_zerocopy_send_server": true, 00:26:54.572 "enable_zerocopy_send_client": false, 00:26:54.572 "zerocopy_threshold": 0, 00:26:54.572 "tls_version": 0, 00:26:54.572 "enable_ktls": false 00:26:54.572 } 00:26:54.572 } 00:26:54.572 ] 00:26:54.572 }, 00:26:54.572 { 00:26:54.572 "subsystem": "vmd", 00:26:54.572 "config": [] 00:26:54.572 }, 00:26:54.572 { 00:26:54.572 "subsystem": "accel", 00:26:54.572 "config": [ 00:26:54.572 { 00:26:54.572 "method": "accel_set_options", 00:26:54.572 "params": { 00:26:54.572 "small_cache_size": 128, 00:26:54.572 "large_cache_size": 16, 00:26:54.572 "task_count": 2048, 00:26:54.572 "sequence_count": 2048, 00:26:54.572 "buf_count": 2048 00:26:54.572 } 00:26:54.572 } 00:26:54.572 ] 00:26:54.572 }, 00:26:54.572 { 00:26:54.572 "subsystem": "bdev", 00:26:54.572 "config": [ 00:26:54.572 { 00:26:54.572 "method": "bdev_set_options", 00:26:54.572 "params": { 00:26:54.572 "bdev_io_pool_size": 65535, 00:26:54.572 "bdev_io_cache_size": 256, 00:26:54.572 "bdev_auto_examine": true, 00:26:54.572 "iobuf_small_cache_size": 128, 00:26:54.572 "iobuf_large_cache_size": 16 00:26:54.572 } 00:26:54.572 }, 00:26:54.572 { 00:26:54.572 "method": "bdev_raid_set_options", 00:26:54.572 "params": { 00:26:54.572 "process_window_size_kb": 1024, 00:26:54.572 "process_max_bandwidth_mb_sec": 0 00:26:54.572 } 00:26:54.572 }, 00:26:54.572 { 00:26:54.572 "method": "bdev_iscsi_set_options", 00:26:54.572 "params": { 00:26:54.572 "timeout_sec": 30 00:26:54.572 } 00:26:54.572 }, 00:26:54.572 { 00:26:54.572 "method": "bdev_nvme_set_options", 00:26:54.572 "params": { 00:26:54.572 "action_on_timeout": "none", 00:26:54.572 "timeout_us": 0, 00:26:54.572 "timeout_admin_us": 0, 00:26:54.572 "keep_alive_timeout_ms": 10000, 00:26:54.572 "arbitration_burst": 0, 00:26:54.572 "low_priority_weight": 0, 00:26:54.572 "medium_priority_weight": 0, 00:26:54.572 "high_priority_weight": 0, 00:26:54.572 "nvme_adminq_poll_period_us": 10000, 00:26:54.572 "nvme_ioq_poll_period_us": 0, 00:26:54.572 "io_queue_requests": 512, 00:26:54.572 "delay_cmd_submit": true, 00:26:54.572 "transport_retry_count": 4, 00:26:54.572 "bdev_retry_count": 3, 00:26:54.572 "transport_ack_timeout": 0, 00:26:54.572 "ctrlr_loss_timeout_sec": 0, 00:26:54.572 "reconnect_delay_sec": 0, 00:26:54.572 "fast_io_fail_timeout_sec": 0, 00:26:54.572 "disable_auto_failback": false, 00:26:54.572 "generate_uuids": false, 00:26:54.572 "transport_tos": 0, 00:26:54.572 "nvme_error_stat": false, 00:26:54.572 "rdma_srq_size": 0, 00:26:54.572 "io_path_stat": false, 00:26:54.572 "allow_accel_sequence": false, 00:26:54.572 "rdma_max_cq_size": 0, 00:26:54.572 "rdma_cm_event_timeout_ms": 0, 00:26:54.572 "dhchap_digests": [ 00:26:54.572 "sha256", 00:26:54.572 "sha384", 00:26:54.572 "sha512" 00:26:54.572 ], 00:26:54.572 "dhchap_dhgroups": [ 00:26:54.572 "null", 00:26:54.572 "ffdhe2048", 00:26:54.572 "ffdhe3072", 00:26:54.572 "ffdhe4096", 00:26:54.572 "ffdhe6144", 00:26:54.572 "ffdhe8192" 00:26:54.572 ] 00:26:54.572 } 00:26:54.572 }, 00:26:54.572 { 00:26:54.572 "method": "bdev_nvme_attach_controller", 00:26:54.572 "params": { 00:26:54.572 "name": "TLSTEST", 00:26:54.572 "trtype": "TCP", 00:26:54.572 "adrfam": "IPv4", 00:26:54.572 "traddr": "10.0.0.2", 00:26:54.572 "trsvcid": "4420", 00:26:54.572 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:26:54.572 "prchk_reftag": false, 00:26:54.572 "prchk_guard": false, 00:26:54.572 "ctrlr_loss_timeout_sec": 0, 00:26:54.572 "reconnect_delay_sec": 0, 00:26:54.572 "fast_io_fail_timeout_sec": 0, 00:26:54.572 "psk": "key0", 00:26:54.572 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:26:54.572 "hdgst": false, 00:26:54.572 "ddgst": false, 00:26:54.572 "multipath": "multipath" 00:26:54.572 } 00:26:54.572 }, 00:26:54.572 { 00:26:54.572 "method": "bdev_nvme_set_hotplug", 00:26:54.572 "params": { 00:26:54.572 "period_us": 100000, 00:26:54.572 "enable": false 00:26:54.572 } 00:26:54.573 }, 00:26:54.573 { 00:26:54.573 "method": "bdev_wait_for_examine" 00:26:54.573 } 00:26:54.573 ] 00:26:54.573 }, 00:26:54.573 { 00:26:54.573 "subsystem": "nbd", 00:26:54.573 "config": [] 00:26:54.573 } 00:26:54.573 ] 00:26:54.573 }' 00:26:54.573 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@201 -- # killprocess 288938 00:26:54.573 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 288938 ']' 00:26:54.573 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 288938 00:26:54.573 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:26:54.573 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:54.573 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 288938 00:26:54.573 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:26:54.573 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:26:54.573 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 288938' 00:26:54.573 killing process with pid 288938 00:26:54.573 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 288938 00:26:54.573 Received shutdown signal, test time was about 10.000000 seconds 00:26:54.573 00:26:54.573 Latency(us) 00:26:54.573 [2024-12-06T16:42:36.412Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:54.573 [2024-12-06T16:42:36.412Z] =================================================================================================================== 00:26:54.573 [2024-12-06T16:42:36.412Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:26:54.573 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 288938 00:26:54.830 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@202 -- # killprocess 288710 00:26:54.830 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 288710 ']' 00:26:54.830 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 288710 00:26:54.830 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:26:54.830 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:26:54.830 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 288710 00:26:54.830 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:26:54.830 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:26:54.831 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 288710' 00:26:54.831 killing process with pid 288710 00:26:54.831 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 288710 00:26:54.831 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 288710 00:26:55.089 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@205 -- # nvmfappstart -m 0x2 -c /dev/fd/62 00:26:55.089 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:26:55.089 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@205 -- # echo '{ 00:26:55.089 "subsystems": [ 00:26:55.089 { 00:26:55.089 "subsystem": "keyring", 00:26:55.089 "config": [ 00:26:55.089 { 00:26:55.089 "method": "keyring_file_add_key", 00:26:55.089 "params": { 00:26:55.089 "name": "key0", 00:26:55.089 "path": "/tmp/tmp.OtECQfGJ9H" 00:26:55.089 } 00:26:55.089 } 00:26:55.089 ] 00:26:55.089 }, 00:26:55.089 { 00:26:55.089 "subsystem": "iobuf", 00:26:55.089 "config": [ 00:26:55.089 { 00:26:55.089 "method": "iobuf_set_options", 00:26:55.089 "params": { 00:26:55.089 "small_pool_count": 8192, 00:26:55.089 "large_pool_count": 1024, 00:26:55.089 "small_bufsize": 8192, 00:26:55.089 "large_bufsize": 135168, 00:26:55.089 "enable_numa": false 00:26:55.089 } 00:26:55.089 } 00:26:55.089 ] 00:26:55.089 }, 00:26:55.089 { 00:26:55.089 "subsystem": "sock", 00:26:55.089 "config": [ 00:26:55.089 { 00:26:55.089 "method": "sock_set_default_impl", 00:26:55.089 "params": { 00:26:55.089 "impl_name": "posix" 00:26:55.089 } 00:26:55.089 }, 00:26:55.089 { 00:26:55.089 "method": "sock_impl_set_options", 00:26:55.089 "params": { 00:26:55.089 "impl_name": "ssl", 00:26:55.089 "recv_buf_size": 4096, 00:26:55.089 "send_buf_size": 4096, 00:26:55.089 "enable_recv_pipe": true, 00:26:55.089 "enable_quickack": false, 00:26:55.089 "enable_placement_id": 0, 00:26:55.089 "enable_zerocopy_send_server": true, 00:26:55.089 "enable_zerocopy_send_client": false, 00:26:55.089 "zerocopy_threshold": 0, 00:26:55.089 "tls_version": 0, 00:26:55.089 "enable_ktls": false 00:26:55.089 } 00:26:55.089 }, 00:26:55.089 { 00:26:55.089 "method": "sock_impl_set_options", 00:26:55.089 "params": { 00:26:55.089 "impl_name": "posix", 00:26:55.089 "recv_buf_size": 2097152, 00:26:55.089 "send_buf_size": 2097152, 00:26:55.089 "enable_recv_pipe": true, 00:26:55.089 "enable_quickack": false, 00:26:55.089 "enable_placement_id": 0, 00:26:55.089 "enable_zerocopy_send_server": true, 00:26:55.089 "enable_zerocopy_send_client": false, 00:26:55.089 "zerocopy_threshold": 0, 00:26:55.089 "tls_version": 0, 00:26:55.089 "enable_ktls": false 00:26:55.089 } 00:26:55.089 } 00:26:55.089 ] 00:26:55.089 }, 00:26:55.089 { 00:26:55.089 "subsystem": "vmd", 00:26:55.089 "config": [] 00:26:55.089 }, 00:26:55.089 { 00:26:55.089 "subsystem": "accel", 00:26:55.089 "config": [ 00:26:55.089 { 00:26:55.089 "method": "accel_set_options", 00:26:55.089 "params": { 00:26:55.089 "small_cache_size": 128, 00:26:55.089 "large_cache_size": 16, 00:26:55.089 "task_count": 2048, 00:26:55.089 "sequence_count": 2048, 00:26:55.089 "buf_count": 2048 00:26:55.089 } 00:26:55.089 } 00:26:55.089 ] 00:26:55.089 }, 00:26:55.089 { 00:26:55.089 "subsystem": "bdev", 00:26:55.089 "config": [ 00:26:55.089 { 00:26:55.089 "method": "bdev_set_options", 00:26:55.089 "params": { 00:26:55.089 "bdev_io_pool_size": 65535, 00:26:55.089 "bdev_io_cache_size": 256, 00:26:55.089 "bdev_auto_examine": true, 00:26:55.089 "iobuf_small_cache_size": 128, 00:26:55.089 "iobuf_large_cache_size": 16 00:26:55.089 } 00:26:55.089 }, 00:26:55.089 { 00:26:55.089 "method": "bdev_raid_set_options", 00:26:55.089 "params": { 00:26:55.089 "process_window_size_kb": 1024, 00:26:55.089 "process_max_bandwidth_mb_sec": 0 00:26:55.089 } 00:26:55.089 }, 00:26:55.089 { 00:26:55.089 "method": "bdev_iscsi_set_options", 00:26:55.089 "params": { 00:26:55.089 "timeout_sec": 30 00:26:55.089 } 00:26:55.089 }, 00:26:55.089 { 00:26:55.089 "method": "bdev_nvme_set_options", 00:26:55.089 "params": { 00:26:55.089 "action_on_timeout": "none", 00:26:55.089 "timeout_us": 0, 00:26:55.089 "timeout_admin_us": 0, 00:26:55.089 "keep_alive_timeout_ms": 10000, 00:26:55.089 "arbitration_burst": 0, 00:26:55.089 "low_priority_weight": 0, 00:26:55.089 "medium_priority_weight": 0, 00:26:55.089 "high_priority_weight": 0, 00:26:55.089 "nvme_adminq_poll_period_us": 10000, 00:26:55.089 "nvme_ioq_poll_period_us": 0, 00:26:55.089 "io_queue_requests": 0, 00:26:55.089 "delay_cmd_submit": true, 00:26:55.089 "transport_retry_count": 4, 00:26:55.090 "bdev_retry_count": 3, 00:26:55.090 "transport_ack_timeout": 0, 00:26:55.090 "ctrlr_loss_timeout_sec": 0, 00:26:55.090 "reconnect_delay_sec": 0, 00:26:55.090 "fast_io_fail_timeout_sec": 0, 00:26:55.090 "disable_auto_failback": false, 00:26:55.090 "generate_uuids": false, 00:26:55.090 "transport_tos": 0, 00:26:55.090 "nvme_error_stat": false, 00:26:55.090 "rdma_srq_size": 0, 00:26:55.090 "io_path_stat": false, 00:26:55.090 "allow_accel_sequence": false, 00:26:55.090 "rdma_max_cq_size": 0, 00:26:55.090 "rdma_cm_event_timeout_ms": 0, 00:26:55.090 "dhchap_digests": [ 00:26:55.090 "sha256", 00:26:55.090 "sha384", 00:26:55.090 "sha512" 00:26:55.090 ], 00:26:55.090 "dhchap_dhgroups": [ 00:26:55.090 "null", 00:26:55.090 "ffdhe2048", 00:26:55.090 "ffdhe3072", 00:26:55.090 "ffdhe4096", 00:26:55.090 "ffdhe6144", 00:26:55.090 "ffdhe8192" 00:26:55.090 ] 00:26:55.090 } 00:26:55.090 }, 00:26:55.090 { 00:26:55.090 "method": "bdev_nvme_set_hotplug", 00:26:55.090 "params": { 00:26:55.090 "period_us": 100000, 00:26:55.090 "enable": false 00:26:55.090 } 00:26:55.090 }, 00:26:55.090 { 00:26:55.090 "method": "bdev_malloc_create", 00:26:55.090 "params": { 00:26:55.090 "name": "malloc0", 00:26:55.090 "num_blocks": 8192, 00:26:55.090 "block_size": 4096, 00:26:55.090 "physical_block_size": 4096, 00:26:55.090 "uuid": "8bfa5834-5ad3-40f1-8373-0cc1f0ddcc04", 00:26:55.090 "optimal_io_boundary": 0, 00:26:55.090 "md_size": 0, 00:26:55.090 "dif_type": 0, 00:26:55.090 "dif_is_head_of_md": false, 00:26:55.090 "dif_pi_format": 0 00:26:55.090 } 00:26:55.090 }, 00:26:55.090 { 00:26:55.090 "method": "bdev_wait_for_examine" 00:26:55.090 } 00:26:55.090 ] 00:26:55.090 }, 00:26:55.090 { 00:26:55.090 "subsystem": "nbd", 00:26:55.090 "config": [] 00:26:55.090 }, 00:26:55.090 { 00:26:55.090 "subsystem": "scheduler", 00:26:55.090 "config": [ 00:26:55.090 { 00:26:55.090 "method": "framework_set_scheduler", 00:26:55.090 "params": { 00:26:55.090 "name": "static" 00:26:55.090 } 00:26:55.090 } 00:26:55.090 ] 00:26:55.090 }, 00:26:55.090 { 00:26:55.090 "subsystem": "nvmf", 00:26:55.090 "config": [ 00:26:55.090 { 00:26:55.090 "method": "nvmf_set_config", 00:26:55.090 "params": { 00:26:55.090 "discovery_filter": "match_any", 00:26:55.090 "admin_cmd_passthru": { 00:26:55.090 "identify_ctrlr": false 00:26:55.090 }, 00:26:55.090 "dhchap_digests": [ 00:26:55.090 "sha256", 00:26:55.090 "sha384", 00:26:55.090 "sha512" 00:26:55.090 ], 00:26:55.090 "dhchap_dhgroups": [ 00:26:55.090 "null", 00:26:55.090 "ffdhe2048", 00:26:55.090 "ffdhe3072", 00:26:55.090 "ffdhe4096", 00:26:55.090 "ffdhe6144", 00:26:55.090 "ffdhe8192" 00:26:55.090 ] 00:26:55.090 } 00:26:55.090 }, 00:26:55.090 { 00:26:55.090 "method": "nvmf_set_max_subsystems", 00:26:55.090 "params": { 00:26:55.090 "max_subsystems": 1024 00:26:55.090 } 00:26:55.090 }, 00:26:55.090 { 00:26:55.090 "method": "nvmf_set_crdt", 00:26:55.090 "params": { 00:26:55.090 "crdt1": 0, 00:26:55.090 "crdt2": 0, 00:26:55.090 "crdt3": 0 00:26:55.090 } 00:26:55.090 }, 00:26:55.090 { 00:26:55.090 "method": "nvmf_create_transport", 00:26:55.090 "params": { 00:26:55.090 "trtype": "TCP", 00:26:55.090 "max_queue_depth": 128, 00:26:55.090 "max_io_qpairs_per_ctrlr": 127, 00:26:55.090 "in_capsule_data_size": 4096, 00:26:55.090 "max_io_size": 131072, 00:26:55.090 "io_unit_size": 131072, 00:26:55.090 "max_aq_depth": 128, 00:26:55.090 "num_shared_buffers": 511, 00:26:55.090 "buf_cache_size": 4294967295, 00:26:55.090 "dif_insert_or_strip": false, 00:26:55.090 "zcopy": false, 00:26:55.090 "c2h_success": false, 00:26:55.090 "sock_priority": 0, 00:26:55.090 "abort_timeout_sec": 1, 00:26:55.090 "ack_timeout": 0, 00:26:55.090 "data_wr_pool_size": 0 00:26:55.090 } 00:26:55.090 }, 00:26:55.090 { 00:26:55.090 "method": "nvmf_create_subsystem", 00:26:55.090 "params": { 00:26:55.090 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:26:55.090 "allow_any_host": false, 00:26:55.090 "serial_number": "SPDK00000000000001", 00:26:55.090 "model_number": "SPDK bdev Controller", 00:26:55.090 "max_namespaces": 10, 00:26:55.090 "min_cntlid": 1, 00:26:55.090 "max_cntlid": 65519, 00:26:55.090 "ana_reporting": false 00:26:55.090 } 00:26:55.090 }, 00:26:55.090 { 00:26:55.090 "method": "nvmf_subsystem_add_host", 00:26:55.090 "params": { 00:26:55.090 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:26:55.090 "host": "nqn.2016-06.io.spdk:host1", 00:26:55.090 "psk": "key0" 00:26:55.090 } 00:26:55.090 }, 00:26:55.090 { 00:26:55.090 "method": "nvmf_subsystem_add_ns", 00:26:55.090 "params": { 00:26:55.090 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:26:55.090 "namespace": { 00:26:55.090 "nsid": 1, 00:26:55.090 "bdev_name": "malloc0", 00:26:55.090 "nguid": "8BFA58345AD340F183730CC1F0DDCC04", 00:26:55.090 "uuid": "8bfa5834-5ad3-40f1-8373-0cc1f0ddcc04", 00:26:55.090 "no_auto_visible": false 00:26:55.090 } 00:26:55.090 } 00:26:55.090 }, 00:26:55.090 { 00:26:55.090 "method": "nvmf_subsystem_add_listener", 00:26:55.090 "params": { 00:26:55.090 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:26:55.090 "listen_address": { 00:26:55.090 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@726 -- # xtrace_disable 00:26:55.090 "trtype": "TCP", 00:26:55.090 "adrfam": "IPv4", 00:26:55.090 "traddr": "10.0.0.2", 00:26:55.090 "trsvcid": "4420" 00:26:55.090 }, 00:26:55.090 "secure_channel": true 00:26:55.090 } 00:26:55.090 } 00:26:55.090 ] 00:26:55.090 } 00:26:55.090 ] 00:26:55.090 }' 00:26:55.090 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:26:55.090 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # nvmfpid=289207 00:26:55.090 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 -c /dev/fd/62 00:26:55.090 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # waitforlisten 289207 00:26:55.090 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 289207 ']' 00:26:55.090 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:55.090 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:55.090 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:55.090 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:55.090 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:55.090 17:42:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:26:55.090 [2024-12-06 17:42:36.892210] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:26:55.090 [2024-12-06 17:42:36.892306] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:55.348 [2024-12-06 17:42:36.966167] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:55.348 [2024-12-06 17:42:37.009681] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:26:55.348 [2024-12-06 17:42:37.009737] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:26:55.348 [2024-12-06 17:42:37.009760] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:26:55.348 [2024-12-06 17:42:37.009771] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:26:55.348 [2024-12-06 17:42:37.009781] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:26:55.348 [2024-12-06 17:42:37.010378] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:26:55.606 [2024-12-06 17:42:37.255200] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:26:55.606 [2024-12-06 17:42:37.287238] tcp.c:1049:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:26:55.606 [2024-12-06 17:42:37.287483] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:26:56.171 17:42:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:56.171 17:42:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:26:56.171 17:42:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:26:56.171 17:42:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@732 -- # xtrace_disable 00:26:56.171 17:42:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:26:56.171 17:42:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:26:56.171 17:42:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@209 -- # bdevperf_pid=289357 00:26:56.171 17:42:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@210 -- # waitforlisten 289357 /var/tmp/bdevperf.sock 00:26:56.171 17:42:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 289357 ']' 00:26:56.171 17:42:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:26:56.171 17:42:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:26:56.171 17:42:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@206 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 -c /dev/fd/63 00:26:56.171 17:42:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@206 -- # echo '{ 00:26:56.171 "subsystems": [ 00:26:56.171 { 00:26:56.171 "subsystem": "keyring", 00:26:56.171 "config": [ 00:26:56.171 { 00:26:56.171 "method": "keyring_file_add_key", 00:26:56.171 "params": { 00:26:56.171 "name": "key0", 00:26:56.171 "path": "/tmp/tmp.OtECQfGJ9H" 00:26:56.171 } 00:26:56.171 } 00:26:56.171 ] 00:26:56.171 }, 00:26:56.171 { 00:26:56.171 "subsystem": "iobuf", 00:26:56.171 "config": [ 00:26:56.171 { 00:26:56.171 "method": "iobuf_set_options", 00:26:56.172 "params": { 00:26:56.172 "small_pool_count": 8192, 00:26:56.172 "large_pool_count": 1024, 00:26:56.172 "small_bufsize": 8192, 00:26:56.172 "large_bufsize": 135168, 00:26:56.172 "enable_numa": false 00:26:56.172 } 00:26:56.172 } 00:26:56.172 ] 00:26:56.172 }, 00:26:56.172 { 00:26:56.172 "subsystem": "sock", 00:26:56.172 "config": [ 00:26:56.172 { 00:26:56.172 "method": "sock_set_default_impl", 00:26:56.172 "params": { 00:26:56.172 "impl_name": "posix" 00:26:56.172 } 00:26:56.172 }, 00:26:56.172 { 00:26:56.172 "method": "sock_impl_set_options", 00:26:56.172 "params": { 00:26:56.172 "impl_name": "ssl", 00:26:56.172 "recv_buf_size": 4096, 00:26:56.172 "send_buf_size": 4096, 00:26:56.172 "enable_recv_pipe": true, 00:26:56.172 "enable_quickack": false, 00:26:56.172 "enable_placement_id": 0, 00:26:56.172 "enable_zerocopy_send_server": true, 00:26:56.172 "enable_zerocopy_send_client": false, 00:26:56.172 "zerocopy_threshold": 0, 00:26:56.172 "tls_version": 0, 00:26:56.172 "enable_ktls": false 00:26:56.172 } 00:26:56.172 }, 00:26:56.172 { 00:26:56.172 "method": "sock_impl_set_options", 00:26:56.172 "params": { 00:26:56.172 "impl_name": "posix", 00:26:56.172 "recv_buf_size": 2097152, 00:26:56.172 "send_buf_size": 2097152, 00:26:56.172 "enable_recv_pipe": true, 00:26:56.172 "enable_quickack": false, 00:26:56.172 "enable_placement_id": 0, 00:26:56.172 "enable_zerocopy_send_server": true, 00:26:56.172 "enable_zerocopy_send_client": false, 00:26:56.172 "zerocopy_threshold": 0, 00:26:56.172 "tls_version": 0, 00:26:56.172 "enable_ktls": false 00:26:56.172 } 00:26:56.172 } 00:26:56.172 ] 00:26:56.172 }, 00:26:56.172 { 00:26:56.172 "subsystem": "vmd", 00:26:56.172 "config": [] 00:26:56.172 }, 00:26:56.172 { 00:26:56.172 "subsystem": "accel", 00:26:56.172 "config": [ 00:26:56.172 { 00:26:56.172 "method": "accel_set_options", 00:26:56.172 "params": { 00:26:56.172 "small_cache_size": 128, 00:26:56.172 "large_cache_size": 16, 00:26:56.172 "task_count": 2048, 00:26:56.172 "sequence_count": 2048, 00:26:56.172 "buf_count": 2048 00:26:56.172 } 00:26:56.172 } 00:26:56.172 ] 00:26:56.172 }, 00:26:56.172 { 00:26:56.172 "subsystem": "bdev", 00:26:56.172 "config": [ 00:26:56.172 { 00:26:56.172 "method": "bdev_set_options", 00:26:56.172 "params": { 00:26:56.172 "bdev_io_pool_size": 65535, 00:26:56.172 "bdev_io_cache_size": 256, 00:26:56.172 "bdev_auto_examine": true, 00:26:56.172 "iobuf_small_cache_size": 128, 00:26:56.172 "iobuf_large_cache_size": 16 00:26:56.172 } 00:26:56.172 }, 00:26:56.172 { 00:26:56.172 "method": "bdev_raid_set_options", 00:26:56.172 "params": { 00:26:56.172 "process_window_size_kb": 1024, 00:26:56.172 "process_max_bandwidth_mb_sec": 0 00:26:56.172 } 00:26:56.172 }, 00:26:56.172 { 00:26:56.172 "method": "bdev_iscsi_set_options", 00:26:56.172 "params": { 00:26:56.172 "timeout_sec": 30 00:26:56.172 } 00:26:56.172 }, 00:26:56.172 { 00:26:56.172 "method": "bdev_nvme_set_options", 00:26:56.172 "params": { 00:26:56.172 "action_on_timeout": "none", 00:26:56.172 "timeout_us": 0, 00:26:56.172 "timeout_admin_us": 0, 00:26:56.172 "keep_alive_timeout_ms": 10000, 00:26:56.172 "arbitration_burst": 0, 00:26:56.172 "low_priority_weight": 0, 00:26:56.172 "medium_priority_weight": 0, 00:26:56.172 "high_priority_weight": 0, 00:26:56.172 "nvme_adminq_poll_period_us": 10000, 00:26:56.172 "nvme_ioq_poll_period_us": 0, 00:26:56.172 "io_queue_requests": 512, 00:26:56.172 "delay_cmd_submit": true, 00:26:56.172 "transport_retry_count": 4, 00:26:56.172 "bdev_retry_count": 3, 00:26:56.172 "transport_ack_timeout": 0, 00:26:56.172 "ctrlr_loss_timeout_sec": 0, 00:26:56.172 "reconnect_delay_sec": 0, 00:26:56.172 "fast_io_fail_timeout_sec": 0, 00:26:56.172 "disable_auto_failback": false, 00:26:56.172 "generate_uuids": false, 00:26:56.172 "transport_tos": 0, 00:26:56.172 "nvme_error_stat": false, 00:26:56.172 "rdma_srq_size": 0, 00:26:56.172 "io_path_stat": false, 00:26:56.172 "allow_accel_sequence": false, 00:26:56.172 "rdma_max_cq_size": 0, 00:26:56.172 "rdma_cm_event_timeout_ms": 0, 00:26:56.172 "dhchap_digests": [ 00:26:56.172 "sha256", 00:26:56.172 "sha384", 00:26:56.172 "sha512" 00:26:56.172 ], 00:26:56.172 "dhchap_dhgroups": [ 00:26:56.172 "null", 00:26:56.172 "ffdhe2048", 00:26:56.172 "ffdhe3072", 00:26:56.172 "ffdhe4096", 00:26:56.172 "ffdhe6144", 00:26:56.172 "ffdhe8192" 00:26:56.172 ] 00:26:56.172 } 00:26:56.172 }, 00:26:56.172 { 00:26:56.172 "method": "bdev_nvme_attach_controller", 00:26:56.172 "params": { 00:26:56.172 "name": "TLSTEST", 00:26:56.172 "trtype": "TCP", 00:26:56.172 "adrfam": "IPv4", 00:26:56.172 "traddr": "10.0.0.2", 00:26:56.172 "trsvcid": "4420", 00:26:56.172 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:26:56.172 "prchk_reftag": false, 00:26:56.172 "prchk_guard": false, 00:26:56.172 "ctrlr_loss_timeout_sec": 0, 00:26:56.172 "reconnect_delay_sec": 0, 00:26:56.172 "fast_io_fail_timeout_sec": 0, 00:26:56.172 "psk": "key0", 00:26:56.172 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:26:56.172 "hdgst": false, 00:26:56.172 "ddgst": false, 00:26:56.172 "multipath": "multipath" 00:26:56.173 } 00:26:56.173 }, 00:26:56.173 { 00:26:56.173 "method": "bdev_nvme_set_hotplug", 00:26:56.173 "params": { 00:26:56.173 "period_us": 100000, 00:26:56.173 "enable": false 00:26:56.173 } 00:26:56.173 }, 00:26:56.173 { 00:26:56.173 "method": "bdev_wait_for_examine" 00:26:56.173 } 00:26:56.173 ] 00:26:56.173 }, 00:26:56.173 { 00:26:56.173 "subsystem": "nbd", 00:26:56.173 "config": [] 00:26:56.173 } 00:26:56.173 ] 00:26:56.173 }' 00:26:56.173 17:42:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:26:56.173 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:26:56.173 17:42:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:26:56.173 17:42:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:26:56.173 [2024-12-06 17:42:37.951239] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:26:56.173 [2024-12-06 17:42:37.951332] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid289357 ] 00:26:56.448 [2024-12-06 17:42:38.019124] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:56.448 [2024-12-06 17:42:38.063992] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:26:56.448 [2024-12-06 17:42:38.238789] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:26:56.705 17:42:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:26:56.705 17:42:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:26:56.705 17:42:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@213 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -t 20 -s /var/tmp/bdevperf.sock perform_tests 00:26:56.705 Running I/O for 10 seconds... 00:26:59.056 2977.00 IOPS, 11.63 MiB/s [2024-12-06T16:42:41.825Z] 3022.50 IOPS, 11.81 MiB/s [2024-12-06T16:42:42.757Z] 3015.67 IOPS, 11.78 MiB/s [2024-12-06T16:42:43.692Z] 3028.50 IOPS, 11.83 MiB/s [2024-12-06T16:42:44.623Z] 3036.80 IOPS, 11.86 MiB/s [2024-12-06T16:42:45.554Z] 3020.67 IOPS, 11.80 MiB/s [2024-12-06T16:42:46.924Z] 3022.86 IOPS, 11.81 MiB/s [2024-12-06T16:42:47.855Z] 3030.12 IOPS, 11.84 MiB/s [2024-12-06T16:42:48.787Z] 3031.33 IOPS, 11.84 MiB/s [2024-12-06T16:42:48.787Z] 3024.50 IOPS, 11.81 MiB/s 00:27:06.948 Latency(us) 00:27:06.948 [2024-12-06T16:42:48.787Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:27:06.948 Job: TLSTESTn1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:27:06.948 Verification LBA range: start 0x0 length 0x2000 00:27:06.948 TLSTESTn1 : 10.04 3025.59 11.82 0.00 0.00 42200.18 6990.51 49321.91 00:27:06.948 [2024-12-06T16:42:48.787Z] =================================================================================================================== 00:27:06.948 [2024-12-06T16:42:48.787Z] Total : 3025.59 11.82 0.00 0.00 42200.18 6990.51 49321.91 00:27:06.948 { 00:27:06.948 "results": [ 00:27:06.948 { 00:27:06.948 "job": "TLSTESTn1", 00:27:06.949 "core_mask": "0x4", 00:27:06.949 "workload": "verify", 00:27:06.949 "status": "finished", 00:27:06.949 "verify_range": { 00:27:06.949 "start": 0, 00:27:06.949 "length": 8192 00:27:06.949 }, 00:27:06.949 "queue_depth": 128, 00:27:06.949 "io_size": 4096, 00:27:06.949 "runtime": 10.038372, 00:27:06.949 "iops": 3025.5902052643596, 00:27:06.949 "mibps": 11.818711739313905, 00:27:06.949 "io_failed": 0, 00:27:06.949 "io_timeout": 0, 00:27:06.949 "avg_latency_us": 42200.181473969686, 00:27:06.949 "min_latency_us": 6990.506666666667, 00:27:06.949 "max_latency_us": 49321.90814814815 00:27:06.949 } 00:27:06.949 ], 00:27:06.949 "core_count": 1 00:27:06.949 } 00:27:06.949 17:42:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@215 -- # trap 'nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:27:06.949 17:42:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@216 -- # killprocess 289357 00:27:06.949 17:42:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 289357 ']' 00:27:06.949 17:42:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 289357 00:27:06.949 17:42:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:27:06.949 17:42:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:06.949 17:42:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 289357 00:27:06.949 17:42:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:27:06.949 17:42:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:27:06.949 17:42:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 289357' 00:27:06.949 killing process with pid 289357 00:27:06.949 17:42:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 289357 00:27:06.949 Received shutdown signal, test time was about 10.000000 seconds 00:27:06.949 00:27:06.949 Latency(us) 00:27:06.949 [2024-12-06T16:42:48.788Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:27:06.949 [2024-12-06T16:42:48.788Z] =================================================================================================================== 00:27:06.949 [2024-12-06T16:42:48.788Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:27:06.949 17:42:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 289357 00:27:06.949 17:42:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@217 -- # killprocess 289207 00:27:06.949 17:42:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 289207 ']' 00:27:06.949 17:42:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 289207 00:27:06.949 17:42:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:27:06.949 17:42:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:06.949 17:42:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 289207 00:27:07.208 17:42:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:27:07.208 17:42:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:27:07.208 17:42:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 289207' 00:27:07.208 killing process with pid 289207 00:27:07.208 17:42:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 289207 00:27:07.208 17:42:48 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 289207 00:27:07.208 17:42:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@220 -- # nvmfappstart 00:27:07.208 17:42:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:27:07.208 17:42:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@726 -- # xtrace_disable 00:27:07.208 17:42:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:27:07.208 17:42:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # nvmfpid=290676 00:27:07.208 17:42:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF 00:27:07.208 17:42:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # waitforlisten 290676 00:27:07.208 17:42:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 290676 ']' 00:27:07.208 17:42:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:07.208 17:42:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:27:07.208 17:42:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:07.208 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:07.208 17:42:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:27:07.208 17:42:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:27:07.466 [2024-12-06 17:42:49.087935] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:27:07.466 [2024-12-06 17:42:49.088038] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:27:07.466 [2024-12-06 17:42:49.161936] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:07.466 [2024-12-06 17:42:49.200923] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:27:07.466 [2024-12-06 17:42:49.201001] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:27:07.466 [2024-12-06 17:42:49.201025] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:27:07.466 [2024-12-06 17:42:49.201036] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:27:07.466 [2024-12-06 17:42:49.201045] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:27:07.466 [2024-12-06 17:42:49.201573] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:27:07.724 17:42:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:27:07.724 17:42:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:27:07.724 17:42:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:27:07.724 17:42:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@732 -- # xtrace_disable 00:27:07.724 17:42:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:27:07.724 17:42:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:27:07.724 17:42:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@221 -- # setup_nvmf_tgt /tmp/tmp.OtECQfGJ9H 00:27:07.724 17:42:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@50 -- # local key=/tmp/tmp.OtECQfGJ9H 00:27:07.724 17:42:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:27:07.981 [2024-12-06 17:42:49.622469] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:27:07.981 17:42:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:27:08.239 17:42:49 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:27:08.496 [2024-12-06 17:42:50.163978] tcp.c:1049:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:27:08.496 [2024-12-06 17:42:50.164240] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:27:08.496 17:42:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:27:08.755 malloc0 00:27:08.755 17:42:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:27:09.012 17:42:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py keyring_file_add_key key0 /tmp/tmp.OtECQfGJ9H 00:27:09.270 17:42:50 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk key0 00:27:09.527 17:42:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@224 -- # bdevperf_pid=290960 00:27:09.527 17:42:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@222 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -z -r /var/tmp/bdevperf.sock -q 128 -o 4k -w verify -t 1 00:27:09.527 17:42:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@226 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:27:09.527 17:42:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@227 -- # waitforlisten 290960 /var/tmp/bdevperf.sock 00:27:09.527 17:42:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 290960 ']' 00:27:09.527 17:42:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:27:09.527 17:42:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:27:09.527 17:42:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:27:09.527 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:27:09.527 17:42:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:27:09.527 17:42:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:27:09.527 [2024-12-06 17:42:51.308119] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:27:09.527 [2024-12-06 17:42:51.308208] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid290960 ] 00:27:09.784 [2024-12-06 17:42:51.375921] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:09.784 [2024-12-06 17:42:51.423162] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:27:09.784 17:42:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:27:09.784 17:42:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:27:09.784 17:42:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@229 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.OtECQfGJ9H 00:27:10.042 17:42:51 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@230 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 --psk key0 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 00:27:10.300 [2024-12-06 17:42:52.108525] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:27:10.557 nvme0n1 00:27:10.557 17:42:52 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@234 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:27:10.557 Running I/O for 1 seconds... 00:27:11.490 3268.00 IOPS, 12.77 MiB/s 00:27:11.490 Latency(us) 00:27:11.490 [2024-12-06T16:42:53.329Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:27:11.490 Job: nvme0n1 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:27:11.490 Verification LBA range: start 0x0 length 0x2000 00:27:11.490 nvme0n1 : 1.02 3333.75 13.02 0.00 0.00 38061.66 5971.06 54758.97 00:27:11.490 [2024-12-06T16:42:53.329Z] =================================================================================================================== 00:27:11.490 [2024-12-06T16:42:53.329Z] Total : 3333.75 13.02 0.00 0.00 38061.66 5971.06 54758.97 00:27:11.490 { 00:27:11.490 "results": [ 00:27:11.490 { 00:27:11.490 "job": "nvme0n1", 00:27:11.490 "core_mask": "0x2", 00:27:11.490 "workload": "verify", 00:27:11.490 "status": "finished", 00:27:11.490 "verify_range": { 00:27:11.490 "start": 0, 00:27:11.490 "length": 8192 00:27:11.490 }, 00:27:11.490 "queue_depth": 128, 00:27:11.490 "io_size": 4096, 00:27:11.490 "runtime": 1.018672, 00:27:11.490 "iops": 3333.752179307962, 00:27:11.490 "mibps": 13.022469450421726, 00:27:11.490 "io_failed": 0, 00:27:11.490 "io_timeout": 0, 00:27:11.490 "avg_latency_us": 38061.66232343062, 00:27:11.490 "min_latency_us": 5971.057777777778, 00:27:11.490 "max_latency_us": 54758.96888888889 00:27:11.490 } 00:27:11.490 ], 00:27:11.490 "core_count": 1 00:27:11.490 } 00:27:11.747 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@236 -- # killprocess 290960 00:27:11.747 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 290960 ']' 00:27:11.747 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 290960 00:27:11.747 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:27:11.747 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:11.747 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 290960 00:27:11.747 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:27:11.747 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:27:11.748 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 290960' 00:27:11.748 killing process with pid 290960 00:27:11.748 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 290960 00:27:11.748 Received shutdown signal, test time was about 1.000000 seconds 00:27:11.748 00:27:11.748 Latency(us) 00:27:11.748 [2024-12-06T16:42:53.587Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:27:11.748 [2024-12-06T16:42:53.587Z] =================================================================================================================== 00:27:11.748 [2024-12-06T16:42:53.587Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:27:11.748 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 290960 00:27:11.748 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@237 -- # killprocess 290676 00:27:11.748 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 290676 ']' 00:27:11.748 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 290676 00:27:11.748 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:27:11.748 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:11.748 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 290676 00:27:11.748 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:27:11.748 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:27:11.748 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 290676' 00:27:11.748 killing process with pid 290676 00:27:11.748 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 290676 00:27:11.748 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 290676 00:27:12.005 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@242 -- # nvmfappstart 00:27:12.005 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:27:12.005 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@726 -- # xtrace_disable 00:27:12.005 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:27:12.005 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # nvmfpid=291245 00:27:12.005 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF 00:27:12.005 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # waitforlisten 291245 00:27:12.005 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 291245 ']' 00:27:12.005 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:12.005 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:27:12.005 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:12.005 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:12.005 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:27:12.005 17:42:53 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:27:12.005 [2024-12-06 17:42:53.818719] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:27:12.005 [2024-12-06 17:42:53.818809] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:27:12.263 [2024-12-06 17:42:53.890359] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:12.263 [2024-12-06 17:42:53.937839] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:27:12.263 [2024-12-06 17:42:53.937903] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:27:12.263 [2024-12-06 17:42:53.937917] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:27:12.263 [2024-12-06 17:42:53.937928] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:27:12.263 [2024-12-06 17:42:53.937938] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:27:12.263 [2024-12-06 17:42:53.938527] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:27:12.263 17:42:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:27:12.263 17:42:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:27:12.263 17:42:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:27:12.263 17:42:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@732 -- # xtrace_disable 00:27:12.263 17:42:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:27:12.263 17:42:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:27:12.263 17:42:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@243 -- # rpc_cmd 00:27:12.263 17:42:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:12.263 17:42:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:27:12.263 [2024-12-06 17:42:54.084970] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:27:12.520 malloc0 00:27:12.520 [2024-12-06 17:42:54.116547] tcp.c:1049:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:27:12.520 [2024-12-06 17:42:54.116831] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:27:12.520 17:42:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:12.520 17:42:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@256 -- # bdevperf_pid=291362 00:27:12.520 17:42:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@258 -- # waitforlisten 291362 /var/tmp/bdevperf.sock 00:27:12.520 17:42:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@254 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -z -r /var/tmp/bdevperf.sock -q 128 -o 4k -w verify -t 1 00:27:12.520 17:42:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 291362 ']' 00:27:12.520 17:42:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:27:12.520 17:42:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:27:12.520 17:42:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:27:12.520 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:27:12.520 17:42:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:27:12.520 17:42:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:27:12.520 [2024-12-06 17:42:54.193991] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:27:12.520 [2024-12-06 17:42:54.194095] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid291362 ] 00:27:12.520 [2024-12-06 17:42:54.260360] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:12.520 [2024-12-06 17:42:54.305187] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:27:12.777 17:42:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:27:12.777 17:42:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:27:12.777 17:42:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@259 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.OtECQfGJ9H 00:27:13.034 17:42:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@260 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 --psk key0 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 00:27:13.291 [2024-12-06 17:42:54.930306] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:27:13.291 nvme0n1 00:27:13.291 17:42:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@264 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:27:13.291 Running I/O for 1 seconds... 00:27:14.658 3271.00 IOPS, 12.78 MiB/s 00:27:14.658 Latency(us) 00:27:14.658 [2024-12-06T16:42:56.497Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:27:14.658 Job: nvme0n1 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:27:14.658 Verification LBA range: start 0x0 length 0x2000 00:27:14.658 nvme0n1 : 1.03 3315.58 12.95 0.00 0.00 38207.55 8786.68 38059.43 00:27:14.658 [2024-12-06T16:42:56.497Z] =================================================================================================================== 00:27:14.658 [2024-12-06T16:42:56.497Z] Total : 3315.58 12.95 0.00 0.00 38207.55 8786.68 38059.43 00:27:14.658 { 00:27:14.658 "results": [ 00:27:14.658 { 00:27:14.658 "job": "nvme0n1", 00:27:14.658 "core_mask": "0x2", 00:27:14.658 "workload": "verify", 00:27:14.658 "status": "finished", 00:27:14.658 "verify_range": { 00:27:14.658 "start": 0, 00:27:14.658 "length": 8192 00:27:14.658 }, 00:27:14.658 "queue_depth": 128, 00:27:14.658 "io_size": 4096, 00:27:14.658 "runtime": 1.025161, 00:27:14.658 "iops": 3315.5767728190986, 00:27:14.658 "mibps": 12.951471768824604, 00:27:14.658 "io_failed": 0, 00:27:14.658 "io_timeout": 0, 00:27:14.658 "avg_latency_us": 38207.55298399311, 00:27:14.658 "min_latency_us": 8786.678518518518, 00:27:14.658 "max_latency_us": 38059.42518518519 00:27:14.658 } 00:27:14.658 ], 00:27:14.658 "core_count": 1 00:27:14.658 } 00:27:14.658 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@267 -- # rpc_cmd save_config 00:27:14.658 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:14.658 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:27:14.658 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:14.658 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@267 -- # tgtcfg='{ 00:27:14.658 "subsystems": [ 00:27:14.658 { 00:27:14.658 "subsystem": "keyring", 00:27:14.658 "config": [ 00:27:14.658 { 00:27:14.658 "method": "keyring_file_add_key", 00:27:14.658 "params": { 00:27:14.658 "name": "key0", 00:27:14.658 "path": "/tmp/tmp.OtECQfGJ9H" 00:27:14.658 } 00:27:14.658 } 00:27:14.658 ] 00:27:14.658 }, 00:27:14.658 { 00:27:14.658 "subsystem": "iobuf", 00:27:14.658 "config": [ 00:27:14.658 { 00:27:14.658 "method": "iobuf_set_options", 00:27:14.658 "params": { 00:27:14.658 "small_pool_count": 8192, 00:27:14.658 "large_pool_count": 1024, 00:27:14.658 "small_bufsize": 8192, 00:27:14.658 "large_bufsize": 135168, 00:27:14.658 "enable_numa": false 00:27:14.658 } 00:27:14.658 } 00:27:14.658 ] 00:27:14.658 }, 00:27:14.658 { 00:27:14.658 "subsystem": "sock", 00:27:14.658 "config": [ 00:27:14.658 { 00:27:14.658 "method": "sock_set_default_impl", 00:27:14.658 "params": { 00:27:14.658 "impl_name": "posix" 00:27:14.658 } 00:27:14.658 }, 00:27:14.658 { 00:27:14.658 "method": "sock_impl_set_options", 00:27:14.658 "params": { 00:27:14.658 "impl_name": "ssl", 00:27:14.658 "recv_buf_size": 4096, 00:27:14.658 "send_buf_size": 4096, 00:27:14.658 "enable_recv_pipe": true, 00:27:14.658 "enable_quickack": false, 00:27:14.658 "enable_placement_id": 0, 00:27:14.658 "enable_zerocopy_send_server": true, 00:27:14.658 "enable_zerocopy_send_client": false, 00:27:14.658 "zerocopy_threshold": 0, 00:27:14.658 "tls_version": 0, 00:27:14.658 "enable_ktls": false 00:27:14.658 } 00:27:14.658 }, 00:27:14.658 { 00:27:14.658 "method": "sock_impl_set_options", 00:27:14.658 "params": { 00:27:14.658 "impl_name": "posix", 00:27:14.658 "recv_buf_size": 2097152, 00:27:14.658 "send_buf_size": 2097152, 00:27:14.658 "enable_recv_pipe": true, 00:27:14.658 "enable_quickack": false, 00:27:14.658 "enable_placement_id": 0, 00:27:14.658 "enable_zerocopy_send_server": true, 00:27:14.658 "enable_zerocopy_send_client": false, 00:27:14.658 "zerocopy_threshold": 0, 00:27:14.658 "tls_version": 0, 00:27:14.658 "enable_ktls": false 00:27:14.658 } 00:27:14.658 } 00:27:14.658 ] 00:27:14.658 }, 00:27:14.658 { 00:27:14.658 "subsystem": "vmd", 00:27:14.658 "config": [] 00:27:14.658 }, 00:27:14.658 { 00:27:14.658 "subsystem": "accel", 00:27:14.658 "config": [ 00:27:14.658 { 00:27:14.658 "method": "accel_set_options", 00:27:14.658 "params": { 00:27:14.658 "small_cache_size": 128, 00:27:14.658 "large_cache_size": 16, 00:27:14.658 "task_count": 2048, 00:27:14.658 "sequence_count": 2048, 00:27:14.659 "buf_count": 2048 00:27:14.659 } 00:27:14.659 } 00:27:14.659 ] 00:27:14.659 }, 00:27:14.659 { 00:27:14.659 "subsystem": "bdev", 00:27:14.659 "config": [ 00:27:14.659 { 00:27:14.659 "method": "bdev_set_options", 00:27:14.659 "params": { 00:27:14.659 "bdev_io_pool_size": 65535, 00:27:14.659 "bdev_io_cache_size": 256, 00:27:14.659 "bdev_auto_examine": true, 00:27:14.659 "iobuf_small_cache_size": 128, 00:27:14.659 "iobuf_large_cache_size": 16 00:27:14.659 } 00:27:14.659 }, 00:27:14.659 { 00:27:14.659 "method": "bdev_raid_set_options", 00:27:14.659 "params": { 00:27:14.659 "process_window_size_kb": 1024, 00:27:14.659 "process_max_bandwidth_mb_sec": 0 00:27:14.659 } 00:27:14.659 }, 00:27:14.659 { 00:27:14.659 "method": "bdev_iscsi_set_options", 00:27:14.659 "params": { 00:27:14.659 "timeout_sec": 30 00:27:14.659 } 00:27:14.659 }, 00:27:14.659 { 00:27:14.659 "method": "bdev_nvme_set_options", 00:27:14.659 "params": { 00:27:14.659 "action_on_timeout": "none", 00:27:14.659 "timeout_us": 0, 00:27:14.659 "timeout_admin_us": 0, 00:27:14.659 "keep_alive_timeout_ms": 10000, 00:27:14.659 "arbitration_burst": 0, 00:27:14.659 "low_priority_weight": 0, 00:27:14.659 "medium_priority_weight": 0, 00:27:14.659 "high_priority_weight": 0, 00:27:14.659 "nvme_adminq_poll_period_us": 10000, 00:27:14.659 "nvme_ioq_poll_period_us": 0, 00:27:14.659 "io_queue_requests": 0, 00:27:14.659 "delay_cmd_submit": true, 00:27:14.659 "transport_retry_count": 4, 00:27:14.659 "bdev_retry_count": 3, 00:27:14.659 "transport_ack_timeout": 0, 00:27:14.659 "ctrlr_loss_timeout_sec": 0, 00:27:14.659 "reconnect_delay_sec": 0, 00:27:14.659 "fast_io_fail_timeout_sec": 0, 00:27:14.659 "disable_auto_failback": false, 00:27:14.659 "generate_uuids": false, 00:27:14.659 "transport_tos": 0, 00:27:14.659 "nvme_error_stat": false, 00:27:14.659 "rdma_srq_size": 0, 00:27:14.659 "io_path_stat": false, 00:27:14.659 "allow_accel_sequence": false, 00:27:14.659 "rdma_max_cq_size": 0, 00:27:14.659 "rdma_cm_event_timeout_ms": 0, 00:27:14.659 "dhchap_digests": [ 00:27:14.659 "sha256", 00:27:14.659 "sha384", 00:27:14.659 "sha512" 00:27:14.659 ], 00:27:14.659 "dhchap_dhgroups": [ 00:27:14.659 "null", 00:27:14.659 "ffdhe2048", 00:27:14.659 "ffdhe3072", 00:27:14.659 "ffdhe4096", 00:27:14.659 "ffdhe6144", 00:27:14.659 "ffdhe8192" 00:27:14.659 ] 00:27:14.659 } 00:27:14.659 }, 00:27:14.659 { 00:27:14.659 "method": "bdev_nvme_set_hotplug", 00:27:14.659 "params": { 00:27:14.659 "period_us": 100000, 00:27:14.659 "enable": false 00:27:14.659 } 00:27:14.659 }, 00:27:14.659 { 00:27:14.659 "method": "bdev_malloc_create", 00:27:14.659 "params": { 00:27:14.659 "name": "malloc0", 00:27:14.659 "num_blocks": 8192, 00:27:14.659 "block_size": 4096, 00:27:14.659 "physical_block_size": 4096, 00:27:14.659 "uuid": "ed4c67e7-ef13-41e2-bd68-aac10de75c2a", 00:27:14.659 "optimal_io_boundary": 0, 00:27:14.659 "md_size": 0, 00:27:14.659 "dif_type": 0, 00:27:14.659 "dif_is_head_of_md": false, 00:27:14.659 "dif_pi_format": 0 00:27:14.659 } 00:27:14.659 }, 00:27:14.659 { 00:27:14.659 "method": "bdev_wait_for_examine" 00:27:14.659 } 00:27:14.659 ] 00:27:14.659 }, 00:27:14.659 { 00:27:14.659 "subsystem": "nbd", 00:27:14.659 "config": [] 00:27:14.659 }, 00:27:14.659 { 00:27:14.659 "subsystem": "scheduler", 00:27:14.659 "config": [ 00:27:14.659 { 00:27:14.659 "method": "framework_set_scheduler", 00:27:14.659 "params": { 00:27:14.659 "name": "static" 00:27:14.659 } 00:27:14.659 } 00:27:14.659 ] 00:27:14.659 }, 00:27:14.659 { 00:27:14.659 "subsystem": "nvmf", 00:27:14.659 "config": [ 00:27:14.659 { 00:27:14.659 "method": "nvmf_set_config", 00:27:14.659 "params": { 00:27:14.659 "discovery_filter": "match_any", 00:27:14.659 "admin_cmd_passthru": { 00:27:14.659 "identify_ctrlr": false 00:27:14.659 }, 00:27:14.659 "dhchap_digests": [ 00:27:14.659 "sha256", 00:27:14.659 "sha384", 00:27:14.659 "sha512" 00:27:14.659 ], 00:27:14.659 "dhchap_dhgroups": [ 00:27:14.659 "null", 00:27:14.659 "ffdhe2048", 00:27:14.659 "ffdhe3072", 00:27:14.659 "ffdhe4096", 00:27:14.659 "ffdhe6144", 00:27:14.659 "ffdhe8192" 00:27:14.659 ] 00:27:14.659 } 00:27:14.659 }, 00:27:14.659 { 00:27:14.659 "method": "nvmf_set_max_subsystems", 00:27:14.659 "params": { 00:27:14.659 "max_subsystems": 1024 00:27:14.659 } 00:27:14.659 }, 00:27:14.659 { 00:27:14.659 "method": "nvmf_set_crdt", 00:27:14.659 "params": { 00:27:14.659 "crdt1": 0, 00:27:14.659 "crdt2": 0, 00:27:14.659 "crdt3": 0 00:27:14.659 } 00:27:14.659 }, 00:27:14.659 { 00:27:14.659 "method": "nvmf_create_transport", 00:27:14.659 "params": { 00:27:14.659 "trtype": "TCP", 00:27:14.659 "max_queue_depth": 128, 00:27:14.659 "max_io_qpairs_per_ctrlr": 127, 00:27:14.659 "in_capsule_data_size": 4096, 00:27:14.659 "max_io_size": 131072, 00:27:14.659 "io_unit_size": 131072, 00:27:14.659 "max_aq_depth": 128, 00:27:14.659 "num_shared_buffers": 511, 00:27:14.659 "buf_cache_size": 4294967295, 00:27:14.659 "dif_insert_or_strip": false, 00:27:14.659 "zcopy": false, 00:27:14.659 "c2h_success": false, 00:27:14.659 "sock_priority": 0, 00:27:14.659 "abort_timeout_sec": 1, 00:27:14.659 "ack_timeout": 0, 00:27:14.659 "data_wr_pool_size": 0 00:27:14.659 } 00:27:14.659 }, 00:27:14.659 { 00:27:14.659 "method": "nvmf_create_subsystem", 00:27:14.659 "params": { 00:27:14.659 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:27:14.659 "allow_any_host": false, 00:27:14.659 "serial_number": "00000000000000000000", 00:27:14.659 "model_number": "SPDK bdev Controller", 00:27:14.659 "max_namespaces": 32, 00:27:14.659 "min_cntlid": 1, 00:27:14.659 "max_cntlid": 65519, 00:27:14.659 "ana_reporting": false 00:27:14.659 } 00:27:14.659 }, 00:27:14.659 { 00:27:14.659 "method": "nvmf_subsystem_add_host", 00:27:14.659 "params": { 00:27:14.659 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:27:14.659 "host": "nqn.2016-06.io.spdk:host1", 00:27:14.659 "psk": "key0" 00:27:14.659 } 00:27:14.659 }, 00:27:14.659 { 00:27:14.659 "method": "nvmf_subsystem_add_ns", 00:27:14.659 "params": { 00:27:14.659 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:27:14.659 "namespace": { 00:27:14.659 "nsid": 1, 00:27:14.659 "bdev_name": "malloc0", 00:27:14.659 "nguid": "ED4C67E7EF1341E2BD68AAC10DE75C2A", 00:27:14.659 "uuid": "ed4c67e7-ef13-41e2-bd68-aac10de75c2a", 00:27:14.659 "no_auto_visible": false 00:27:14.659 } 00:27:14.659 } 00:27:14.659 }, 00:27:14.659 { 00:27:14.659 "method": "nvmf_subsystem_add_listener", 00:27:14.659 "params": { 00:27:14.659 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:27:14.659 "listen_address": { 00:27:14.659 "trtype": "TCP", 00:27:14.659 "adrfam": "IPv4", 00:27:14.659 "traddr": "10.0.0.2", 00:27:14.659 "trsvcid": "4420" 00:27:14.659 }, 00:27:14.659 "secure_channel": false, 00:27:14.659 "sock_impl": "ssl" 00:27:14.659 } 00:27:14.659 } 00:27:14.659 ] 00:27:14.659 } 00:27:14.659 ] 00:27:14.659 }' 00:27:14.659 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@268 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock save_config 00:27:14.917 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@268 -- # bperfcfg='{ 00:27:14.917 "subsystems": [ 00:27:14.917 { 00:27:14.917 "subsystem": "keyring", 00:27:14.917 "config": [ 00:27:14.917 { 00:27:14.917 "method": "keyring_file_add_key", 00:27:14.917 "params": { 00:27:14.917 "name": "key0", 00:27:14.917 "path": "/tmp/tmp.OtECQfGJ9H" 00:27:14.917 } 00:27:14.917 } 00:27:14.917 ] 00:27:14.917 }, 00:27:14.917 { 00:27:14.917 "subsystem": "iobuf", 00:27:14.917 "config": [ 00:27:14.917 { 00:27:14.917 "method": "iobuf_set_options", 00:27:14.917 "params": { 00:27:14.917 "small_pool_count": 8192, 00:27:14.917 "large_pool_count": 1024, 00:27:14.917 "small_bufsize": 8192, 00:27:14.917 "large_bufsize": 135168, 00:27:14.917 "enable_numa": false 00:27:14.917 } 00:27:14.917 } 00:27:14.917 ] 00:27:14.917 }, 00:27:14.917 { 00:27:14.917 "subsystem": "sock", 00:27:14.917 "config": [ 00:27:14.917 { 00:27:14.917 "method": "sock_set_default_impl", 00:27:14.917 "params": { 00:27:14.917 "impl_name": "posix" 00:27:14.917 } 00:27:14.917 }, 00:27:14.917 { 00:27:14.917 "method": "sock_impl_set_options", 00:27:14.917 "params": { 00:27:14.917 "impl_name": "ssl", 00:27:14.917 "recv_buf_size": 4096, 00:27:14.917 "send_buf_size": 4096, 00:27:14.917 "enable_recv_pipe": true, 00:27:14.917 "enable_quickack": false, 00:27:14.917 "enable_placement_id": 0, 00:27:14.917 "enable_zerocopy_send_server": true, 00:27:14.917 "enable_zerocopy_send_client": false, 00:27:14.917 "zerocopy_threshold": 0, 00:27:14.917 "tls_version": 0, 00:27:14.917 "enable_ktls": false 00:27:14.917 } 00:27:14.917 }, 00:27:14.917 { 00:27:14.917 "method": "sock_impl_set_options", 00:27:14.917 "params": { 00:27:14.917 "impl_name": "posix", 00:27:14.917 "recv_buf_size": 2097152, 00:27:14.917 "send_buf_size": 2097152, 00:27:14.917 "enable_recv_pipe": true, 00:27:14.917 "enable_quickack": false, 00:27:14.917 "enable_placement_id": 0, 00:27:14.917 "enable_zerocopy_send_server": true, 00:27:14.917 "enable_zerocopy_send_client": false, 00:27:14.917 "zerocopy_threshold": 0, 00:27:14.917 "tls_version": 0, 00:27:14.917 "enable_ktls": false 00:27:14.917 } 00:27:14.917 } 00:27:14.917 ] 00:27:14.917 }, 00:27:14.917 { 00:27:14.917 "subsystem": "vmd", 00:27:14.917 "config": [] 00:27:14.917 }, 00:27:14.917 { 00:27:14.917 "subsystem": "accel", 00:27:14.917 "config": [ 00:27:14.917 { 00:27:14.917 "method": "accel_set_options", 00:27:14.917 "params": { 00:27:14.917 "small_cache_size": 128, 00:27:14.917 "large_cache_size": 16, 00:27:14.917 "task_count": 2048, 00:27:14.917 "sequence_count": 2048, 00:27:14.917 "buf_count": 2048 00:27:14.917 } 00:27:14.917 } 00:27:14.917 ] 00:27:14.917 }, 00:27:14.917 { 00:27:14.917 "subsystem": "bdev", 00:27:14.917 "config": [ 00:27:14.917 { 00:27:14.917 "method": "bdev_set_options", 00:27:14.917 "params": { 00:27:14.917 "bdev_io_pool_size": 65535, 00:27:14.917 "bdev_io_cache_size": 256, 00:27:14.917 "bdev_auto_examine": true, 00:27:14.917 "iobuf_small_cache_size": 128, 00:27:14.917 "iobuf_large_cache_size": 16 00:27:14.917 } 00:27:14.917 }, 00:27:14.917 { 00:27:14.917 "method": "bdev_raid_set_options", 00:27:14.917 "params": { 00:27:14.917 "process_window_size_kb": 1024, 00:27:14.917 "process_max_bandwidth_mb_sec": 0 00:27:14.917 } 00:27:14.917 }, 00:27:14.917 { 00:27:14.917 "method": "bdev_iscsi_set_options", 00:27:14.917 "params": { 00:27:14.917 "timeout_sec": 30 00:27:14.917 } 00:27:14.917 }, 00:27:14.917 { 00:27:14.917 "method": "bdev_nvme_set_options", 00:27:14.917 "params": { 00:27:14.917 "action_on_timeout": "none", 00:27:14.917 "timeout_us": 0, 00:27:14.917 "timeout_admin_us": 0, 00:27:14.917 "keep_alive_timeout_ms": 10000, 00:27:14.917 "arbitration_burst": 0, 00:27:14.917 "low_priority_weight": 0, 00:27:14.917 "medium_priority_weight": 0, 00:27:14.917 "high_priority_weight": 0, 00:27:14.917 "nvme_adminq_poll_period_us": 10000, 00:27:14.917 "nvme_ioq_poll_period_us": 0, 00:27:14.917 "io_queue_requests": 512, 00:27:14.917 "delay_cmd_submit": true, 00:27:14.917 "transport_retry_count": 4, 00:27:14.917 "bdev_retry_count": 3, 00:27:14.917 "transport_ack_timeout": 0, 00:27:14.917 "ctrlr_loss_timeout_sec": 0, 00:27:14.917 "reconnect_delay_sec": 0, 00:27:14.917 "fast_io_fail_timeout_sec": 0, 00:27:14.917 "disable_auto_failback": false, 00:27:14.917 "generate_uuids": false, 00:27:14.917 "transport_tos": 0, 00:27:14.917 "nvme_error_stat": false, 00:27:14.917 "rdma_srq_size": 0, 00:27:14.917 "io_path_stat": false, 00:27:14.917 "allow_accel_sequence": false, 00:27:14.917 "rdma_max_cq_size": 0, 00:27:14.917 "rdma_cm_event_timeout_ms": 0, 00:27:14.917 "dhchap_digests": [ 00:27:14.917 "sha256", 00:27:14.917 "sha384", 00:27:14.917 "sha512" 00:27:14.917 ], 00:27:14.917 "dhchap_dhgroups": [ 00:27:14.917 "null", 00:27:14.917 "ffdhe2048", 00:27:14.917 "ffdhe3072", 00:27:14.917 "ffdhe4096", 00:27:14.917 "ffdhe6144", 00:27:14.917 "ffdhe8192" 00:27:14.917 ] 00:27:14.917 } 00:27:14.917 }, 00:27:14.917 { 00:27:14.917 "method": "bdev_nvme_attach_controller", 00:27:14.917 "params": { 00:27:14.917 "name": "nvme0", 00:27:14.917 "trtype": "TCP", 00:27:14.917 "adrfam": "IPv4", 00:27:14.917 "traddr": "10.0.0.2", 00:27:14.917 "trsvcid": "4420", 00:27:14.917 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:27:14.917 "prchk_reftag": false, 00:27:14.917 "prchk_guard": false, 00:27:14.917 "ctrlr_loss_timeout_sec": 0, 00:27:14.917 "reconnect_delay_sec": 0, 00:27:14.917 "fast_io_fail_timeout_sec": 0, 00:27:14.917 "psk": "key0", 00:27:14.917 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:27:14.917 "hdgst": false, 00:27:14.917 "ddgst": false, 00:27:14.917 "multipath": "multipath" 00:27:14.917 } 00:27:14.917 }, 00:27:14.917 { 00:27:14.917 "method": "bdev_nvme_set_hotplug", 00:27:14.917 "params": { 00:27:14.917 "period_us": 100000, 00:27:14.917 "enable": false 00:27:14.917 } 00:27:14.917 }, 00:27:14.917 { 00:27:14.917 "method": "bdev_enable_histogram", 00:27:14.917 "params": { 00:27:14.917 "name": "nvme0n1", 00:27:14.917 "enable": true 00:27:14.917 } 00:27:14.917 }, 00:27:14.917 { 00:27:14.917 "method": "bdev_wait_for_examine" 00:27:14.917 } 00:27:14.917 ] 00:27:14.917 }, 00:27:14.917 { 00:27:14.917 "subsystem": "nbd", 00:27:14.917 "config": [] 00:27:14.917 } 00:27:14.917 ] 00:27:14.917 }' 00:27:14.917 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@270 -- # killprocess 291362 00:27:14.917 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 291362 ']' 00:27:14.917 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 291362 00:27:14.917 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:27:14.917 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:14.917 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 291362 00:27:14.917 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:27:14.917 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:27:14.917 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 291362' 00:27:14.917 killing process with pid 291362 00:27:14.918 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 291362 00:27:14.918 Received shutdown signal, test time was about 1.000000 seconds 00:27:14.918 00:27:14.918 Latency(us) 00:27:14.918 [2024-12-06T16:42:56.757Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:27:14.918 [2024-12-06T16:42:56.757Z] =================================================================================================================== 00:27:14.918 [2024-12-06T16:42:56.757Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:27:14.918 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 291362 00:27:15.175 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@271 -- # killprocess 291245 00:27:15.175 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 291245 ']' 00:27:15.175 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 291245 00:27:15.175 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:27:15.175 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:15.175 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 291245 00:27:15.175 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:27:15.175 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:27:15.175 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 291245' 00:27:15.175 killing process with pid 291245 00:27:15.175 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 291245 00:27:15.175 17:42:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 291245 00:27:15.432 17:42:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@273 -- # nvmfappstart -c /dev/fd/62 00:27:15.432 17:42:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:27:15.432 17:42:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@273 -- # echo '{ 00:27:15.432 "subsystems": [ 00:27:15.432 { 00:27:15.432 "subsystem": "keyring", 00:27:15.432 "config": [ 00:27:15.432 { 00:27:15.432 "method": "keyring_file_add_key", 00:27:15.432 "params": { 00:27:15.432 "name": "key0", 00:27:15.432 "path": "/tmp/tmp.OtECQfGJ9H" 00:27:15.432 } 00:27:15.432 } 00:27:15.432 ] 00:27:15.432 }, 00:27:15.432 { 00:27:15.432 "subsystem": "iobuf", 00:27:15.432 "config": [ 00:27:15.433 { 00:27:15.433 "method": "iobuf_set_options", 00:27:15.433 "params": { 00:27:15.433 "small_pool_count": 8192, 00:27:15.433 "large_pool_count": 1024, 00:27:15.433 "small_bufsize": 8192, 00:27:15.433 "large_bufsize": 135168, 00:27:15.433 "enable_numa": false 00:27:15.433 } 00:27:15.433 } 00:27:15.433 ] 00:27:15.433 }, 00:27:15.433 { 00:27:15.433 "subsystem": "sock", 00:27:15.433 "config": [ 00:27:15.433 { 00:27:15.433 "method": "sock_set_default_impl", 00:27:15.433 "params": { 00:27:15.433 "impl_name": "posix" 00:27:15.433 } 00:27:15.433 }, 00:27:15.433 { 00:27:15.433 "method": "sock_impl_set_options", 00:27:15.433 "params": { 00:27:15.433 "impl_name": "ssl", 00:27:15.433 "recv_buf_size": 4096, 00:27:15.433 "send_buf_size": 4096, 00:27:15.433 "enable_recv_pipe": true, 00:27:15.433 "enable_quickack": false, 00:27:15.433 "enable_placement_id": 0, 00:27:15.433 "enable_zerocopy_send_server": true, 00:27:15.433 "enable_zerocopy_send_client": false, 00:27:15.433 "zerocopy_threshold": 0, 00:27:15.433 "tls_version": 0, 00:27:15.433 "enable_ktls": false 00:27:15.433 } 00:27:15.433 }, 00:27:15.433 { 00:27:15.433 "method": "sock_impl_set_options", 00:27:15.433 "params": { 00:27:15.433 "impl_name": "posix", 00:27:15.433 "recv_buf_size": 2097152, 00:27:15.433 "send_buf_size": 2097152, 00:27:15.433 "enable_recv_pipe": true, 00:27:15.433 "enable_quickack": false, 00:27:15.433 "enable_placement_id": 0, 00:27:15.433 "enable_zerocopy_send_server": true, 00:27:15.433 "enable_zerocopy_send_client": false, 00:27:15.433 "zerocopy_threshold": 0, 00:27:15.433 "tls_version": 0, 00:27:15.433 "enable_ktls": false 00:27:15.433 } 00:27:15.433 } 00:27:15.433 ] 00:27:15.433 }, 00:27:15.433 { 00:27:15.433 "subsystem": "vmd", 00:27:15.433 "config": [] 00:27:15.433 }, 00:27:15.433 { 00:27:15.433 "subsystem": "accel", 00:27:15.433 "config": [ 00:27:15.433 { 00:27:15.433 "method": "accel_set_options", 00:27:15.433 "params": { 00:27:15.433 "small_cache_size": 128, 00:27:15.433 "large_cache_size": 16, 00:27:15.433 "task_count": 2048, 00:27:15.433 "sequence_count": 2048, 00:27:15.433 "buf_count": 2048 00:27:15.433 } 00:27:15.433 } 00:27:15.433 ] 00:27:15.433 }, 00:27:15.433 { 00:27:15.433 "subsystem": "bdev", 00:27:15.433 "config": [ 00:27:15.433 { 00:27:15.433 "method": "bdev_set_options", 00:27:15.433 "params": { 00:27:15.433 "bdev_io_pool_size": 65535, 00:27:15.433 "bdev_io_cache_size": 256, 00:27:15.433 "bdev_auto_examine": true, 00:27:15.433 "iobuf_small_cache_size": 128, 00:27:15.433 "iobuf_large_cache_size": 16 00:27:15.433 } 00:27:15.433 }, 00:27:15.433 { 00:27:15.433 "method": "bdev_raid_set_options", 00:27:15.433 "params": { 00:27:15.433 "process_window_size_kb": 1024, 00:27:15.433 "process_max_bandwidth_mb_sec": 0 00:27:15.433 } 00:27:15.433 }, 00:27:15.433 { 00:27:15.433 "method": "bdev_iscsi_set_options", 00:27:15.433 "params": { 00:27:15.433 "timeout_sec": 30 00:27:15.433 } 00:27:15.433 }, 00:27:15.433 { 00:27:15.433 "method": "bdev_nvme_set_options", 00:27:15.433 "params": { 00:27:15.433 "action_on_timeout": "none", 00:27:15.433 "timeout_us": 0, 00:27:15.433 "timeout_admin_us": 0, 00:27:15.433 "keep_alive_timeout_ms": 10000, 00:27:15.433 "arbitration_burst": 0, 00:27:15.433 "low_priority_weight": 0, 00:27:15.433 "medium_priority_weight": 0, 00:27:15.433 "high_priority_weight": 0, 00:27:15.433 "nvme_adminq_poll_period_us": 10000, 00:27:15.433 "nvme_ioq_poll_period_us": 0, 00:27:15.433 "io_queue_requests": 0, 00:27:15.433 "delay_cmd_submit": true, 00:27:15.433 "transport_retry_count": 4, 00:27:15.433 "bdev_retry_count": 3, 00:27:15.433 "transport_ack_timeout": 0, 00:27:15.433 "ctrlr_loss_timeout_sec": 0, 00:27:15.433 "reconnect_delay_sec": 0, 00:27:15.433 "fast_io_fail_timeout_sec": 0, 00:27:15.433 "disable_auto_failback": false, 00:27:15.433 "generate_uuids": false, 00:27:15.433 "transport_tos": 0, 00:27:15.433 "nvme_error_stat": false, 00:27:15.433 "rdma_srq_size": 0, 00:27:15.433 "io_path_stat": false, 00:27:15.433 "allow_accel_sequence": false, 00:27:15.433 "rdma_max_cq_size": 0, 00:27:15.433 "rdma_cm_event_timeout_ms": 0, 00:27:15.433 "dhchap_digests": [ 00:27:15.433 "sha256", 00:27:15.433 "sha384", 00:27:15.433 "sha512" 00:27:15.433 ], 00:27:15.433 "dhchap_dhgroups": [ 00:27:15.433 "null", 00:27:15.433 "ffdhe2048", 00:27:15.433 "ffdhe3072", 00:27:15.433 "ffdhe4096", 00:27:15.433 "ffdhe6144", 00:27:15.433 "ffdhe8192" 00:27:15.433 ] 00:27:15.433 } 00:27:15.433 }, 00:27:15.433 { 00:27:15.433 "method": "bdev_nvme_set_hotplug", 00:27:15.433 "params": { 00:27:15.433 "period_us": 100000, 00:27:15.433 "enable": false 00:27:15.433 } 00:27:15.433 }, 00:27:15.433 { 00:27:15.433 "method": "bdev_malloc_create", 00:27:15.433 "params": { 00:27:15.433 "name": "malloc0", 00:27:15.433 "num_blocks": 8192, 00:27:15.433 "block_size": 4096, 00:27:15.433 "physical_block_size": 4096, 00:27:15.433 "uuid": "ed4c67e7-ef13-41e2-bd68-aac10de75c2a", 00:27:15.433 "optimal_io_boundary": 0, 00:27:15.433 "md_size": 0, 00:27:15.433 "dif_type": 0, 00:27:15.433 "dif_is_head_of_md": false, 00:27:15.434 "dif_pi_format": 0 00:27:15.434 } 00:27:15.434 }, 00:27:15.434 { 00:27:15.434 "method": "bdev_wait_for_examine" 00:27:15.434 } 00:27:15.434 ] 00:27:15.434 }, 00:27:15.434 { 00:27:15.434 "subsystem": "nbd", 00:27:15.434 "config": [] 00:27:15.434 }, 00:27:15.434 { 00:27:15.434 "subsystem": "scheduler", 00:27:15.434 "config": [ 00:27:15.434 { 00:27:15.434 "method": "framework_set_scheduler", 00:27:15.434 "params": { 00:27:15.434 "name": "static" 00:27:15.434 } 00:27:15.434 } 00:27:15.434 ] 00:27:15.434 }, 00:27:15.434 { 00:27:15.434 "subsystem": "nvmf", 00:27:15.434 "config": [ 00:27:15.434 { 00:27:15.434 "method": "nvmf_set_config", 00:27:15.434 "params": { 00:27:15.434 "discovery_filter": "match_any", 00:27:15.434 "admin_cmd_passthru": { 00:27:15.434 "identify_ctrlr": false 00:27:15.434 }, 00:27:15.434 "dhchap_digests": [ 00:27:15.434 "sha256", 00:27:15.434 "sha384", 00:27:15.434 "sha512" 00:27:15.434 ], 00:27:15.434 "dhchap_dhgroups": [ 00:27:15.434 "null", 00:27:15.434 "ffdhe2048", 00:27:15.434 "ffdhe3072", 00:27:15.434 "ffdhe4096", 00:27:15.434 "ffdhe6144", 00:27:15.434 "ffdhe8192" 00:27:15.434 ] 00:27:15.434 } 00:27:15.434 }, 00:27:15.434 { 00:27:15.434 "method": "nvmf_set_max_subsystems", 00:27:15.434 "params": { 00:27:15.434 "max_subsystems": 1024 00:27:15.434 } 00:27:15.434 }, 00:27:15.434 { 00:27:15.434 "method": "nvmf_set_crdt", 00:27:15.434 "params": { 00:27:15.434 "crdt1": 0, 00:27:15.434 "crdt2": 0, 00:27:15.434 "crdt3": 0 00:27:15.434 } 00:27:15.434 }, 00:27:15.434 { 00:27:15.434 "method": "nvmf_create_transport", 00:27:15.434 "params": { 00:27:15.434 "trtype": "TCP", 00:27:15.434 "max_queue_depth": 128, 00:27:15.434 "max_io_qpairs_per_ctrlr": 127, 00:27:15.434 "in_capsule_data_size": 4096, 00:27:15.434 "max_io_size": 131072, 00:27:15.434 "io_unit_size": 131072, 00:27:15.434 "max_aq_depth": 128, 00:27:15.434 "num_shared_buffers": 511, 00:27:15.434 "buf_cache_size": 4294967295, 00:27:15.434 "dif_insert_or_strip": false, 00:27:15.434 "zcopy": false, 00:27:15.434 "c2h_success": false, 00:27:15.434 "sock_priority": 0, 00:27:15.434 "abort_timeout_sec": 1, 00:27:15.434 "ack_timeout": 0, 00:27:15.434 "data_wr_pool_size": 0 00:27:15.434 } 00:27:15.434 }, 00:27:15.434 { 00:27:15.434 "method": "nvmf_create_subsystem", 00:27:15.434 "params": { 00:27:15.434 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:27:15.434 "allow_any_host": false, 00:27:15.434 "serial_number": "00000000000000000000", 00:27:15.434 "model_number": "SPDK bdev Controller", 00:27:15.434 "max_namespaces": 32, 00:27:15.434 "min_cntlid": 1, 00:27:15.434 "max_cntlid": 65519, 00:27:15.434 "ana_reporting": false 00:27:15.434 } 00:27:15.434 }, 00:27:15.434 { 00:27:15.434 "method": "nvmf_subsystem_add_host", 00:27:15.434 "params": { 00:27:15.434 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:27:15.434 "host": "nqn.2016-06.io.spdk:host1", 00:27:15.434 "psk": "key0" 00:27:15.434 } 00:27:15.434 }, 00:27:15.434 { 00:27:15.434 "method": "nvmf_subsystem_add_ns", 00:27:15.434 "params": { 00:27:15.434 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:27:15.434 "namespace": { 00:27:15.434 "nsid": 1, 00:27:15.434 "bdev_name": "malloc0", 00:27:15.434 "nguid": "ED4C67E7EF1341E2BD68AAC10DE75C2A", 00:27:15.434 "uuid": "ed4c67e7-ef13-41e2-bd68-aac10de75c2a", 00:27:15.434 "no_auto_visible": false 00:27:15.434 } 00:27:15.434 } 00:27:15.434 }, 00:27:15.434 { 00:27:15.434 "method": "nvmf_subsystem_add_listener", 00:27:15.434 "params": { 00:27:15.434 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:27:15.434 "listen_address": { 00:27:15.434 "trtype": "TCP", 00:27:15.434 "adrfam": "IPv4", 00:27:15.434 "traddr": "10.0.0.2", 00:27:15.434 "trsvcid": "4420" 00:27:15.434 }, 00:27:15.434 "secure_channel": false, 00:27:15.434 "sock_impl": "ssl" 00:27:15.434 } 00:27:15.434 } 00:27:15.434 ] 00:27:15.434 } 00:27:15.434 ] 00:27:15.434 }' 00:27:15.434 17:42:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@726 -- # xtrace_disable 00:27:15.434 17:42:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:27:15.434 17:42:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # nvmfpid=291677 00:27:15.434 17:42:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -c /dev/fd/62 00:27:15.434 17:42:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # waitforlisten 291677 00:27:15.434 17:42:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 291677 ']' 00:27:15.434 17:42:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:15.434 17:42:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:27:15.434 17:42:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:15.434 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:15.434 17:42:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:27:15.434 17:42:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:27:15.434 [2024-12-06 17:42:57.178571] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:27:15.435 [2024-12-06 17:42:57.178686] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:27:15.435 [2024-12-06 17:42:57.249643] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:15.692 [2024-12-06 17:42:57.294224] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:27:15.692 [2024-12-06 17:42:57.294285] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:27:15.692 [2024-12-06 17:42:57.294306] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:27:15.692 [2024-12-06 17:42:57.294317] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:27:15.692 [2024-12-06 17:42:57.294327] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:27:15.692 [2024-12-06 17:42:57.294928] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:27:15.949 [2024-12-06 17:42:57.534696] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:27:15.949 [2024-12-06 17:42:57.566719] tcp.c:1049:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:27:15.949 [2024-12-06 17:42:57.566947] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:27:16.513 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:27:16.513 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:27:16.513 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:27:16.513 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@732 -- # xtrace_disable 00:27:16.513 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:27:16.513 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:27:16.513 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@276 -- # bdevperf_pid=291827 00:27:16.513 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@277 -- # waitforlisten 291827 /var/tmp/bdevperf.sock 00:27:16.513 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # '[' -z 291827 ']' 00:27:16.513 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:27:16.513 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@274 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -z -r /var/tmp/bdevperf.sock -q 128 -o 4k -w verify -t 1 -c /dev/fd/63 00:27:16.513 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # local max_retries=100 00:27:16.513 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:27:16.513 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:27:16.513 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@274 -- # echo '{ 00:27:16.513 "subsystems": [ 00:27:16.513 { 00:27:16.513 "subsystem": "keyring", 00:27:16.513 "config": [ 00:27:16.513 { 00:27:16.513 "method": "keyring_file_add_key", 00:27:16.513 "params": { 00:27:16.513 "name": "key0", 00:27:16.513 "path": "/tmp/tmp.OtECQfGJ9H" 00:27:16.513 } 00:27:16.513 } 00:27:16.513 ] 00:27:16.513 }, 00:27:16.513 { 00:27:16.513 "subsystem": "iobuf", 00:27:16.513 "config": [ 00:27:16.513 { 00:27:16.513 "method": "iobuf_set_options", 00:27:16.513 "params": { 00:27:16.513 "small_pool_count": 8192, 00:27:16.513 "large_pool_count": 1024, 00:27:16.513 "small_bufsize": 8192, 00:27:16.513 "large_bufsize": 135168, 00:27:16.513 "enable_numa": false 00:27:16.513 } 00:27:16.513 } 00:27:16.513 ] 00:27:16.513 }, 00:27:16.513 { 00:27:16.513 "subsystem": "sock", 00:27:16.513 "config": [ 00:27:16.513 { 00:27:16.513 "method": "sock_set_default_impl", 00:27:16.513 "params": { 00:27:16.513 "impl_name": "posix" 00:27:16.513 } 00:27:16.513 }, 00:27:16.513 { 00:27:16.513 "method": "sock_impl_set_options", 00:27:16.513 "params": { 00:27:16.513 "impl_name": "ssl", 00:27:16.513 "recv_buf_size": 4096, 00:27:16.513 "send_buf_size": 4096, 00:27:16.513 "enable_recv_pipe": true, 00:27:16.513 "enable_quickack": false, 00:27:16.513 "enable_placement_id": 0, 00:27:16.513 "enable_zerocopy_send_server": true, 00:27:16.513 "enable_zerocopy_send_client": false, 00:27:16.513 "zerocopy_threshold": 0, 00:27:16.513 "tls_version": 0, 00:27:16.513 "enable_ktls": false 00:27:16.513 } 00:27:16.513 }, 00:27:16.513 { 00:27:16.513 "method": "sock_impl_set_options", 00:27:16.513 "params": { 00:27:16.513 "impl_name": "posix", 00:27:16.513 "recv_buf_size": 2097152, 00:27:16.513 "send_buf_size": 2097152, 00:27:16.513 "enable_recv_pipe": true, 00:27:16.513 "enable_quickack": false, 00:27:16.513 "enable_placement_id": 0, 00:27:16.513 "enable_zerocopy_send_server": true, 00:27:16.513 "enable_zerocopy_send_client": false, 00:27:16.513 "zerocopy_threshold": 0, 00:27:16.513 "tls_version": 0, 00:27:16.513 "enable_ktls": false 00:27:16.513 } 00:27:16.513 } 00:27:16.513 ] 00:27:16.513 }, 00:27:16.513 { 00:27:16.513 "subsystem": "vmd", 00:27:16.513 "config": [] 00:27:16.513 }, 00:27:16.513 { 00:27:16.513 "subsystem": "accel", 00:27:16.513 "config": [ 00:27:16.513 { 00:27:16.513 "method": "accel_set_options", 00:27:16.513 "params": { 00:27:16.513 "small_cache_size": 128, 00:27:16.513 "large_cache_size": 16, 00:27:16.513 "task_count": 2048, 00:27:16.513 "sequence_count": 2048, 00:27:16.513 "buf_count": 2048 00:27:16.513 } 00:27:16.513 } 00:27:16.513 ] 00:27:16.513 }, 00:27:16.513 { 00:27:16.513 "subsystem": "bdev", 00:27:16.513 "config": [ 00:27:16.513 { 00:27:16.513 "method": "bdev_set_options", 00:27:16.513 "params": { 00:27:16.513 "bdev_io_pool_size": 65535, 00:27:16.513 "bdev_io_cache_size": 256, 00:27:16.513 "bdev_auto_examine": true, 00:27:16.513 "iobuf_small_cache_size": 128, 00:27:16.513 "iobuf_large_cache_size": 16 00:27:16.513 } 00:27:16.513 }, 00:27:16.513 { 00:27:16.513 "method": "bdev_raid_set_options", 00:27:16.513 "params": { 00:27:16.513 "process_window_size_kb": 1024, 00:27:16.513 "process_max_bandwidth_mb_sec": 0 00:27:16.513 } 00:27:16.513 }, 00:27:16.513 { 00:27:16.513 "method": "bdev_iscsi_set_options", 00:27:16.513 "params": { 00:27:16.513 "timeout_sec": 30 00:27:16.513 } 00:27:16.513 }, 00:27:16.513 { 00:27:16.513 "method": "bdev_nvme_set_options", 00:27:16.513 "params": { 00:27:16.513 "action_on_timeout": "none", 00:27:16.513 "timeout_us": 0, 00:27:16.513 "timeout_admin_us": 0, 00:27:16.513 "keep_alive_timeout_ms": 10000, 00:27:16.513 "arbitration_burst": 0, 00:27:16.513 "low_priority_weight": 0, 00:27:16.513 "medium_priority_weight": 0, 00:27:16.513 "high_priority_weight": 0, 00:27:16.513 "nvme_adminq_poll_period_us": 10000, 00:27:16.513 "nvme_ioq_poll_period_us": 0, 00:27:16.513 "io_queue_requests": 512, 00:27:16.513 "delay_cmd_submit": true, 00:27:16.513 "transport_retry_count": 4, 00:27:16.513 "bdev_retry_count": 3, 00:27:16.513 "transport_ack_timeout": 0, 00:27:16.513 "ctrlr_loss_timeout_sec": 0, 00:27:16.513 "reconnect_delay_sec": 0, 00:27:16.513 "fast_io_fail_timeout_sec": 0, 00:27:16.513 "disable_auto_failback": false, 00:27:16.513 "generate_uuids": false, 00:27:16.513 "transport_tos": 0, 00:27:16.513 "nvme_error_stat": false, 00:27:16.513 "rdma_srq_size": 0, 00:27:16.513 "io_path_stat": false, 00:27:16.513 "allow_accel_sequence": false, 00:27:16.513 "rdma_max_cq_size": 0, 00:27:16.514 "rdma_cm_event_timeout_ms": 0, 00:27:16.514 "dhchap_digests": [ 00:27:16.514 "sha256", 00:27:16.514 "sha384", 00:27:16.514 "sha512" 00:27:16.514 ], 00:27:16.514 "dhchap_dhgroups": [ 00:27:16.514 "null", 00:27:16.514 "ffdhe2048", 00:27:16.514 "ffdhe3072", 00:27:16.514 "ffdhe4096", 00:27:16.514 "ffdhe6144", 00:27:16.514 "ffdhe8192" 00:27:16.514 ] 00:27:16.514 } 00:27:16.514 }, 00:27:16.514 { 00:27:16.514 "method": "bdev_nvme_attach_controller", 00:27:16.514 "params": { 00:27:16.514 "name": "nvme0", 00:27:16.514 "trtype": "TCP", 00:27:16.514 "adrfam": "IPv4", 00:27:16.514 "traddr": "10.0.0.2", 00:27:16.514 "trsvcid": "4420", 00:27:16.514 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:27:16.514 "prchk_reftag": false, 00:27:16.514 "prchk_guard": false, 00:27:16.514 "ctrlr_loss_timeout_sec": 0, 00:27:16.514 "reconnect_delay_sec": 0, 00:27:16.514 "fast_io_fail_timeout_sec": 0, 00:27:16.514 "psk": "key0", 00:27:16.514 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:27:16.514 "hdgst": false, 00:27:16.514 "ddgst": false, 00:27:16.514 "multipath": "multipath" 00:27:16.514 } 00:27:16.514 }, 00:27:16.514 { 00:27:16.514 "method": "bdev_nvme_set_hotplug", 00:27:16.514 "params": { 00:27:16.514 "period_us": 100000, 00:27:16.514 "enable": false 00:27:16.514 } 00:27:16.514 }, 00:27:16.514 { 00:27:16.514 "method": "bdev_enable_histogram", 00:27:16.514 "params": { 00:27:16.514 "name": "nvme0n1", 00:27:16.514 "enable": true 00:27:16.514 } 00:27:16.514 }, 00:27:16.514 { 00:27:16.514 "method": "bdev_wait_for_examine" 00:27:16.514 } 00:27:16.514 ] 00:27:16.514 }, 00:27:16.514 { 00:27:16.514 "subsystem": "nbd", 00:27:16.514 "config": [] 00:27:16.514 } 00:27:16.514 ] 00:27:16.514 }' 00:27:16.514 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@844 -- # xtrace_disable 00:27:16.514 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:27:16.514 [2024-12-06 17:42:58.286307] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:27:16.514 [2024-12-06 17:42:58.286395] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid291827 ] 00:27:16.771 [2024-12-06 17:42:58.352738] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:16.771 [2024-12-06 17:42:58.399348] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:27:16.771 [2024-12-06 17:42:58.580772] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:27:17.028 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:27:17.028 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@868 -- # return 0 00:27:17.028 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@279 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:27:17.028 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@279 -- # jq -r '.[].name' 00:27:17.286 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@279 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:27:17.286 17:42:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@280 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:27:17.286 Running I/O for 1 seconds... 00:27:18.656 3073.00 IOPS, 12.00 MiB/s 00:27:18.656 Latency(us) 00:27:18.656 [2024-12-06T16:43:00.495Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:27:18.656 Job: nvme0n1 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:27:18.656 Verification LBA range: start 0x0 length 0x2000 00:27:18.656 nvme0n1 : 1.02 3127.93 12.22 0.00 0.00 40512.01 8398.32 56700.78 00:27:18.656 [2024-12-06T16:43:00.495Z] =================================================================================================================== 00:27:18.656 [2024-12-06T16:43:00.495Z] Total : 3127.93 12.22 0.00 0.00 40512.01 8398.32 56700.78 00:27:18.656 { 00:27:18.656 "results": [ 00:27:18.656 { 00:27:18.656 "job": "nvme0n1", 00:27:18.656 "core_mask": "0x2", 00:27:18.656 "workload": "verify", 00:27:18.656 "status": "finished", 00:27:18.656 "verify_range": { 00:27:18.656 "start": 0, 00:27:18.656 "length": 8192 00:27:18.656 }, 00:27:18.656 "queue_depth": 128, 00:27:18.656 "io_size": 4096, 00:27:18.656 "runtime": 1.023361, 00:27:18.656 "iops": 3127.9284631718424, 00:27:18.656 "mibps": 12.21847055926501, 00:27:18.656 "io_failed": 0, 00:27:18.656 "io_timeout": 0, 00:27:18.656 "avg_latency_us": 40512.011048399225, 00:27:18.656 "min_latency_us": 8398.317037037037, 00:27:18.656 "max_latency_us": 56700.776296296295 00:27:18.656 } 00:27:18.656 ], 00:27:18.656 "core_count": 1 00:27:18.656 } 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@282 -- # trap - SIGINT SIGTERM EXIT 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@283 -- # cleanup 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@15 -- # process_shm --id 0 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@812 -- # type=--id 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@813 -- # id=0 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@814 -- # '[' --id = --pid ']' 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@818 -- # find /dev/shm -name '*.0' -printf '%f\n' 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@818 -- # shm_files=nvmf_trace.0 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@820 -- # [[ -z nvmf_trace.0 ]] 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@824 -- # for n in $shm_files 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@825 -- # tar -C /dev/shm/ -cvzf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf_trace.0_shm.tar.gz nvmf_trace.0 00:27:18.656 nvmf_trace.0 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@827 -- # return 0 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@16 -- # killprocess 291827 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 291827 ']' 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 291827 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 291827 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 291827' 00:27:18.656 killing process with pid 291827 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 291827 00:27:18.656 Received shutdown signal, test time was about 1.000000 seconds 00:27:18.656 00:27:18.656 Latency(us) 00:27:18.656 [2024-12-06T16:43:00.495Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:27:18.656 [2024-12-06T16:43:00.495Z] =================================================================================================================== 00:27:18.656 [2024-12-06T16:43:00.495Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 291827 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@17 -- # nvmftestfini 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@516 -- # nvmfcleanup 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@121 -- # sync 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@124 -- # set +e 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@125 -- # for i in {1..20} 00:27:18.656 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:27:18.656 rmmod nvme_tcp 00:27:18.656 rmmod nvme_fabrics 00:27:18.656 rmmod nvme_keyring 00:27:18.914 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:27:18.914 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@128 -- # set -e 00:27:18.914 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@129 -- # return 0 00:27:18.914 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@517 -- # '[' -n 291677 ']' 00:27:18.914 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@518 -- # killprocess 291677 00:27:18.914 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # '[' -z 291677 ']' 00:27:18.914 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@958 -- # kill -0 291677 00:27:18.914 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # uname 00:27:18.914 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:18.914 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 291677 00:27:18.914 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:27:18.914 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:27:18.914 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@972 -- # echo 'killing process with pid 291677' 00:27:18.914 killing process with pid 291677 00:27:18.914 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@973 -- # kill 291677 00:27:18.914 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@978 -- # wait 291677 00:27:19.171 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:27:19.171 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:27:19.171 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:27:19.171 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@297 -- # iptr 00:27:19.171 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@791 -- # iptables-save 00:27:19.171 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:27:19.171 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@791 -- # iptables-restore 00:27:19.171 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:27:19.171 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@302 -- # remove_spdk_ns 00:27:19.171 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:19.171 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:27:19.171 17:43:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:27:21.076 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:27:21.076 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@18 -- # rm -f /tmp/tmp.7KTTXF2gPL /tmp/tmp.WvD1hwJPAj /tmp/tmp.OtECQfGJ9H 00:27:21.076 00:27:21.076 real 1m22.395s 00:27:21.076 user 2m16.236s 00:27:21.076 sys 0m25.410s 00:27:21.076 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1130 -- # xtrace_disable 00:27:21.076 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:27:21.076 ************************************ 00:27:21.076 END TEST nvmf_tls 00:27:21.076 ************************************ 00:27:21.076 17:43:02 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@42 -- # run_test nvmf_fips /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips/fips.sh --transport=tcp 00:27:21.076 17:43:02 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:27:21.076 17:43:02 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:27:21.076 17:43:02 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:27:21.076 ************************************ 00:27:21.076 START TEST nvmf_fips 00:27:21.076 ************************************ 00:27:21.076 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips/fips.sh --transport=tcp 00:27:21.076 * Looking for test storage... 00:27:21.076 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips 00:27:21.076 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:27:21.076 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1711 -- # lcov --version 00:27:21.076 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@333 -- # local ver1 ver1_l 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@334 -- # local ver2 ver2_l 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@336 -- # IFS=.-: 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@336 -- # read -ra ver1 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@337 -- # IFS=.-: 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@337 -- # read -ra ver2 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@338 -- # local 'op=<' 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@340 -- # ver1_l=2 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@341 -- # ver2_l=1 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@344 -- # case "$op" in 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@345 -- # : 1 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@364 -- # (( v = 0 )) 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@365 -- # decimal 1 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@353 -- # local d=1 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@355 -- # echo 1 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@365 -- # ver1[v]=1 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@366 -- # decimal 2 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@353 -- # local d=2 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@355 -- # echo 2 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@366 -- # ver2[v]=2 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@368 -- # return 0 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:27:21.336 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:21.336 --rc genhtml_branch_coverage=1 00:27:21.336 --rc genhtml_function_coverage=1 00:27:21.336 --rc genhtml_legend=1 00:27:21.336 --rc geninfo_all_blocks=1 00:27:21.336 --rc geninfo_unexecuted_blocks=1 00:27:21.336 00:27:21.336 ' 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:27:21.336 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:21.336 --rc genhtml_branch_coverage=1 00:27:21.336 --rc genhtml_function_coverage=1 00:27:21.336 --rc genhtml_legend=1 00:27:21.336 --rc geninfo_all_blocks=1 00:27:21.336 --rc geninfo_unexecuted_blocks=1 00:27:21.336 00:27:21.336 ' 00:27:21.336 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:27:21.336 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:21.336 --rc genhtml_branch_coverage=1 00:27:21.336 --rc genhtml_function_coverage=1 00:27:21.336 --rc genhtml_legend=1 00:27:21.336 --rc geninfo_all_blocks=1 00:27:21.337 --rc geninfo_unexecuted_blocks=1 00:27:21.337 00:27:21.337 ' 00:27:21.337 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:27:21.337 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:21.337 --rc genhtml_branch_coverage=1 00:27:21.337 --rc genhtml_function_coverage=1 00:27:21.337 --rc genhtml_legend=1 00:27:21.337 --rc geninfo_all_blocks=1 00:27:21.337 --rc geninfo_unexecuted_blocks=1 00:27:21.337 00:27:21.337 ' 00:27:21.337 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@11 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:27:21.337 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@7 -- # uname -s 00:27:21.337 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:27:21.337 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:27:21.337 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:27:21.337 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:27:21.337 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:27:21.337 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:27:21.337 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:27:21.337 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:27:21.337 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:27:21.337 17:43:02 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@15 -- # shopt -s extglob 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- paths/export.sh@5 -- # export PATH 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@51 -- # : 0 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:27:21.337 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@55 -- # have_pci_nics=0 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@12 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@90 -- # check_openssl_version 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@84 -- # local target=3.0.0 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@86 -- # openssl version 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@86 -- # awk '{print $2}' 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@86 -- # ge 3.1.1 3.0.0 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@376 -- # cmp_versions 3.1.1 '>=' 3.0.0 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@333 -- # local ver1 ver1_l 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@334 -- # local ver2 ver2_l 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@336 -- # IFS=.-: 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@336 -- # read -ra ver1 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@337 -- # IFS=.-: 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@337 -- # read -ra ver2 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@338 -- # local 'op=>=' 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@340 -- # ver1_l=3 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@341 -- # ver2_l=3 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@344 -- # case "$op" in 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@348 -- # : 1 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@364 -- # (( v = 0 )) 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@365 -- # decimal 3 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@353 -- # local d=3 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@354 -- # [[ 3 =~ ^[0-9]+$ ]] 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@355 -- # echo 3 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@365 -- # ver1[v]=3 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@366 -- # decimal 3 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@353 -- # local d=3 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@354 -- # [[ 3 =~ ^[0-9]+$ ]] 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@355 -- # echo 3 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@366 -- # ver2[v]=3 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@364 -- # (( v++ )) 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@365 -- # decimal 1 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@353 -- # local d=1 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@355 -- # echo 1 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@365 -- # ver1[v]=1 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@366 -- # decimal 0 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@353 -- # local d=0 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@354 -- # [[ 0 =~ ^[0-9]+$ ]] 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@355 -- # echo 0 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@366 -- # ver2[v]=0 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@367 -- # return 0 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@96 -- # openssl info -modulesdir 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@96 -- # [[ ! -f /usr/lib64/ossl-modules/fips.so ]] 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@101 -- # openssl fipsinstall -help 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@101 -- # warn='This command is not enabled in the Red Hat Enterprise Linux OpenSSL build, please consult Red Hat documentation to learn how to enable FIPS mode' 00:27:21.337 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@102 -- # [[ This command is not enabled in the Red Hat Enterprise Linux OpenSSL build, please consult Red Hat documentation to learn how to enable FIPS mode == \T\h\i\s\ \c\o\m\m\a\n\d\ \i\s\ \n\o\t\ \e\n\a\b\l\e\d* ]] 00:27:21.338 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@105 -- # export callback=build_openssl_config 00:27:21.338 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@105 -- # callback=build_openssl_config 00:27:21.338 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@114 -- # build_openssl_config 00:27:21.338 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@38 -- # cat 00:27:21.338 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@58 -- # [[ ! -t 0 ]] 00:27:21.338 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@59 -- # cat - 00:27:21.338 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@115 -- # export OPENSSL_CONF=spdk_fips.conf 00:27:21.338 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@115 -- # OPENSSL_CONF=spdk_fips.conf 00:27:21.338 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@117 -- # mapfile -t providers 00:27:21.338 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@117 -- # openssl list -providers 00:27:21.338 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@117 -- # grep name 00:27:21.338 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@121 -- # (( 2 != 2 )) 00:27:21.338 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@121 -- # [[ name: openssl base provider != *base* ]] 00:27:21.338 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@121 -- # [[ name: red hat enterprise linux 9 - openssl fips provider != *fips* ]] 00:27:21.338 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@128 -- # NOT openssl md5 /dev/fd/62 00:27:21.338 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@128 -- # : 00:27:21.338 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@652 -- # local es=0 00:27:21.338 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@654 -- # valid_exec_arg openssl md5 /dev/fd/62 00:27:21.338 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@640 -- # local arg=openssl 00:27:21.338 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:27:21.338 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@644 -- # type -t openssl 00:27:21.338 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:27:21.338 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@646 -- # type -P openssl 00:27:21.338 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:27:21.338 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@646 -- # arg=/usr/bin/openssl 00:27:21.338 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@646 -- # [[ -x /usr/bin/openssl ]] 00:27:21.338 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@655 -- # openssl md5 /dev/fd/62 00:27:21.597 Error setting digest 00:27:21.597 40E2D66B4D7F0000:error:0308010C:digital envelope routines:inner_evp_generic_fetch:unsupported:crypto/evp/evp_fetch.c:341:Global default library context, Algorithm (MD5 : 95), Properties () 00:27:21.597 40E2D66B4D7F0000:error:03000086:digital envelope routines:evp_md_init_internal:initialization error:crypto/evp/digest.c:272: 00:27:21.597 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@655 -- # es=1 00:27:21.597 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:27:21.597 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:27:21.597 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:27:21.597 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@131 -- # nvmftestinit 00:27:21.597 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:27:21.597 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:27:21.597 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@476 -- # prepare_net_devs 00:27:21.597 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@438 -- # local -g is_hw=no 00:27:21.597 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@440 -- # remove_spdk_ns 00:27:21.597 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:21.597 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:27:21.597 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:27:21.597 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:27:21.597 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:27:21.597 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@309 -- # xtrace_disable 00:27:21.597 17:43:03 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@315 -- # pci_devs=() 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@315 -- # local -a pci_devs 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@316 -- # pci_net_devs=() 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@317 -- # pci_drivers=() 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@317 -- # local -A pci_drivers 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@319 -- # net_devs=() 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@319 -- # local -ga net_devs 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@320 -- # e810=() 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@320 -- # local -ga e810 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@321 -- # x722=() 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@321 -- # local -ga x722 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@322 -- # mlx=() 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@322 -- # local -ga mlx 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:27:23.499 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:27:23.499 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:27:23.499 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@418 -- # [[ up == up ]] 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:27:23.500 Found net devices under 0000:0a:00.0: cvl_0_0 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@418 -- # [[ up == up ]] 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:27:23.500 Found net devices under 0000:0a:00.1: cvl_0_1 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@442 -- # is_hw=yes 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:27:23.500 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:27:23.759 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:27:23.759 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.343 ms 00:27:23.759 00:27:23.759 --- 10.0.0.2 ping statistics --- 00:27:23.759 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:27:23.759 rtt min/avg/max/mdev = 0.343/0.343/0.343/0.000 ms 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:27:23.759 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:27:23.759 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.160 ms 00:27:23.759 00:27:23.759 --- 10.0.0.1 ping statistics --- 00:27:23.759 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:27:23.759 rtt min/avg/max/mdev = 0.160/0.160/0.160/0.000 ms 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@450 -- # return 0 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@132 -- # nvmfappstart -m 0x2 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@726 -- # xtrace_disable 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@509 -- # nvmfpid=294071 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@510 -- # waitforlisten 294071 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@835 -- # '[' -z 294071 ']' 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@840 -- # local max_retries=100 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:23.759 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@844 -- # xtrace_disable 00:27:23.759 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:27:23.759 [2024-12-06 17:43:05.521311] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:27:23.759 [2024-12-06 17:43:05.521394] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:27:23.759 [2024-12-06 17:43:05.593095] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:24.016 [2024-12-06 17:43:05.641486] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:27:24.016 [2024-12-06 17:43:05.641553] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:27:24.016 [2024-12-06 17:43:05.641577] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:27:24.016 [2024-12-06 17:43:05.641589] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:27:24.016 [2024-12-06 17:43:05.641598] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:27:24.016 [2024-12-06 17:43:05.642204] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:27:24.017 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:27:24.017 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@868 -- # return 0 00:27:24.017 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:27:24.017 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@732 -- # xtrace_disable 00:27:24.017 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:27:24.017 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:27:24.017 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@134 -- # trap cleanup EXIT 00:27:24.017 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@137 -- # key=NVMeTLSkey-1:01:VRLbtnN9AQb2WXW3c9+wEf/DRLz0QuLdbYvEhwtdWwNf9LrZ: 00:27:24.017 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@138 -- # mktemp -t spdk-psk.XXX 00:27:24.017 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@138 -- # key_path=/tmp/spdk-psk.o1F 00:27:24.017 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@139 -- # echo -n NVMeTLSkey-1:01:VRLbtnN9AQb2WXW3c9+wEf/DRLz0QuLdbYvEhwtdWwNf9LrZ: 00:27:24.017 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@140 -- # chmod 0600 /tmp/spdk-psk.o1F 00:27:24.017 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@142 -- # setup_nvmf_tgt_conf /tmp/spdk-psk.o1F 00:27:24.017 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@22 -- # local key=/tmp/spdk-psk.o1F 00:27:24.017 17:43:05 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:27:24.274 [2024-12-06 17:43:06.048245] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:27:24.274 [2024-12-06 17:43:06.064241] tcp.c:1049:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:27:24.274 [2024-12-06 17:43:06.064458] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:27:24.274 malloc0 00:27:24.532 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@145 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:27:24.532 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@148 -- # bdevperf_pid=294217 00:27:24.532 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@146 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:27:24.532 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@149 -- # waitforlisten 294217 /var/tmp/bdevperf.sock 00:27:24.532 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@835 -- # '[' -z 294217 ']' 00:27:24.532 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:27:24.532 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@840 -- # local max_retries=100 00:27:24.532 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:27:24.532 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:27:24.532 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@844 -- # xtrace_disable 00:27:24.532 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:27:24.532 [2024-12-06 17:43:06.190759] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:27:24.532 [2024-12-06 17:43:06.190853] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid294217 ] 00:27:24.532 [2024-12-06 17:43:06.258844] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:24.532 [2024-12-06 17:43:06.304911] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:27:24.791 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:27:24.791 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@868 -- # return 0 00:27:24.791 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@151 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/spdk-psk.o1F 00:27:25.048 17:43:06 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@152 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk key0 00:27:25.305 [2024-12-06 17:43:06.961916] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:27:25.305 TLSTESTn1 00:27:25.305 17:43:07 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@156 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:27:25.563 Running I/O for 10 seconds... 00:27:27.425 3382.00 IOPS, 13.21 MiB/s [2024-12-06T16:43:10.192Z] 3446.50 IOPS, 13.46 MiB/s [2024-12-06T16:43:11.559Z] 3411.33 IOPS, 13.33 MiB/s [2024-12-06T16:43:12.489Z] 3404.00 IOPS, 13.30 MiB/s [2024-12-06T16:43:13.421Z] 3314.60 IOPS, 12.95 MiB/s [2024-12-06T16:43:14.352Z] 3339.17 IOPS, 13.04 MiB/s [2024-12-06T16:43:15.283Z] 3364.86 IOPS, 13.14 MiB/s [2024-12-06T16:43:16.213Z] 3370.12 IOPS, 13.16 MiB/s [2024-12-06T16:43:17.584Z] 3367.67 IOPS, 13.15 MiB/s [2024-12-06T16:43:17.584Z] 3385.90 IOPS, 13.23 MiB/s 00:27:35.745 Latency(us) 00:27:35.745 [2024-12-06T16:43:17.584Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:27:35.745 Job: TLSTESTn1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:27:35.745 Verification LBA range: start 0x0 length 0x2000 00:27:35.745 TLSTESTn1 : 10.02 3391.99 13.25 0.00 0.00 37673.07 7961.41 54370.61 00:27:35.745 [2024-12-06T16:43:17.584Z] =================================================================================================================== 00:27:35.745 [2024-12-06T16:43:17.584Z] Total : 3391.99 13.25 0.00 0.00 37673.07 7961.41 54370.61 00:27:35.745 { 00:27:35.745 "results": [ 00:27:35.745 { 00:27:35.745 "job": "TLSTESTn1", 00:27:35.745 "core_mask": "0x4", 00:27:35.745 "workload": "verify", 00:27:35.745 "status": "finished", 00:27:35.746 "verify_range": { 00:27:35.746 "start": 0, 00:27:35.746 "length": 8192 00:27:35.746 }, 00:27:35.746 "queue_depth": 128, 00:27:35.746 "io_size": 4096, 00:27:35.746 "runtime": 10.019485, 00:27:35.746 "iops": 3391.9907061091462, 00:27:35.746 "mibps": 13.249963695738852, 00:27:35.746 "io_failed": 0, 00:27:35.746 "io_timeout": 0, 00:27:35.746 "avg_latency_us": 37673.07201684354, 00:27:35.746 "min_latency_us": 7961.41037037037, 00:27:35.746 "max_latency_us": 54370.607407407406 00:27:35.746 } 00:27:35.746 ], 00:27:35.746 "core_count": 1 00:27:35.746 } 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@1 -- # cleanup 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@15 -- # process_shm --id 0 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@812 -- # type=--id 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@813 -- # id=0 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@814 -- # '[' --id = --pid ']' 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@818 -- # find /dev/shm -name '*.0' -printf '%f\n' 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@818 -- # shm_files=nvmf_trace.0 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@820 -- # [[ -z nvmf_trace.0 ]] 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@824 -- # for n in $shm_files 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@825 -- # tar -C /dev/shm/ -cvzf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf_trace.0_shm.tar.gz nvmf_trace.0 00:27:35.746 nvmf_trace.0 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@827 -- # return 0 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@16 -- # killprocess 294217 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@954 -- # '[' -z 294217 ']' 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@958 -- # kill -0 294217 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@959 -- # uname 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 294217 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@972 -- # echo 'killing process with pid 294217' 00:27:35.746 killing process with pid 294217 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@973 -- # kill 294217 00:27:35.746 Received shutdown signal, test time was about 10.000000 seconds 00:27:35.746 00:27:35.746 Latency(us) 00:27:35.746 [2024-12-06T16:43:17.585Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:27:35.746 [2024-12-06T16:43:17.585Z] =================================================================================================================== 00:27:35.746 [2024-12-06T16:43:17.585Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@978 -- # wait 294217 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@17 -- # nvmftestfini 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@516 -- # nvmfcleanup 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@121 -- # sync 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@124 -- # set +e 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@125 -- # for i in {1..20} 00:27:35.746 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:27:35.746 rmmod nvme_tcp 00:27:35.746 rmmod nvme_fabrics 00:27:35.746 rmmod nvme_keyring 00:27:36.005 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:27:36.005 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@128 -- # set -e 00:27:36.005 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@129 -- # return 0 00:27:36.005 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@517 -- # '[' -n 294071 ']' 00:27:36.005 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@518 -- # killprocess 294071 00:27:36.005 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@954 -- # '[' -z 294071 ']' 00:27:36.005 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@958 -- # kill -0 294071 00:27:36.005 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@959 -- # uname 00:27:36.005 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:36.005 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 294071 00:27:36.005 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:27:36.005 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:27:36.005 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@972 -- # echo 'killing process with pid 294071' 00:27:36.005 killing process with pid 294071 00:27:36.005 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@973 -- # kill 294071 00:27:36.005 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@978 -- # wait 294071 00:27:36.263 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:27:36.263 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:27:36.263 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:27:36.263 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@297 -- # iptr 00:27:36.263 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@791 -- # iptables-save 00:27:36.263 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:27:36.263 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@791 -- # iptables-restore 00:27:36.263 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:27:36.263 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@302 -- # remove_spdk_ns 00:27:36.263 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:36.263 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:27:36.263 17:43:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:27:38.172 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:27:38.172 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@18 -- # rm -f /tmp/spdk-psk.o1F 00:27:38.172 00:27:38.172 real 0m17.040s 00:27:38.172 user 0m22.896s 00:27:38.172 sys 0m5.148s 00:27:38.172 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1130 -- # xtrace_disable 00:27:38.172 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:27:38.172 ************************************ 00:27:38.172 END TEST nvmf_fips 00:27:38.172 ************************************ 00:27:38.172 17:43:19 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@43 -- # run_test nvmf_control_msg_list /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/control_msg_list.sh --transport=tcp 00:27:38.172 17:43:19 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:27:38.172 17:43:19 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:27:38.172 17:43:19 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:27:38.172 ************************************ 00:27:38.172 START TEST nvmf_control_msg_list 00:27:38.172 ************************************ 00:27:38.172 17:43:19 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/control_msg_list.sh --transport=tcp 00:27:38.172 * Looking for test storage... 00:27:38.432 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1711 -- # lcov --version 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@333 -- # local ver1 ver1_l 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@334 -- # local ver2 ver2_l 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@336 -- # IFS=.-: 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@336 -- # read -ra ver1 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@337 -- # IFS=.-: 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@337 -- # read -ra ver2 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@338 -- # local 'op=<' 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@340 -- # ver1_l=2 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@341 -- # ver2_l=1 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@344 -- # case "$op" in 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@345 -- # : 1 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@364 -- # (( v = 0 )) 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@365 -- # decimal 1 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@353 -- # local d=1 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@355 -- # echo 1 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@365 -- # ver1[v]=1 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@366 -- # decimal 2 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@353 -- # local d=2 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@355 -- # echo 2 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@366 -- # ver2[v]=2 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@368 -- # return 0 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:27:38.432 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:38.432 --rc genhtml_branch_coverage=1 00:27:38.432 --rc genhtml_function_coverage=1 00:27:38.432 --rc genhtml_legend=1 00:27:38.432 --rc geninfo_all_blocks=1 00:27:38.432 --rc geninfo_unexecuted_blocks=1 00:27:38.432 00:27:38.432 ' 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:27:38.432 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:38.432 --rc genhtml_branch_coverage=1 00:27:38.432 --rc genhtml_function_coverage=1 00:27:38.432 --rc genhtml_legend=1 00:27:38.432 --rc geninfo_all_blocks=1 00:27:38.432 --rc geninfo_unexecuted_blocks=1 00:27:38.432 00:27:38.432 ' 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:27:38.432 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:38.432 --rc genhtml_branch_coverage=1 00:27:38.432 --rc genhtml_function_coverage=1 00:27:38.432 --rc genhtml_legend=1 00:27:38.432 --rc geninfo_all_blocks=1 00:27:38.432 --rc geninfo_unexecuted_blocks=1 00:27:38.432 00:27:38.432 ' 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:27:38.432 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:38.432 --rc genhtml_branch_coverage=1 00:27:38.432 --rc genhtml_function_coverage=1 00:27:38.432 --rc genhtml_legend=1 00:27:38.432 --rc geninfo_all_blocks=1 00:27:38.432 --rc geninfo_unexecuted_blocks=1 00:27:38.432 00:27:38.432 ' 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@7 -- # uname -s 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@15 -- # shopt -s extglob 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:27:38.432 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:38.433 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:38.433 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:38.433 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- paths/export.sh@5 -- # export PATH 00:27:38.433 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:38.433 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@51 -- # : 0 00:27:38.433 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:27:38.433 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:27:38.433 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:27:38.433 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:27:38.433 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:27:38.433 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:27:38.433 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:27:38.433 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:27:38.433 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:27:38.433 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@55 -- # have_pci_nics=0 00:27:38.433 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@12 -- # nvmftestinit 00:27:38.433 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:27:38.433 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:27:38.433 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@476 -- # prepare_net_devs 00:27:38.433 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@438 -- # local -g is_hw=no 00:27:38.433 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@440 -- # remove_spdk_ns 00:27:38.433 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:38.433 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:27:38.433 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:27:38.433 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:27:38.433 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:27:38.433 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@309 -- # xtrace_disable 00:27:38.433 17:43:20 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:27:40.970 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:27:40.970 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@315 -- # pci_devs=() 00:27:40.970 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@315 -- # local -a pci_devs 00:27:40.970 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@316 -- # pci_net_devs=() 00:27:40.970 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:27:40.970 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@317 -- # pci_drivers=() 00:27:40.970 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@317 -- # local -A pci_drivers 00:27:40.970 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@319 -- # net_devs=() 00:27:40.970 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@319 -- # local -ga net_devs 00:27:40.970 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@320 -- # e810=() 00:27:40.970 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@320 -- # local -ga e810 00:27:40.970 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@321 -- # x722=() 00:27:40.970 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@321 -- # local -ga x722 00:27:40.970 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@322 -- # mlx=() 00:27:40.970 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@322 -- # local -ga mlx 00:27:40.970 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:27:40.970 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:27:40.970 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:27:40.971 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:27:40.971 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@418 -- # [[ up == up ]] 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:27:40.971 Found net devices under 0000:0a:00.0: cvl_0_0 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@418 -- # [[ up == up ]] 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:27:40.971 Found net devices under 0000:0a:00.1: cvl_0_1 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@442 -- # is_hw=yes 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:27:40.971 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:27:40.971 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.328 ms 00:27:40.971 00:27:40.971 --- 10.0.0.2 ping statistics --- 00:27:40.971 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:27:40.971 rtt min/avg/max/mdev = 0.328/0.328/0.328/0.000 ms 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:27:40.971 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:27:40.971 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.160 ms 00:27:40.971 00:27:40.971 --- 10.0.0.1 ping statistics --- 00:27:40.971 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:27:40.971 rtt min/avg/max/mdev = 0.160/0.160/0.160/0.000 ms 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@450 -- # return 0 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@13 -- # nvmfappstart 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@726 -- # xtrace_disable 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:27:40.971 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@509 -- # nvmfpid=297478 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@510 -- # waitforlisten 297478 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@835 -- # '[' -z 297478 ']' 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@840 -- # local max_retries=100 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:40.972 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@844 -- # xtrace_disable 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:27:40.972 [2024-12-06 17:43:22.443417] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:27:40.972 [2024-12-06 17:43:22.443491] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:27:40.972 [2024-12-06 17:43:22.514391] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:40.972 [2024-12-06 17:43:22.560622] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:27:40.972 [2024-12-06 17:43:22.560708] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:27:40.972 [2024-12-06 17:43:22.560725] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:27:40.972 [2024-12-06 17:43:22.560736] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:27:40.972 [2024-12-06 17:43:22.560746] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:27:40.972 [2024-12-06 17:43:22.561313] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@868 -- # return 0 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@732 -- # xtrace_disable 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@15 -- # subnqn=nqn.2024-07.io.spdk:cnode0 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@16 -- # perf=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@19 -- # rpc_cmd nvmf_create_transport '-t tcp -o' --in-capsule-data-size 768 --control-msg-num 1 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:27:40.972 [2024-12-06 17:43:22.701953] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@20 -- # rpc_cmd nvmf_create_subsystem nqn.2024-07.io.spdk:cnode0 -a 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@21 -- # rpc_cmd bdev_malloc_create -b Malloc0 32 512 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:27:40.972 Malloc0 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@22 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2024-07.io.spdk:cnode0 Malloc0 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@23 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2024-07.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:27:40.972 [2024-12-06 17:43:22.740479] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@27 -- # perf_pid1=297612 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0x2 -q 1 -o 4096 -w randread -t 1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@29 -- # perf_pid2=297613 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0x4 -q 1 -o 4096 -w randread -t 1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@31 -- # perf_pid3=297614 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0x8 -q 1 -o 4096 -w randread -t 1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:27:40.972 17:43:22 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@33 -- # wait 297612 00:27:41.230 [2024-12-06 17:43:22.819548] subsystem.c:1641:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:27:41.230 [2024-12-06 17:43:22.819867] subsystem.c:1641:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:27:41.230 [2024-12-06 17:43:22.820138] subsystem.c:1641:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:27:42.163 Initializing NVMe Controllers 00:27:42.163 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2024-07.io.spdk:cnode0 00:27:42.163 Associating TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 with lcore 3 00:27:42.163 Initialization complete. Launching workers. 00:27:42.163 ======================================================== 00:27:42.163 Latency(us) 00:27:42.163 Device Information : IOPS MiB/s Average min max 00:27:42.163 TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 from core 3: 6271.00 24.50 159.08 151.62 513.47 00:27:42.163 ======================================================== 00:27:42.163 Total : 6271.00 24.50 159.08 151.62 513.47 00:27:42.163 00:27:42.163 Initializing NVMe Controllers 00:27:42.163 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2024-07.io.spdk:cnode0 00:27:42.163 Associating TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 with lcore 1 00:27:42.163 Initialization complete. Launching workers. 00:27:42.163 ======================================================== 00:27:42.164 Latency(us) 00:27:42.164 Device Information : IOPS MiB/s Average min max 00:27:42.164 TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 from core 1: 25.00 0.10 40903.88 40831.16 40999.00 00:27:42.164 ======================================================== 00:27:42.164 Total : 25.00 0.10 40903.88 40831.16 40999.00 00:27:42.164 00:27:42.164 17:43:23 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@34 -- # wait 297613 00:27:42.421 Initializing NVMe Controllers 00:27:42.421 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2024-07.io.spdk:cnode0 00:27:42.421 Associating TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 with lcore 2 00:27:42.421 Initialization complete. Launching workers. 00:27:42.421 ======================================================== 00:27:42.421 Latency(us) 00:27:42.421 Device Information : IOPS MiB/s Average min max 00:27:42.421 TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 from core 2: 25.00 0.10 40901.99 40800.71 40968.41 00:27:42.421 ======================================================== 00:27:42.421 Total : 25.00 0.10 40901.99 40800.71 40968.41 00:27:42.421 00:27:42.421 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@35 -- # wait 297614 00:27:42.421 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:27:42.421 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@38 -- # nvmftestfini 00:27:42.421 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@516 -- # nvmfcleanup 00:27:42.421 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@121 -- # sync 00:27:42.421 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:27:42.421 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@124 -- # set +e 00:27:42.421 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@125 -- # for i in {1..20} 00:27:42.421 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:27:42.421 rmmod nvme_tcp 00:27:42.421 rmmod nvme_fabrics 00:27:42.421 rmmod nvme_keyring 00:27:42.421 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:27:42.421 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@128 -- # set -e 00:27:42.421 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@129 -- # return 0 00:27:42.421 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@517 -- # '[' -n 297478 ']' 00:27:42.421 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@518 -- # killprocess 297478 00:27:42.421 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@954 -- # '[' -z 297478 ']' 00:27:42.421 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@958 -- # kill -0 297478 00:27:42.421 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@959 -- # uname 00:27:42.421 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:42.421 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 297478 00:27:42.421 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:27:42.421 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:27:42.421 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@972 -- # echo 'killing process with pid 297478' 00:27:42.421 killing process with pid 297478 00:27:42.421 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@973 -- # kill 297478 00:27:42.421 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@978 -- # wait 297478 00:27:42.680 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:27:42.680 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:27:42.680 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:27:42.680 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@297 -- # iptr 00:27:42.680 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@791 -- # iptables-save 00:27:42.680 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@791 -- # iptables-restore 00:27:42.680 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:27:42.680 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:27:42.680 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@302 -- # remove_spdk_ns 00:27:42.680 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:42.680 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:27:42.680 17:43:24 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:27:44.591 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:27:44.591 00:27:44.591 real 0m6.428s 00:27:44.591 user 0m5.807s 00:27:44.591 sys 0m2.646s 00:27:44.591 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1130 -- # xtrace_disable 00:27:44.591 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:27:44.591 ************************************ 00:27:44.591 END TEST nvmf_control_msg_list 00:27:44.591 ************************************ 00:27:44.591 17:43:26 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@44 -- # run_test nvmf_wait_for_buf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/wait_for_buf.sh --transport=tcp 00:27:44.591 17:43:26 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:27:44.591 17:43:26 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:27:44.591 17:43:26 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:27:44.849 ************************************ 00:27:44.849 START TEST nvmf_wait_for_buf 00:27:44.849 ************************************ 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/wait_for_buf.sh --transport=tcp 00:27:44.849 * Looking for test storage... 00:27:44.849 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1711 -- # lcov --version 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@333 -- # local ver1 ver1_l 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@334 -- # local ver2 ver2_l 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@336 -- # IFS=.-: 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@336 -- # read -ra ver1 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@337 -- # IFS=.-: 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@337 -- # read -ra ver2 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@338 -- # local 'op=<' 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@340 -- # ver1_l=2 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@341 -- # ver2_l=1 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@344 -- # case "$op" in 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@345 -- # : 1 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@364 -- # (( v = 0 )) 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@365 -- # decimal 1 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@353 -- # local d=1 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@355 -- # echo 1 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@365 -- # ver1[v]=1 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@366 -- # decimal 2 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@353 -- # local d=2 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@355 -- # echo 2 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@366 -- # ver2[v]=2 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@368 -- # return 0 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:27:44.849 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:44.849 --rc genhtml_branch_coverage=1 00:27:44.849 --rc genhtml_function_coverage=1 00:27:44.849 --rc genhtml_legend=1 00:27:44.849 --rc geninfo_all_blocks=1 00:27:44.849 --rc geninfo_unexecuted_blocks=1 00:27:44.849 00:27:44.849 ' 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:27:44.849 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:44.849 --rc genhtml_branch_coverage=1 00:27:44.849 --rc genhtml_function_coverage=1 00:27:44.849 --rc genhtml_legend=1 00:27:44.849 --rc geninfo_all_blocks=1 00:27:44.849 --rc geninfo_unexecuted_blocks=1 00:27:44.849 00:27:44.849 ' 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:27:44.849 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:44.849 --rc genhtml_branch_coverage=1 00:27:44.849 --rc genhtml_function_coverage=1 00:27:44.849 --rc genhtml_legend=1 00:27:44.849 --rc geninfo_all_blocks=1 00:27:44.849 --rc geninfo_unexecuted_blocks=1 00:27:44.849 00:27:44.849 ' 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:27:44.849 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:44.849 --rc genhtml_branch_coverage=1 00:27:44.849 --rc genhtml_function_coverage=1 00:27:44.849 --rc genhtml_legend=1 00:27:44.849 --rc geninfo_all_blocks=1 00:27:44.849 --rc geninfo_unexecuted_blocks=1 00:27:44.849 00:27:44.849 ' 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@7 -- # uname -s 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:27:44.849 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@15 -- # shopt -s extglob 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- paths/export.sh@5 -- # export PATH 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@51 -- # : 0 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:27:44.850 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@55 -- # have_pci_nics=0 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@12 -- # nvmftestinit 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@476 -- # prepare_net_devs 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@438 -- # local -g is_hw=no 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@440 -- # remove_spdk_ns 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@309 -- # xtrace_disable 00:27:44.850 17:43:26 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@315 -- # pci_devs=() 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@315 -- # local -a pci_devs 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@316 -- # pci_net_devs=() 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@317 -- # pci_drivers=() 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@317 -- # local -A pci_drivers 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@319 -- # net_devs=() 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@319 -- # local -ga net_devs 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@320 -- # e810=() 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@320 -- # local -ga e810 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@321 -- # x722=() 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@321 -- # local -ga x722 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@322 -- # mlx=() 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@322 -- # local -ga mlx 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:27:47.373 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:27:47.373 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@418 -- # [[ up == up ]] 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:27:47.373 Found net devices under 0000:0a:00.0: cvl_0_0 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@418 -- # [[ up == up ]] 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:27:47.373 Found net devices under 0000:0a:00.1: cvl_0_1 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@442 -- # is_hw=yes 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:27:47.373 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:27:47.374 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:27:47.374 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.191 ms 00:27:47.374 00:27:47.374 --- 10.0.0.2 ping statistics --- 00:27:47.374 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:27:47.374 rtt min/avg/max/mdev = 0.191/0.191/0.191/0.000 ms 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:27:47.374 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:27:47.374 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.094 ms 00:27:47.374 00:27:47.374 --- 10.0.0.1 ping statistics --- 00:27:47.374 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:27:47.374 rtt min/avg/max/mdev = 0.094/0.094/0.094/0.000 ms 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@450 -- # return 0 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@13 -- # nvmfappstart --wait-for-rpc 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@726 -- # xtrace_disable 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@509 -- # nvmfpid=299699 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --wait-for-rpc 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@510 -- # waitforlisten 299699 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@835 -- # '[' -z 299699 ']' 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@840 -- # local max_retries=100 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:47.374 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@844 -- # xtrace_disable 00:27:47.374 17:43:28 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:27:47.374 [2024-12-06 17:43:29.037588] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:27:47.374 [2024-12-06 17:43:29.037679] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:27:47.374 [2024-12-06 17:43:29.108403] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:47.374 [2024-12-06 17:43:29.149893] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:27:47.374 [2024-12-06 17:43:29.149955] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:27:47.374 [2024-12-06 17:43:29.149983] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:27:47.374 [2024-12-06 17:43:29.149994] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:27:47.374 [2024-12-06 17:43:29.150004] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:27:47.374 [2024-12-06 17:43:29.150550] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@868 -- # return 0 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@732 -- # xtrace_disable 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@15 -- # subnqn=nqn.2024-07.io.spdk:cnode0 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@16 -- # perf=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@19 -- # rpc_cmd accel_set_options --small-cache-size 0 --large-cache-size 0 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@20 -- # rpc_cmd iobuf_set_options --small-pool-count 154 --small_bufsize=8192 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@21 -- # rpc_cmd framework_start_init 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@22 -- # rpc_cmd bdev_malloc_create -b Malloc0 32 512 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:27:47.633 Malloc0 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@23 -- # rpc_cmd nvmf_create_transport '-t tcp -o' -u 8192 -n 24 -b 24 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:27:47.633 [2024-12-06 17:43:29.403906] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2024-07.io.spdk:cnode0 -a -s SPDK00000000000001 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@25 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2024-07.io.spdk:cnode0 Malloc0 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@26 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2024-07.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:27:47.633 [2024-12-06 17:43:29.428108] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:47.633 17:43:29 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 4 -o 131072 -w randread -t 1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:27:47.891 [2024-12-06 17:43:29.517823] subsystem.c:1641:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:27:49.263 Initializing NVMe Controllers 00:27:49.263 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2024-07.io.spdk:cnode0 00:27:49.263 Associating TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 with lcore 0 00:27:49.263 Initialization complete. Launching workers. 00:27:49.263 ======================================================== 00:27:49.263 Latency(us) 00:27:49.263 Device Information : IOPS MiB/s Average min max 00:27:49.263 TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 from core 0: 129.00 16.12 32291.15 7986.49 63869.80 00:27:49.263 ======================================================== 00:27:49.263 Total : 129.00 16.12 32291.15 7986.49 63869.80 00:27:49.263 00:27:49.263 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@32 -- # rpc_cmd iobuf_get_stats 00:27:49.263 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@32 -- # jq -r '.[] | select(.module == "nvmf_TCP") | .small_pool.retry' 00:27:49.263 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:49.263 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:27:49.263 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:49.263 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@32 -- # retry_count=2038 00:27:49.263 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@33 -- # [[ 2038 -eq 0 ]] 00:27:49.263 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:27:49.263 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@38 -- # nvmftestfini 00:27:49.263 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@516 -- # nvmfcleanup 00:27:49.263 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@121 -- # sync 00:27:49.263 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:27:49.263 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@124 -- # set +e 00:27:49.263 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@125 -- # for i in {1..20} 00:27:49.263 17:43:30 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:27:49.263 rmmod nvme_tcp 00:27:49.263 rmmod nvme_fabrics 00:27:49.263 rmmod nvme_keyring 00:27:49.263 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:27:49.263 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@128 -- # set -e 00:27:49.263 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@129 -- # return 0 00:27:49.263 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@517 -- # '[' -n 299699 ']' 00:27:49.264 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@518 -- # killprocess 299699 00:27:49.264 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@954 -- # '[' -z 299699 ']' 00:27:49.264 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@958 -- # kill -0 299699 00:27:49.264 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@959 -- # uname 00:27:49.264 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:27:49.264 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 299699 00:27:49.264 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:27:49.264 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:27:49.264 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@972 -- # echo 'killing process with pid 299699' 00:27:49.264 killing process with pid 299699 00:27:49.264 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@973 -- # kill 299699 00:27:49.264 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@978 -- # wait 299699 00:27:49.523 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:27:49.523 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:27:49.523 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:27:49.523 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@297 -- # iptr 00:27:49.523 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@791 -- # iptables-save 00:27:49.523 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:27:49.523 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@791 -- # iptables-restore 00:27:49.523 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:27:49.523 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@302 -- # remove_spdk_ns 00:27:49.523 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:49.523 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:27:49.523 17:43:31 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:27:52.061 00:27:52.061 real 0m6.849s 00:27:52.061 user 0m3.214s 00:27:52.061 sys 0m2.103s 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:27:52.061 ************************************ 00:27:52.061 END TEST nvmf_wait_for_buf 00:27:52.061 ************************************ 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@47 -- # '[' 1 -eq 1 ']' 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@48 -- # run_test nvmf_fuzz /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fabrics_fuzz.sh --transport=tcp 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:27:52.061 ************************************ 00:27:52.061 START TEST nvmf_fuzz 00:27:52.061 ************************************ 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fabrics_fuzz.sh --transport=tcp 00:27:52.061 * Looking for test storage... 00:27:52.061 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@1711 -- # lcov --version 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@333 -- # local ver1 ver1_l 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@334 -- # local ver2 ver2_l 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@336 -- # IFS=.-: 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@336 -- # read -ra ver1 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@337 -- # IFS=.-: 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@337 -- # read -ra ver2 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@338 -- # local 'op=<' 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@340 -- # ver1_l=2 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@341 -- # ver2_l=1 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@344 -- # case "$op" in 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@345 -- # : 1 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@364 -- # (( v = 0 )) 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@365 -- # decimal 1 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@353 -- # local d=1 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@355 -- # echo 1 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@365 -- # ver1[v]=1 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@366 -- # decimal 2 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@353 -- # local d=2 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@355 -- # echo 2 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@366 -- # ver2[v]=2 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@368 -- # return 0 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:27:52.061 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:52.061 --rc genhtml_branch_coverage=1 00:27:52.061 --rc genhtml_function_coverage=1 00:27:52.061 --rc genhtml_legend=1 00:27:52.061 --rc geninfo_all_blocks=1 00:27:52.061 --rc geninfo_unexecuted_blocks=1 00:27:52.061 00:27:52.061 ' 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:27:52.061 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:52.061 --rc genhtml_branch_coverage=1 00:27:52.061 --rc genhtml_function_coverage=1 00:27:52.061 --rc genhtml_legend=1 00:27:52.061 --rc geninfo_all_blocks=1 00:27:52.061 --rc geninfo_unexecuted_blocks=1 00:27:52.061 00:27:52.061 ' 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:27:52.061 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:52.061 --rc genhtml_branch_coverage=1 00:27:52.061 --rc genhtml_function_coverage=1 00:27:52.061 --rc genhtml_legend=1 00:27:52.061 --rc geninfo_all_blocks=1 00:27:52.061 --rc geninfo_unexecuted_blocks=1 00:27:52.061 00:27:52.061 ' 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:27:52.061 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:52.061 --rc genhtml_branch_coverage=1 00:27:52.061 --rc genhtml_function_coverage=1 00:27:52.061 --rc genhtml_legend=1 00:27:52.061 --rc geninfo_all_blocks=1 00:27:52.061 --rc geninfo_unexecuted_blocks=1 00:27:52.061 00:27:52.061 ' 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- target/fabrics_fuzz.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@7 -- # uname -s 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:27:52.061 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@15 -- # shopt -s extglob 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- paths/export.sh@5 -- # export PATH 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@51 -- # : 0 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:27:52.062 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@55 -- # have_pci_nics=0 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- target/fabrics_fuzz.sh@11 -- # nvmftestinit 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@476 -- # prepare_net_devs 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@438 -- # local -g is_hw=no 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@440 -- # remove_spdk_ns 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@309 -- # xtrace_disable 00:27:52.062 17:43:33 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@10 -- # set +x 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@315 -- # pci_devs=() 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@315 -- # local -a pci_devs 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@316 -- # pci_net_devs=() 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@317 -- # pci_drivers=() 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@317 -- # local -A pci_drivers 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@319 -- # net_devs=() 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@319 -- # local -ga net_devs 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@320 -- # e810=() 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@320 -- # local -ga e810 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@321 -- # x722=() 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@321 -- # local -ga x722 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@322 -- # mlx=() 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@322 -- # local -ga mlx 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:27:53.969 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:27:53.969 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@418 -- # [[ up == up ]] 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:27:53.969 Found net devices under 0000:0a:00.0: cvl_0_0 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@418 -- # [[ up == up ]] 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:27:53.969 Found net devices under 0000:0a:00.1: cvl_0_1 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@442 -- # is_hw=yes 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:27:53.969 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:27:54.228 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:27:54.228 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.322 ms 00:27:54.228 00:27:54.228 --- 10.0.0.2 ping statistics --- 00:27:54.228 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:27:54.228 rtt min/avg/max/mdev = 0.322/0.322/0.322/0.000 ms 00:27:54.228 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:27:54.228 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:27:54.228 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.167 ms 00:27:54.228 00:27:54.228 --- 10.0.0.1 ping statistics --- 00:27:54.228 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:27:54.228 rtt min/avg/max/mdev = 0.167/0.167/0.167/0.000 ms 00:27:54.228 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:27:54.228 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@450 -- # return 0 00:27:54.228 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:27:54.228 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:27:54.228 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:27:54.228 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:27:54.228 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:27:54.228 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:27:54.228 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:27:54.228 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- target/fabrics_fuzz.sh@14 -- # nvmfpid=301910 00:27:54.228 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- target/fabrics_fuzz.sh@13 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1 00:27:54.228 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- target/fabrics_fuzz.sh@16 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; killprocess $nvmfpid; nvmftestfini $1; exit 1' SIGINT SIGTERM EXIT 00:27:54.228 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- target/fabrics_fuzz.sh@18 -- # waitforlisten 301910 00:27:54.228 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@835 -- # '[' -z 301910 ']' 00:27:54.228 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:54.228 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@840 -- # local max_retries=100 00:27:54.228 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:54.228 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:54.228 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@844 -- # xtrace_disable 00:27:54.228 17:43:35 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@10 -- # set +x 00:27:54.486 17:43:36 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:27:54.486 17:43:36 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@868 -- # return 0 00:27:54.486 17:43:36 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- target/fabrics_fuzz.sh@19 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:27:54.486 17:43:36 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:54.486 17:43:36 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@10 -- # set +x 00:27:54.486 17:43:36 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:54.486 17:43:36 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- target/fabrics_fuzz.sh@21 -- # rpc_cmd bdev_malloc_create -b Malloc0 64 512 00:27:54.486 17:43:36 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:54.486 17:43:36 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@10 -- # set +x 00:27:54.486 Malloc0 00:27:54.486 17:43:36 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:54.487 17:43:36 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- target/fabrics_fuzz.sh@23 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:27:54.487 17:43:36 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:54.487 17:43:36 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@10 -- # set +x 00:27:54.487 17:43:36 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:54.487 17:43:36 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- target/fabrics_fuzz.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:27:54.487 17:43:36 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:54.487 17:43:36 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@10 -- # set +x 00:27:54.487 17:43:36 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:54.487 17:43:36 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- target/fabrics_fuzz.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:27:54.487 17:43:36 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@563 -- # xtrace_disable 00:27:54.487 17:43:36 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@10 -- # set +x 00:27:54.487 17:43:36 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:27:54.487 17:43:36 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- target/fabrics_fuzz.sh@27 -- # trid='trtype:tcp adrfam:IPv4 subnqn:nqn.2016-06.io.spdk:cnode1 traddr:10.0.0.2 trsvcid:4420' 00:27:54.487 17:43:36 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- target/fabrics_fuzz.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/fuzz/nvme_fuzz/nvme_fuzz -m 0x2 -t 30 -S 123456 -F 'trtype:tcp adrfam:IPv4 subnqn:nqn.2016-06.io.spdk:cnode1 traddr:10.0.0.2 trsvcid:4420' -N -a 00:28:26.542 Fuzzing completed. Shutting down the fuzz application 00:28:26.542 00:28:26.542 Dumping successful admin opcodes: 00:28:26.542 9, 10, 00:28:26.542 Dumping successful io opcodes: 00:28:26.542 0, 9, 00:28:26.542 NS: 0x2000008eff00 I/O qp, Total commands completed: 507747, total successful commands: 2925, random_seed: 2916502784 00:28:26.542 NS: 0x2000008eff00 admin qp, Total commands completed: 61040, total successful commands: 15, random_seed: 1685323904 00:28:26.542 17:44:06 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- target/fabrics_fuzz.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/fuzz/nvme_fuzz/nvme_fuzz -m 0x2 -F 'trtype:tcp adrfam:IPv4 subnqn:nqn.2016-06.io.spdk:cnode1 traddr:10.0.0.2 trsvcid:4420' -j /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/fuzz/nvme_fuzz/example.json -a 00:28:26.542 Fuzzing completed. Shutting down the fuzz application 00:28:26.542 00:28:26.542 Dumping successful admin opcodes: 00:28:26.542 00:28:26.542 Dumping successful io opcodes: 00:28:26.542 00:28:26.542 NS: 0x2000008eff00 I/O qp, Total commands completed: 0, total successful commands: 0, random_seed: 3759298413 00:28:26.542 NS: 0x2000008eff00 admin qp, Total commands completed: 16, total successful commands: 0, random_seed: 3759405435 00:28:26.542 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- target/fabrics_fuzz.sh@34 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:28:26.542 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:26.542 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@10 -- # set +x 00:28:26.543 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:26.543 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- target/fabrics_fuzz.sh@36 -- # trap - SIGINT SIGTERM EXIT 00:28:26.543 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- target/fabrics_fuzz.sh@38 -- # nvmftestfini 00:28:26.543 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@516 -- # nvmfcleanup 00:28:26.543 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@121 -- # sync 00:28:26.543 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:28:26.543 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@124 -- # set +e 00:28:26.543 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@125 -- # for i in {1..20} 00:28:26.543 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:28:26.543 rmmod nvme_tcp 00:28:26.543 rmmod nvme_fabrics 00:28:26.543 rmmod nvme_keyring 00:28:26.543 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:28:26.543 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@128 -- # set -e 00:28:26.543 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@129 -- # return 0 00:28:26.543 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@517 -- # '[' -n 301910 ']' 00:28:26.543 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@518 -- # killprocess 301910 00:28:26.543 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@954 -- # '[' -z 301910 ']' 00:28:26.543 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@958 -- # kill -0 301910 00:28:26.543 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@959 -- # uname 00:28:26.543 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:28:26.543 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 301910 00:28:26.543 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:28:26.543 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:28:26.543 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@972 -- # echo 'killing process with pid 301910' 00:28:26.543 killing process with pid 301910 00:28:26.543 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@973 -- # kill 301910 00:28:26.543 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@978 -- # wait 301910 00:28:26.802 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:28:26.803 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:28:26.803 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:28:26.803 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@297 -- # iptr 00:28:26.803 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@791 -- # iptables-save 00:28:26.803 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:28:26.803 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@791 -- # iptables-restore 00:28:26.803 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:28:26.803 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@302 -- # remove_spdk_ns 00:28:26.803 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:28:26.803 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:28:26.803 17:44:08 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- target/fabrics_fuzz.sh@39 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf_fuzz_logs1.txt /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf_fuzz_logs2.txt 00:28:29.337 00:28:29.337 real 0m37.334s 00:28:29.337 user 0m51.520s 00:28:29.337 sys 0m15.008s 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@1130 -- # xtrace_disable 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_fuzz -- common/autotest_common.sh@10 -- # set +x 00:28:29.337 ************************************ 00:28:29.337 END TEST nvmf_fuzz 00:28:29.337 ************************************ 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@49 -- # run_test nvmf_multiconnection /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multiconnection.sh --transport=tcp 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:28:29.337 ************************************ 00:28:29.337 START TEST nvmf_multiconnection 00:28:29.337 ************************************ 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multiconnection.sh --transport=tcp 00:28:29.337 * Looking for test storage... 00:28:29.337 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1711 -- # lcov --version 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@333 -- # local ver1 ver1_l 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@334 -- # local ver2 ver2_l 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@336 -- # IFS=.-: 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@336 -- # read -ra ver1 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@337 -- # IFS=.-: 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@337 -- # read -ra ver2 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@338 -- # local 'op=<' 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@340 -- # ver1_l=2 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@341 -- # ver2_l=1 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@344 -- # case "$op" in 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@345 -- # : 1 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@364 -- # (( v = 0 )) 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@365 -- # decimal 1 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@353 -- # local d=1 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@355 -- # echo 1 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@365 -- # ver1[v]=1 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@366 -- # decimal 2 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@353 -- # local d=2 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@355 -- # echo 2 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@366 -- # ver2[v]=2 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@368 -- # return 0 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:28:29.337 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:28:29.337 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:28:29.338 --rc genhtml_branch_coverage=1 00:28:29.338 --rc genhtml_function_coverage=1 00:28:29.338 --rc genhtml_legend=1 00:28:29.338 --rc geninfo_all_blocks=1 00:28:29.338 --rc geninfo_unexecuted_blocks=1 00:28:29.338 00:28:29.338 ' 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:28:29.338 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:28:29.338 --rc genhtml_branch_coverage=1 00:28:29.338 --rc genhtml_function_coverage=1 00:28:29.338 --rc genhtml_legend=1 00:28:29.338 --rc geninfo_all_blocks=1 00:28:29.338 --rc geninfo_unexecuted_blocks=1 00:28:29.338 00:28:29.338 ' 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:28:29.338 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:28:29.338 --rc genhtml_branch_coverage=1 00:28:29.338 --rc genhtml_function_coverage=1 00:28:29.338 --rc genhtml_legend=1 00:28:29.338 --rc geninfo_all_blocks=1 00:28:29.338 --rc geninfo_unexecuted_blocks=1 00:28:29.338 00:28:29.338 ' 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:28:29.338 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:28:29.338 --rc genhtml_branch_coverage=1 00:28:29.338 --rc genhtml_function_coverage=1 00:28:29.338 --rc genhtml_legend=1 00:28:29.338 --rc geninfo_all_blocks=1 00:28:29.338 --rc geninfo_unexecuted_blocks=1 00:28:29.338 00:28:29.338 ' 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@7 -- # uname -s 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@15 -- # shopt -s extglob 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- paths/export.sh@5 -- # export PATH 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@51 -- # : 0 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:28:29.338 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@55 -- # have_pci_nics=0 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@11 -- # MALLOC_BDEV_SIZE=64 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@14 -- # NVMF_SUBSYS=11 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@16 -- # nvmftestinit 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@476 -- # prepare_net_devs 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@438 -- # local -g is_hw=no 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@440 -- # remove_spdk_ns 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@309 -- # xtrace_disable 00:28:29.338 17:44:10 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@315 -- # pci_devs=() 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@315 -- # local -a pci_devs 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@316 -- # pci_net_devs=() 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@317 -- # pci_drivers=() 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@317 -- # local -A pci_drivers 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@319 -- # net_devs=() 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@319 -- # local -ga net_devs 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@320 -- # e810=() 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@320 -- # local -ga e810 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@321 -- # x722=() 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@321 -- # local -ga x722 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@322 -- # mlx=() 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@322 -- # local -ga mlx 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:28:31.236 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:28:31.236 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:28:31.236 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@418 -- # [[ up == up ]] 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:28:31.237 Found net devices under 0000:0a:00.0: cvl_0_0 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@418 -- # [[ up == up ]] 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:28:31.237 Found net devices under 0000:0a:00.1: cvl_0_1 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@442 -- # is_hw=yes 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:28:31.237 17:44:12 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:28:31.237 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:28:31.237 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:28:31.237 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:28:31.237 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:28:31.494 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:28:31.494 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:28:31.494 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:28:31.494 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:28:31.494 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:28:31.494 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.289 ms 00:28:31.494 00:28:31.494 --- 10.0.0.2 ping statistics --- 00:28:31.494 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:28:31.494 rtt min/avg/max/mdev = 0.289/0.289/0.289/0.000 ms 00:28:31.494 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:28:31.494 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:28:31.494 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.136 ms 00:28:31.494 00:28:31.494 --- 10.0.0.1 ping statistics --- 00:28:31.494 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:28:31.494 rtt min/avg/max/mdev = 0.136/0.136/0.136/0.000 ms 00:28:31.494 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:28:31.494 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@450 -- # return 0 00:28:31.494 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:28:31.494 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:28:31.494 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:28:31.494 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:28:31.494 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:28:31.495 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:28:31.495 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:28:31.495 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@17 -- # nvmfappstart -m 0xF 00:28:31.495 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:28:31.495 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@726 -- # xtrace_disable 00:28:31.495 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:31.495 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@509 -- # nvmfpid=307639 00:28:31.495 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:28:31.495 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@510 -- # waitforlisten 307639 00:28:31.495 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@835 -- # '[' -z 307639 ']' 00:28:31.495 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:28:31.495 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@840 -- # local max_retries=100 00:28:31.495 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:28:31.495 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:28:31.495 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@844 -- # xtrace_disable 00:28:31.495 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:31.495 [2024-12-06 17:44:13.183778] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:28:31.495 [2024-12-06 17:44:13.183848] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:28:31.495 [2024-12-06 17:44:13.256216] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:28:31.495 [2024-12-06 17:44:13.303933] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:28:31.495 [2024-12-06 17:44:13.303996] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:28:31.495 [2024-12-06 17:44:13.304020] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:28:31.495 [2024-12-06 17:44:13.304031] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:28:31.495 [2024-12-06 17:44:13.304041] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:28:31.495 [2024-12-06 17:44:13.305509] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:28:31.495 [2024-12-06 17:44:13.305575] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:28:31.495 [2024-12-06 17:44:13.305639] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:28:31.495 [2024-12-06 17:44:13.305642] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@868 -- # return 0 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@732 -- # xtrace_disable 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@19 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:31.752 [2024-12-06 17:44:13.448442] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@21 -- # seq 1 11 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@21 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@22 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:31.752 Malloc1 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@23 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK1 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:31.752 [2024-12-06 17:44:13.514973] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@21 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@22 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc2 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:31.752 Malloc2 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@23 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 -a -s SPDK2 00:28:31.752 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:31.753 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:31.753 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:31.753 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 Malloc2 00:28:31.753 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:31.753 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:31.753 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:31.753 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:28:31.753 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:31.753 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:31.753 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:31.753 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@21 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:28:31.753 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@22 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc3 00:28:31.753 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:31.753 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.011 Malloc3 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@23 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode3 -a -s SPDK3 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode3 Malloc3 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode3 -t tcp -a 10.0.0.2 -s 4420 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@21 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@22 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc4 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.011 Malloc4 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@23 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode4 -a -s SPDK4 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode4 Malloc4 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode4 -t tcp -a 10.0.0.2 -s 4420 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@21 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@22 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc5 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.011 Malloc5 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@23 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode5 -a -s SPDK5 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode5 Malloc5 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode5 -t tcp -a 10.0.0.2 -s 4420 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@21 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@22 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc6 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.011 Malloc6 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@23 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode6 -a -s SPDK6 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode6 Malloc6 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode6 -t tcp -a 10.0.0.2 -s 4420 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@21 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@22 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc7 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.011 Malloc7 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@23 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode7 -a -s SPDK7 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode7 Malloc7 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.011 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.012 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode7 -t tcp -a 10.0.0.2 -s 4420 00:28:32.012 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.012 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.012 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.012 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@21 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:28:32.012 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@22 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc8 00:28:32.012 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.012 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.012 Malloc8 00:28:32.012 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.012 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@23 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode8 -a -s SPDK8 00:28:32.012 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.012 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode8 Malloc8 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode8 -t tcp -a 10.0.0.2 -s 4420 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@21 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@22 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc9 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.270 Malloc9 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@23 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode9 -a -s SPDK9 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode9 Malloc9 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode9 -t tcp -a 10.0.0.2 -s 4420 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@21 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@22 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc10 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.270 Malloc10 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@23 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode10 -a -s SPDK10 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode10 Malloc10 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode10 -t tcp -a 10.0.0.2 -s 4420 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@21 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@22 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc11 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.270 Malloc11 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@23 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode11 -a -s SPDK11 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode11 Malloc11 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.270 17:44:13 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.270 17:44:14 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.270 17:44:14 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode11 -t tcp -a 10.0.0.2 -s 4420 00:28:32.270 17:44:14 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:28:32.270 17:44:14 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:28:32.270 17:44:14 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:28:32.270 17:44:14 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@28 -- # seq 1 11 00:28:32.270 17:44:14 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@28 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:28:32.270 17:44:14 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@29 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:28:33.202 17:44:14 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@30 -- # waitforserial SPDK1 00:28:33.202 17:44:14 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1202 -- # local i=0 00:28:33.202 17:44:14 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:28:33.202 17:44:14 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:28:33.202 17:44:14 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1209 -- # sleep 2 00:28:35.096 17:44:16 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:28:35.096 17:44:16 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:28:35.096 17:44:16 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # grep -c SPDK1 00:28:35.096 17:44:16 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:28:35.096 17:44:16 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:28:35.096 17:44:16 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1212 -- # return 0 00:28:35.096 17:44:16 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@28 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:28:35.096 17:44:16 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@29 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode2 -a 10.0.0.2 -s 4420 00:28:35.661 17:44:17 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@30 -- # waitforserial SPDK2 00:28:35.661 17:44:17 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1202 -- # local i=0 00:28:35.661 17:44:17 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:28:35.661 17:44:17 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:28:35.661 17:44:17 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1209 -- # sleep 2 00:28:37.555 17:44:19 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:28:37.555 17:44:19 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:28:37.555 17:44:19 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # grep -c SPDK2 00:28:37.555 17:44:19 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:28:37.555 17:44:19 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:28:37.555 17:44:19 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1212 -- # return 0 00:28:37.555 17:44:19 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@28 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:28:37.555 17:44:19 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@29 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode3 -a 10.0.0.2 -s 4420 00:28:38.487 17:44:19 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@30 -- # waitforserial SPDK3 00:28:38.487 17:44:19 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1202 -- # local i=0 00:28:38.487 17:44:19 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:28:38.487 17:44:19 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:28:38.487 17:44:19 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1209 -- # sleep 2 00:28:40.380 17:44:21 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:28:40.380 17:44:21 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:28:40.380 17:44:21 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # grep -c SPDK3 00:28:40.380 17:44:22 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:28:40.380 17:44:22 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:28:40.380 17:44:22 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1212 -- # return 0 00:28:40.380 17:44:22 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@28 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:28:40.380 17:44:22 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@29 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode4 -a 10.0.0.2 -s 4420 00:28:40.946 17:44:22 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@30 -- # waitforserial SPDK4 00:28:40.946 17:44:22 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1202 -- # local i=0 00:28:40.946 17:44:22 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:28:40.946 17:44:22 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:28:40.946 17:44:22 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1209 -- # sleep 2 00:28:43.471 17:44:24 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:28:43.471 17:44:24 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:28:43.471 17:44:24 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # grep -c SPDK4 00:28:43.471 17:44:24 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:28:43.471 17:44:24 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:28:43.471 17:44:24 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1212 -- # return 0 00:28:43.471 17:44:24 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@28 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:28:43.471 17:44:24 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@29 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode5 -a 10.0.0.2 -s 4420 00:28:43.728 17:44:25 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@30 -- # waitforserial SPDK5 00:28:43.729 17:44:25 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1202 -- # local i=0 00:28:43.729 17:44:25 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:28:43.729 17:44:25 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:28:43.729 17:44:25 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1209 -- # sleep 2 00:28:46.253 17:44:27 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:28:46.253 17:44:27 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:28:46.253 17:44:27 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # grep -c SPDK5 00:28:46.253 17:44:27 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:28:46.253 17:44:27 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:28:46.253 17:44:27 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1212 -- # return 0 00:28:46.253 17:44:27 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@28 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:28:46.253 17:44:27 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@29 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode6 -a 10.0.0.2 -s 4420 00:28:46.511 17:44:28 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@30 -- # waitforserial SPDK6 00:28:46.511 17:44:28 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1202 -- # local i=0 00:28:46.511 17:44:28 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:28:46.511 17:44:28 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:28:46.511 17:44:28 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1209 -- # sleep 2 00:28:49.032 17:44:30 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:28:49.032 17:44:30 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:28:49.033 17:44:30 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # grep -c SPDK6 00:28:49.033 17:44:30 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:28:49.033 17:44:30 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:28:49.033 17:44:30 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1212 -- # return 0 00:28:49.033 17:44:30 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@28 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:28:49.033 17:44:30 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@29 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode7 -a 10.0.0.2 -s 4420 00:28:49.597 17:44:31 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@30 -- # waitforserial SPDK7 00:28:49.597 17:44:31 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1202 -- # local i=0 00:28:49.597 17:44:31 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:28:49.597 17:44:31 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:28:49.597 17:44:31 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1209 -- # sleep 2 00:28:51.490 17:44:33 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:28:51.490 17:44:33 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:28:51.490 17:44:33 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # grep -c SPDK7 00:28:51.490 17:44:33 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:28:51.490 17:44:33 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:28:51.490 17:44:33 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1212 -- # return 0 00:28:51.490 17:44:33 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@28 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:28:51.490 17:44:33 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@29 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode8 -a 10.0.0.2 -s 4420 00:28:52.421 17:44:33 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@30 -- # waitforserial SPDK8 00:28:52.421 17:44:33 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1202 -- # local i=0 00:28:52.421 17:44:33 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:28:52.421 17:44:33 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:28:52.421 17:44:33 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1209 -- # sleep 2 00:28:54.314 17:44:35 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:28:54.314 17:44:35 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:28:54.314 17:44:35 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # grep -c SPDK8 00:28:54.314 17:44:35 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:28:54.314 17:44:35 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:28:54.314 17:44:35 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1212 -- # return 0 00:28:54.314 17:44:35 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@28 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:28:54.314 17:44:35 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@29 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode9 -a 10.0.0.2 -s 4420 00:28:55.245 17:44:36 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@30 -- # waitforserial SPDK9 00:28:55.245 17:44:36 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1202 -- # local i=0 00:28:55.245 17:44:36 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:28:55.245 17:44:36 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:28:55.245 17:44:36 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1209 -- # sleep 2 00:28:57.142 17:44:38 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:28:57.142 17:44:38 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:28:57.142 17:44:38 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # grep -c SPDK9 00:28:57.142 17:44:38 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:28:57.142 17:44:38 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:28:57.142 17:44:38 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1212 -- # return 0 00:28:57.142 17:44:38 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@28 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:28:57.142 17:44:38 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@29 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode10 -a 10.0.0.2 -s 4420 00:28:58.073 17:44:39 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@30 -- # waitforserial SPDK10 00:28:58.073 17:44:39 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1202 -- # local i=0 00:28:58.073 17:44:39 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:28:58.073 17:44:39 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:28:58.073 17:44:39 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1209 -- # sleep 2 00:28:59.969 17:44:41 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:28:59.969 17:44:41 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:28:59.969 17:44:41 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # grep -c SPDK10 00:28:59.969 17:44:41 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:28:59.969 17:44:41 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:28:59.970 17:44:41 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1212 -- # return 0 00:28:59.970 17:44:41 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@28 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:28:59.970 17:44:41 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@29 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode11 -a 10.0.0.2 -s 4420 00:29:00.902 17:44:42 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@30 -- # waitforserial SPDK11 00:29:00.902 17:44:42 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1202 -- # local i=0 00:29:00.902 17:44:42 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:29:00.902 17:44:42 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:29:00.902 17:44:42 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1209 -- # sleep 2 00:29:03.428 17:44:44 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:29:03.428 17:44:44 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:29:03.428 17:44:44 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # grep -c SPDK11 00:29:03.428 17:44:44 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:29:03.428 17:44:44 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:29:03.428 17:44:44 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1212 -- # return 0 00:29:03.428 17:44:44 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 262144 -d 64 -t read -r 10 00:29:03.428 [global] 00:29:03.428 thread=1 00:29:03.428 invalidate=1 00:29:03.428 rw=read 00:29:03.428 time_based=1 00:29:03.428 runtime=10 00:29:03.428 ioengine=libaio 00:29:03.428 direct=1 00:29:03.428 bs=262144 00:29:03.428 iodepth=64 00:29:03.428 norandommap=1 00:29:03.428 numjobs=1 00:29:03.428 00:29:03.428 [job0] 00:29:03.428 filename=/dev/nvme0n1 00:29:03.428 [job1] 00:29:03.428 filename=/dev/nvme10n1 00:29:03.428 [job2] 00:29:03.428 filename=/dev/nvme1n1 00:29:03.428 [job3] 00:29:03.428 filename=/dev/nvme2n1 00:29:03.428 [job4] 00:29:03.428 filename=/dev/nvme3n1 00:29:03.428 [job5] 00:29:03.428 filename=/dev/nvme4n1 00:29:03.428 [job6] 00:29:03.428 filename=/dev/nvme5n1 00:29:03.428 [job7] 00:29:03.428 filename=/dev/nvme6n1 00:29:03.428 [job8] 00:29:03.428 filename=/dev/nvme7n1 00:29:03.428 [job9] 00:29:03.428 filename=/dev/nvme8n1 00:29:03.428 [job10] 00:29:03.428 filename=/dev/nvme9n1 00:29:03.428 Could not set queue depth (nvme0n1) 00:29:03.428 Could not set queue depth (nvme10n1) 00:29:03.428 Could not set queue depth (nvme1n1) 00:29:03.428 Could not set queue depth (nvme2n1) 00:29:03.428 Could not set queue depth (nvme3n1) 00:29:03.428 Could not set queue depth (nvme4n1) 00:29:03.428 Could not set queue depth (nvme5n1) 00:29:03.428 Could not set queue depth (nvme6n1) 00:29:03.428 Could not set queue depth (nvme7n1) 00:29:03.428 Could not set queue depth (nvme8n1) 00:29:03.428 Could not set queue depth (nvme9n1) 00:29:03.428 job0: (g=0): rw=read, bs=(R) 256KiB-256KiB, (W) 256KiB-256KiB, (T) 256KiB-256KiB, ioengine=libaio, iodepth=64 00:29:03.428 job1: (g=0): rw=read, bs=(R) 256KiB-256KiB, (W) 256KiB-256KiB, (T) 256KiB-256KiB, ioengine=libaio, iodepth=64 00:29:03.428 job2: (g=0): rw=read, bs=(R) 256KiB-256KiB, (W) 256KiB-256KiB, (T) 256KiB-256KiB, ioengine=libaio, iodepth=64 00:29:03.428 job3: (g=0): rw=read, bs=(R) 256KiB-256KiB, (W) 256KiB-256KiB, (T) 256KiB-256KiB, ioengine=libaio, iodepth=64 00:29:03.428 job4: (g=0): rw=read, bs=(R) 256KiB-256KiB, (W) 256KiB-256KiB, (T) 256KiB-256KiB, ioengine=libaio, iodepth=64 00:29:03.428 job5: (g=0): rw=read, bs=(R) 256KiB-256KiB, (W) 256KiB-256KiB, (T) 256KiB-256KiB, ioengine=libaio, iodepth=64 00:29:03.428 job6: (g=0): rw=read, bs=(R) 256KiB-256KiB, (W) 256KiB-256KiB, (T) 256KiB-256KiB, ioengine=libaio, iodepth=64 00:29:03.428 job7: (g=0): rw=read, bs=(R) 256KiB-256KiB, (W) 256KiB-256KiB, (T) 256KiB-256KiB, ioengine=libaio, iodepth=64 00:29:03.428 job8: (g=0): rw=read, bs=(R) 256KiB-256KiB, (W) 256KiB-256KiB, (T) 256KiB-256KiB, ioengine=libaio, iodepth=64 00:29:03.428 job9: (g=0): rw=read, bs=(R) 256KiB-256KiB, (W) 256KiB-256KiB, (T) 256KiB-256KiB, ioengine=libaio, iodepth=64 00:29:03.428 job10: (g=0): rw=read, bs=(R) 256KiB-256KiB, (W) 256KiB-256KiB, (T) 256KiB-256KiB, ioengine=libaio, iodepth=64 00:29:03.428 fio-3.35 00:29:03.428 Starting 11 threads 00:29:15.615 00:29:15.615 job0: (groupid=0, jobs=1): err= 0: pid=311901: Fri Dec 6 17:44:55 2024 00:29:15.615 read: IOPS=278, BW=69.7MiB/s (73.0MB/s)(705MiB/10121msec) 00:29:15.615 slat (usec): min=13, max=206252, avg=3550.76, stdev=12846.34 00:29:15.615 clat (msec): min=46, max=648, avg=225.93, stdev=98.38 00:29:15.615 lat (msec): min=46, max=648, avg=229.48, stdev=99.70 00:29:15.615 clat percentiles (msec): 00:29:15.615 | 1.00th=[ 63], 5.00th=[ 124], 10.00th=[ 132], 20.00th=[ 146], 00:29:15.615 | 30.00th=[ 163], 40.00th=[ 178], 50.00th=[ 194], 60.00th=[ 218], 00:29:15.615 | 70.00th=[ 251], 80.00th=[ 300], 90.00th=[ 368], 95.00th=[ 430], 00:29:15.615 | 99.00th=[ 535], 99.50th=[ 584], 99.90th=[ 634], 99.95th=[ 651], 00:29:15.615 | 99.99th=[ 651] 00:29:15.615 bw ( KiB/s): min=32768, max=114176, per=9.88%, avg=70546.30, stdev=22574.16, samples=20 00:29:15.615 iops : min= 128, max= 446, avg=275.55, stdev=88.18, samples=20 00:29:15.615 lat (msec) : 50=0.57%, 100=0.64%, 250=68.26%, 500=28.51%, 750=2.02% 00:29:15.615 cpu : usr=0.20%, sys=1.04%, ctx=379, majf=0, minf=4097 00:29:15.615 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.3%, 16=0.6%, 32=1.1%, >=64=97.8% 00:29:15.615 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:29:15.615 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.1%, >=64=0.0% 00:29:15.615 issued rwts: total=2820,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:29:15.615 latency : target=0, window=0, percentile=100.00%, depth=64 00:29:15.615 job1: (groupid=0, jobs=1): err= 0: pid=311902: Fri Dec 6 17:44:55 2024 00:29:15.615 read: IOPS=287, BW=71.8MiB/s (75.3MB/s)(731MiB/10185msec) 00:29:15.615 slat (usec): min=8, max=276916, avg=1372.39, stdev=8905.80 00:29:15.615 clat (msec): min=13, max=1801, avg=221.32, stdev=221.25 00:29:15.615 lat (msec): min=13, max=1801, avg=222.69, stdev=221.81 00:29:15.615 clat percentiles (msec): 00:29:15.615 | 1.00th=[ 26], 5.00th=[ 44], 10.00th=[ 65], 20.00th=[ 92], 00:29:15.615 | 30.00th=[ 134], 40.00th=[ 150], 50.00th=[ 165], 60.00th=[ 180], 00:29:15.615 | 70.00th=[ 209], 80.00th=[ 284], 90.00th=[ 401], 95.00th=[ 625], 00:29:15.615 | 99.00th=[ 1250], 99.50th=[ 1502], 99.90th=[ 1687], 99.95th=[ 1687], 00:29:15.615 | 99.99th=[ 1804] 00:29:15.615 bw ( KiB/s): min=28672, max=114176, per=10.26%, avg=73204.85, stdev=27552.89, samples=20 00:29:15.615 iops : min= 112, max= 446, avg=285.90, stdev=107.67, samples=20 00:29:15.615 lat (msec) : 20=0.48%, 50=5.68%, 100=15.97%, 250=53.86%, 500=16.69% 00:29:15.615 lat (msec) : 750=4.72%, 1000=0.31%, 2000=2.29% 00:29:15.615 cpu : usr=0.08%, sys=0.86%, ctx=594, majf=0, minf=4097 00:29:15.615 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.3%, 16=0.5%, 32=1.1%, >=64=97.8% 00:29:15.615 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:29:15.615 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.1%, >=64=0.0% 00:29:15.615 issued rwts: total=2924,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:29:15.615 latency : target=0, window=0, percentile=100.00%, depth=64 00:29:15.615 job2: (groupid=0, jobs=1): err= 0: pid=311903: Fri Dec 6 17:44:55 2024 00:29:15.615 read: IOPS=1035, BW=259MiB/s (271MB/s)(2594MiB/10018msec) 00:29:15.615 slat (usec): min=11, max=82003, avg=953.05, stdev=3387.88 00:29:15.615 clat (msec): min=15, max=353, avg=60.78, stdev=41.21 00:29:15.615 lat (msec): min=15, max=362, avg=61.73, stdev=41.81 00:29:15.615 clat percentiles (msec): 00:29:15.615 | 1.00th=[ 34], 5.00th=[ 39], 10.00th=[ 41], 20.00th=[ 44], 00:29:15.615 | 30.00th=[ 46], 40.00th=[ 48], 50.00th=[ 50], 60.00th=[ 52], 00:29:15.615 | 70.00th=[ 55], 80.00th=[ 58], 90.00th=[ 85], 95.00th=[ 140], 00:29:15.615 | 99.00th=[ 262], 99.50th=[ 300], 99.90th=[ 330], 99.95th=[ 334], 00:29:15.615 | 99.99th=[ 342] 00:29:15.615 bw ( KiB/s): min=64512, max=394752, per=36.98%, avg=263916.25, stdev=96400.49, samples=20 00:29:15.615 iops : min= 252, max= 1542, avg=1030.90, stdev=376.54, samples=20 00:29:15.615 lat (msec) : 20=0.13%, 50=52.80%, 100=38.13%, 250=7.53%, 500=1.41% 00:29:15.615 cpu : usr=0.75%, sys=3.47%, ctx=1392, majf=0, minf=4097 00:29:15.615 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.3%, >=64=99.4% 00:29:15.615 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:29:15.615 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.1%, >=64=0.0% 00:29:15.615 issued rwts: total=10375,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:29:15.615 latency : target=0, window=0, percentile=100.00%, depth=64 00:29:15.615 job3: (groupid=0, jobs=1): err= 0: pid=311904: Fri Dec 6 17:44:55 2024 00:29:15.615 read: IOPS=564, BW=141MiB/s (148MB/s)(1430MiB/10133msec) 00:29:15.615 slat (usec): min=9, max=302328, avg=1655.14, stdev=8655.97 00:29:15.615 clat (usec): min=1630, max=892081, avg=111635.24, stdev=132509.90 00:29:15.615 lat (usec): min=1710, max=892123, avg=113290.38, stdev=134203.60 00:29:15.615 clat percentiles (msec): 00:29:15.615 | 1.00th=[ 8], 5.00th=[ 40], 10.00th=[ 44], 20.00th=[ 47], 00:29:15.615 | 30.00th=[ 50], 40.00th=[ 51], 50.00th=[ 53], 60.00th=[ 55], 00:29:15.615 | 70.00th=[ 104], 80.00th=[ 144], 90.00th=[ 275], 95.00th=[ 401], 00:29:15.615 | 99.00th=[ 701], 99.50th=[ 793], 99.90th=[ 869], 99.95th=[ 877], 00:29:15.615 | 99.99th=[ 894] 00:29:15.615 bw ( KiB/s): min=15360, max=333824, per=20.28%, avg=144762.25, stdev=115220.26, samples=20 00:29:15.615 iops : min= 60, max= 1304, avg=565.40, stdev=450.15, samples=20 00:29:15.615 lat (msec) : 2=0.05%, 4=0.12%, 10=1.00%, 20=1.42%, 50=34.16% 00:29:15.615 lat (msec) : 100=32.49%, 250=18.29%, 500=9.51%, 750=2.34%, 1000=0.61% 00:29:15.615 cpu : usr=0.29%, sys=1.97%, ctx=1044, majf=0, minf=4097 00:29:15.615 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.3%, 32=0.6%, >=64=98.9% 00:29:15.615 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:29:15.615 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.1%, >=64=0.0% 00:29:15.615 issued rwts: total=5718,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:29:15.615 latency : target=0, window=0, percentile=100.00%, depth=64 00:29:15.615 job4: (groupid=0, jobs=1): err= 0: pid=311905: Fri Dec 6 17:44:55 2024 00:29:15.615 read: IOPS=79, BW=19.9MiB/s (20.9MB/s)(202MiB/10133msec) 00:29:15.615 slat (usec): min=9, max=471182, avg=9295.72, stdev=42923.90 00:29:15.615 clat (msec): min=22, max=1620, avg=793.51, stdev=482.92 00:29:15.615 lat (msec): min=22, max=1736, avg=802.81, stdev=489.51 00:29:15.615 clat percentiles (msec): 00:29:15.615 | 1.00th=[ 30], 5.00th=[ 39], 10.00th=[ 53], 20.00th=[ 236], 00:29:15.615 | 30.00th=[ 506], 40.00th=[ 634], 50.00th=[ 877], 60.00th=[ 1062], 00:29:15.615 | 70.00th=[ 1234], 80.00th=[ 1284], 90.00th=[ 1334], 95.00th=[ 1385], 00:29:15.615 | 99.00th=[ 1519], 99.50th=[ 1552], 99.90th=[ 1620], 99.95th=[ 1620], 00:29:15.615 | 99.99th=[ 1620] 00:29:15.615 bw ( KiB/s): min= 4608, max=64000, per=2.66%, avg=19016.75, stdev=12880.98, samples=20 00:29:15.615 iops : min= 18, max= 250, avg=74.20, stdev=50.31, samples=20 00:29:15.615 lat (msec) : 50=9.05%, 100=6.69%, 250=5.08%, 500=7.93%, 750=15.49% 00:29:15.615 lat (msec) : 1000=11.77%, 2000=43.99% 00:29:15.615 cpu : usr=0.00%, sys=0.38%, ctx=211, majf=0, minf=4097 00:29:15.615 IO depths : 1=0.1%, 2=0.2%, 4=0.5%, 8=1.0%, 16=2.0%, 32=4.0%, >=64=92.2% 00:29:15.615 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:29:15.615 complete : 0=0.0%, 4=99.9%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.1%, >=64=0.0% 00:29:15.615 issued rwts: total=807,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:29:15.615 latency : target=0, window=0, percentile=100.00%, depth=64 00:29:15.615 job5: (groupid=0, jobs=1): err= 0: pid=311906: Fri Dec 6 17:44:55 2024 00:29:15.615 read: IOPS=54, BW=13.7MiB/s (14.4MB/s)(140MiB/10221msec) 00:29:15.615 slat (usec): min=13, max=1021.9k, avg=18012.80, stdev=77937.00 00:29:15.615 clat (msec): min=129, max=2297, avg=1148.84, stdev=295.68 00:29:15.615 lat (msec): min=249, max=2297, avg=1166.85, stdev=303.12 00:29:15.615 clat percentiles (msec): 00:29:15.615 | 1.00th=[ 249], 5.00th=[ 397], 10.00th=[ 785], 20.00th=[ 1011], 00:29:15.615 | 30.00th=[ 1083], 40.00th=[ 1167], 50.00th=[ 1200], 60.00th=[ 1267], 00:29:15.615 | 70.00th=[ 1301], 80.00th=[ 1385], 90.00th=[ 1418], 95.00th=[ 1536], 00:29:15.616 | 99.00th=[ 1552], 99.50th=[ 2165], 99.90th=[ 2299], 99.95th=[ 2299], 00:29:15.616 | 99.99th=[ 2299] 00:29:15.616 bw ( KiB/s): min= 6656, max=23040, per=2.09%, avg=14937.06, stdev=4956.06, samples=17 00:29:15.616 iops : min= 26, max= 90, avg=58.29, stdev=19.41, samples=17 00:29:15.616 lat (msec) : 250=1.43%, 500=3.75%, 750=3.04%, 1000=11.43%, 2000=79.82% 00:29:15.616 lat (msec) : >=2000=0.54% 00:29:15.616 cpu : usr=0.03%, sys=0.32%, ctx=67, majf=0, minf=4097 00:29:15.616 IO depths : 1=0.2%, 2=0.4%, 4=0.7%, 8=1.4%, 16=2.9%, 32=5.7%, >=64=88.8% 00:29:15.616 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:29:15.616 complete : 0=0.0%, 4=99.8%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.2%, >=64=0.0% 00:29:15.616 issued rwts: total=560,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:29:15.616 latency : target=0, window=0, percentile=100.00%, depth=64 00:29:15.616 job6: (groupid=0, jobs=1): err= 0: pid=311909: Fri Dec 6 17:44:55 2024 00:29:15.616 read: IOPS=62, BW=15.7MiB/s (16.5MB/s)(160MiB/10190msec) 00:29:15.616 slat (usec): min=8, max=774073, avg=13521.60, stdev=68212.91 00:29:15.616 clat (msec): min=32, max=1984, avg=1002.91, stdev=596.56 00:29:15.616 lat (msec): min=34, max=2067, avg=1016.43, stdev=604.20 00:29:15.616 clat percentiles (msec): 00:29:15.616 | 1.00th=[ 56], 5.00th=[ 84], 10.00th=[ 93], 20.00th=[ 109], 00:29:15.616 | 30.00th=[ 651], 40.00th=[ 1062], 50.00th=[ 1183], 60.00th=[ 1301], 00:29:15.616 | 70.00th=[ 1401], 80.00th=[ 1552], 90.00th=[ 1720], 95.00th=[ 1754], 00:29:15.616 | 99.00th=[ 1888], 99.50th=[ 1888], 99.90th=[ 1989], 99.95th=[ 1989], 00:29:15.616 | 99.99th=[ 1989] 00:29:15.616 bw ( KiB/s): min= 4608, max=82944, per=2.07%, avg=14767.85, stdev=16908.62, samples=20 00:29:15.616 iops : min= 18, max= 324, avg=57.60, stdev=66.04, samples=20 00:29:15.616 lat (msec) : 50=0.31%, 100=14.82%, 250=8.42%, 750=11.39%, 1000=3.90% 00:29:15.616 lat (msec) : 2000=61.15% 00:29:15.616 cpu : usr=0.04%, sys=0.27%, ctx=120, majf=0, minf=4097 00:29:15.616 IO depths : 1=0.2%, 2=0.3%, 4=0.6%, 8=1.2%, 16=2.5%, 32=5.0%, >=64=90.2% 00:29:15.616 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:29:15.616 complete : 0=0.0%, 4=99.8%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.2%, >=64=0.0% 00:29:15.616 issued rwts: total=641,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:29:15.616 latency : target=0, window=0, percentile=100.00%, depth=64 00:29:15.616 job7: (groupid=0, jobs=1): err= 0: pid=311910: Fri Dec 6 17:44:55 2024 00:29:15.616 read: IOPS=62, BW=15.6MiB/s (16.3MB/s)(159MiB/10185msec) 00:29:15.616 slat (usec): min=13, max=678480, avg=15202.48, stdev=58641.50 00:29:15.616 clat (msec): min=113, max=1804, avg=1011.76, stdev=413.96 00:29:15.616 lat (msec): min=113, max=1865, avg=1026.96, stdev=420.43 00:29:15.616 clat percentiles (msec): 00:29:15.616 | 1.00th=[ 113], 5.00th=[ 157], 10.00th=[ 232], 20.00th=[ 584], 00:29:15.616 | 30.00th=[ 961], 40.00th=[ 1062], 50.00th=[ 1133], 60.00th=[ 1217], 00:29:15.616 | 70.00th=[ 1250], 80.00th=[ 1318], 90.00th=[ 1385], 95.00th=[ 1485], 00:29:15.616 | 99.00th=[ 1603], 99.50th=[ 1787], 99.90th=[ 1804], 99.95th=[ 1804], 00:29:15.616 | 99.99th=[ 1804] 00:29:15.616 bw ( KiB/s): min= 6656, max=32768, per=2.04%, avg=14590.30, stdev=7079.29, samples=20 00:29:15.616 iops : min= 26, max= 128, avg=56.90, stdev=27.63, samples=20 00:29:15.616 lat (msec) : 250=14.67%, 750=8.04%, 1000=9.15%, 2000=68.14% 00:29:15.616 cpu : usr=0.04%, sys=0.33%, ctx=92, majf=0, minf=4097 00:29:15.616 IO depths : 1=0.2%, 2=0.3%, 4=0.6%, 8=1.3%, 16=2.5%, 32=5.0%, >=64=90.1% 00:29:15.616 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:29:15.616 complete : 0=0.0%, 4=99.8%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.2%, >=64=0.0% 00:29:15.616 issued rwts: total=634,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:29:15.616 latency : target=0, window=0, percentile=100.00%, depth=64 00:29:15.616 job8: (groupid=0, jobs=1): err= 0: pid=311911: Fri Dec 6 17:44:55 2024 00:29:15.616 read: IOPS=58, BW=14.6MiB/s (15.4MB/s)(150MiB/10224msec) 00:29:15.616 slat (usec): min=14, max=633262, avg=16732.58, stdev=60569.00 00:29:15.616 clat (msec): min=37, max=1863, avg=1074.34, stdev=371.22 00:29:15.616 lat (msec): min=37, max=1863, avg=1091.07, stdev=376.66 00:29:15.616 clat percentiles (msec): 00:29:15.616 | 1.00th=[ 253], 5.00th=[ 397], 10.00th=[ 472], 20.00th=[ 751], 00:29:15.616 | 30.00th=[ 844], 40.00th=[ 1036], 50.00th=[ 1167], 60.00th=[ 1250], 00:29:15.616 | 70.00th=[ 1301], 80.00th=[ 1368], 90.00th=[ 1469], 95.00th=[ 1586], 00:29:15.616 | 99.00th=[ 1871], 99.50th=[ 1871], 99.90th=[ 1871], 99.95th=[ 1871], 00:29:15.616 | 99.99th=[ 1871] 00:29:15.616 bw ( KiB/s): min= 3584, max=20992, per=1.92%, avg=13692.20, stdev=5512.80, samples=20 00:29:15.616 iops : min= 14, max= 82, avg=53.40, stdev=21.45, samples=20 00:29:15.616 lat (msec) : 50=0.67%, 250=0.17%, 500=9.52%, 750=6.84%, 1000=20.53% 00:29:15.616 lat (msec) : 2000=62.27% 00:29:15.616 cpu : usr=0.03%, sys=0.33%, ctx=82, majf=0, minf=4097 00:29:15.616 IO depths : 1=0.2%, 2=0.3%, 4=0.7%, 8=1.3%, 16=2.7%, 32=5.3%, >=64=89.5% 00:29:15.616 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:29:15.616 complete : 0=0.0%, 4=99.8%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.2%, >=64=0.0% 00:29:15.616 issued rwts: total=599,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:29:15.616 latency : target=0, window=0, percentile=100.00%, depth=64 00:29:15.616 job9: (groupid=0, jobs=1): err= 0: pid=311916: Fri Dec 6 17:44:55 2024 00:29:15.616 read: IOPS=60, BW=15.1MiB/s (15.9MB/s)(154MiB/10178msec) 00:29:15.616 slat (usec): min=14, max=565696, avg=16268.07, stdev=53468.08 00:29:15.616 clat (msec): min=37, max=1758, avg=1040.05, stdev=334.41 00:29:15.616 lat (msec): min=37, max=1758, avg=1056.32, stdev=339.93 00:29:15.616 clat percentiles (msec): 00:29:15.616 | 1.00th=[ 176], 5.00th=[ 351], 10.00th=[ 542], 20.00th=[ 751], 00:29:15.616 | 30.00th=[ 885], 40.00th=[ 1003], 50.00th=[ 1133], 60.00th=[ 1217], 00:29:15.616 | 70.00th=[ 1267], 80.00th=[ 1318], 90.00th=[ 1418], 95.00th=[ 1435], 00:29:15.616 | 99.00th=[ 1502], 99.50th=[ 1603], 99.90th=[ 1754], 99.95th=[ 1754], 00:29:15.616 | 99.99th=[ 1754] 00:29:15.616 bw ( KiB/s): min= 7680, max=19968, per=1.98%, avg=14127.60, stdev=3907.51, samples=20 00:29:15.616 iops : min= 30, max= 78, avg=55.10, stdev=15.18, samples=20 00:29:15.616 lat (msec) : 50=0.65%, 250=2.44%, 500=5.52%, 750=10.71%, 1000=19.16% 00:29:15.616 lat (msec) : 2000=61.53% 00:29:15.616 cpu : usr=0.07%, sys=0.30%, ctx=89, majf=0, minf=4097 00:29:15.616 IO depths : 1=0.2%, 2=0.3%, 4=0.6%, 8=1.3%, 16=2.6%, 32=5.2%, >=64=89.8% 00:29:15.616 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:29:15.616 complete : 0=0.0%, 4=99.8%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.2%, >=64=0.0% 00:29:15.616 issued rwts: total=616,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:29:15.616 latency : target=0, window=0, percentile=100.00%, depth=64 00:29:15.616 job10: (groupid=0, jobs=1): err= 0: pid=311919: Fri Dec 6 17:44:55 2024 00:29:15.616 read: IOPS=277, BW=69.4MiB/s (72.8MB/s)(703MiB/10125msec) 00:29:15.616 slat (usec): min=9, max=153321, avg=3503.61, stdev=12899.21 00:29:15.616 clat (msec): min=13, max=627, avg=226.88, stdev=101.53 00:29:15.616 lat (msec): min=13, max=627, avg=230.39, stdev=102.83 00:29:15.616 clat percentiles (msec): 00:29:15.616 | 1.00th=[ 46], 5.00th=[ 115], 10.00th=[ 127], 20.00th=[ 142], 00:29:15.616 | 30.00th=[ 155], 40.00th=[ 171], 50.00th=[ 199], 60.00th=[ 241], 00:29:15.616 | 70.00th=[ 271], 80.00th=[ 313], 90.00th=[ 376], 95.00th=[ 426], 00:29:15.616 | 99.00th=[ 510], 99.50th=[ 550], 99.90th=[ 609], 99.95th=[ 609], 00:29:15.616 | 99.99th=[ 625] 00:29:15.616 bw ( KiB/s): min=31232, max=128000, per=9.85%, avg=70283.05, stdev=26563.39, samples=20 00:29:15.616 iops : min= 122, max= 500, avg=274.50, stdev=103.78, samples=20 00:29:15.616 lat (msec) : 20=0.07%, 50=1.32%, 100=0.68%, 250=60.71%, 500=35.55% 00:29:15.616 lat (msec) : 750=1.67% 00:29:15.616 cpu : usr=0.16%, sys=1.05%, ctx=401, majf=0, minf=3722 00:29:15.616 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.3%, 16=0.6%, 32=1.1%, >=64=97.8% 00:29:15.616 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:29:15.616 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.1%, >=64=0.0% 00:29:15.616 issued rwts: total=2810,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:29:15.616 latency : target=0, window=0, percentile=100.00%, depth=64 00:29:15.616 00:29:15.616 Run status group 0 (all jobs): 00:29:15.616 READ: bw=697MiB/s (731MB/s), 13.7MiB/s-259MiB/s (14.4MB/s-271MB/s), io=7126MiB (7472MB), run=10018-10224msec 00:29:15.616 00:29:15.616 Disk stats (read/write): 00:29:15.616 nvme0n1: ios=5468/0, merge=0/0, ticks=1233788/0, in_queue=1233788, util=97.28% 00:29:15.616 nvme10n1: ios=5685/0, merge=0/0, ticks=1238305/0, in_queue=1238305, util=97.50% 00:29:15.616 nvme1n1: ios=20459/0, merge=0/0, ticks=1245305/0, in_queue=1245305, util=97.73% 00:29:15.616 nvme2n1: ios=11301/0, merge=0/0, ticks=1177341/0, in_queue=1177341, util=97.87% 00:29:15.616 nvme3n1: ios=1486/0, merge=0/0, ticks=1191116/0, in_queue=1191116, util=97.95% 00:29:15.616 nvme4n1: ios=1119/0, merge=0/0, ticks=1270578/0, in_queue=1270578, util=98.28% 00:29:15.616 nvme5n1: ios=1155/0, merge=0/0, ticks=1167210/0, in_queue=1167210, util=98.40% 00:29:15.616 nvme6n1: ios=1128/0, merge=0/0, ticks=1186808/0, in_queue=1186808, util=98.51% 00:29:15.616 nvme7n1: ios=1197/0, merge=0/0, ticks=1265831/0, in_queue=1265831, util=98.94% 00:29:15.616 nvme8n1: ios=1104/0, merge=0/0, ticks=1195858/0, in_queue=1195858, util=99.10% 00:29:15.616 nvme9n1: ios=5464/0, merge=0/0, ticks=1232809/0, in_queue=1232809, util=99.24% 00:29:15.616 17:44:55 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 262144 -d 64 -t randwrite -r 10 00:29:15.616 [global] 00:29:15.616 thread=1 00:29:15.616 invalidate=1 00:29:15.616 rw=randwrite 00:29:15.616 time_based=1 00:29:15.616 runtime=10 00:29:15.616 ioengine=libaio 00:29:15.616 direct=1 00:29:15.616 bs=262144 00:29:15.616 iodepth=64 00:29:15.616 norandommap=1 00:29:15.616 numjobs=1 00:29:15.616 00:29:15.616 [job0] 00:29:15.616 filename=/dev/nvme0n1 00:29:15.616 [job1] 00:29:15.616 filename=/dev/nvme10n1 00:29:15.616 [job2] 00:29:15.616 filename=/dev/nvme1n1 00:29:15.616 [job3] 00:29:15.616 filename=/dev/nvme2n1 00:29:15.616 [job4] 00:29:15.616 filename=/dev/nvme3n1 00:29:15.616 [job5] 00:29:15.616 filename=/dev/nvme4n1 00:29:15.616 [job6] 00:29:15.616 filename=/dev/nvme5n1 00:29:15.616 [job7] 00:29:15.616 filename=/dev/nvme6n1 00:29:15.616 [job8] 00:29:15.616 filename=/dev/nvme7n1 00:29:15.616 [job9] 00:29:15.616 filename=/dev/nvme8n1 00:29:15.616 [job10] 00:29:15.616 filename=/dev/nvme9n1 00:29:15.616 Could not set queue depth (nvme0n1) 00:29:15.616 Could not set queue depth (nvme10n1) 00:29:15.617 Could not set queue depth (nvme1n1) 00:29:15.617 Could not set queue depth (nvme2n1) 00:29:15.617 Could not set queue depth (nvme3n1) 00:29:15.617 Could not set queue depth (nvme4n1) 00:29:15.617 Could not set queue depth (nvme5n1) 00:29:15.617 Could not set queue depth (nvme6n1) 00:29:15.617 Could not set queue depth (nvme7n1) 00:29:15.617 Could not set queue depth (nvme8n1) 00:29:15.617 Could not set queue depth (nvme9n1) 00:29:15.617 job0: (g=0): rw=randwrite, bs=(R) 256KiB-256KiB, (W) 256KiB-256KiB, (T) 256KiB-256KiB, ioengine=libaio, iodepth=64 00:29:15.617 job1: (g=0): rw=randwrite, bs=(R) 256KiB-256KiB, (W) 256KiB-256KiB, (T) 256KiB-256KiB, ioengine=libaio, iodepth=64 00:29:15.617 job2: (g=0): rw=randwrite, bs=(R) 256KiB-256KiB, (W) 256KiB-256KiB, (T) 256KiB-256KiB, ioengine=libaio, iodepth=64 00:29:15.617 job3: (g=0): rw=randwrite, bs=(R) 256KiB-256KiB, (W) 256KiB-256KiB, (T) 256KiB-256KiB, ioengine=libaio, iodepth=64 00:29:15.617 job4: (g=0): rw=randwrite, bs=(R) 256KiB-256KiB, (W) 256KiB-256KiB, (T) 256KiB-256KiB, ioengine=libaio, iodepth=64 00:29:15.617 job5: (g=0): rw=randwrite, bs=(R) 256KiB-256KiB, (W) 256KiB-256KiB, (T) 256KiB-256KiB, ioengine=libaio, iodepth=64 00:29:15.617 job6: (g=0): rw=randwrite, bs=(R) 256KiB-256KiB, (W) 256KiB-256KiB, (T) 256KiB-256KiB, ioengine=libaio, iodepth=64 00:29:15.617 job7: (g=0): rw=randwrite, bs=(R) 256KiB-256KiB, (W) 256KiB-256KiB, (T) 256KiB-256KiB, ioengine=libaio, iodepth=64 00:29:15.617 job8: (g=0): rw=randwrite, bs=(R) 256KiB-256KiB, (W) 256KiB-256KiB, (T) 256KiB-256KiB, ioengine=libaio, iodepth=64 00:29:15.617 job9: (g=0): rw=randwrite, bs=(R) 256KiB-256KiB, (W) 256KiB-256KiB, (T) 256KiB-256KiB, ioengine=libaio, iodepth=64 00:29:15.617 job10: (g=0): rw=randwrite, bs=(R) 256KiB-256KiB, (W) 256KiB-256KiB, (T) 256KiB-256KiB, ioengine=libaio, iodepth=64 00:29:15.617 fio-3.35 00:29:15.617 Starting 11 threads 00:29:25.580 00:29:25.580 job0: (groupid=0, jobs=1): err= 0: pid=312492: Fri Dec 6 17:45:06 2024 00:29:25.580 write: IOPS=484, BW=121MiB/s (127MB/s)(1224MiB/10106msec); 0 zone resets 00:29:25.580 slat (usec): min=15, max=129189, avg=1425.90, stdev=4893.46 00:29:25.580 clat (usec): min=640, max=631852, avg=130634.01, stdev=119329.65 00:29:25.580 lat (usec): min=673, max=631963, avg=132059.91, stdev=120402.96 00:29:25.580 clat percentiles (msec): 00:29:25.580 | 1.00th=[ 3], 5.00th=[ 13], 10.00th=[ 25], 20.00th=[ 52], 00:29:25.580 | 30.00th=[ 54], 40.00th=[ 55], 50.00th=[ 72], 60.00th=[ 133], 00:29:25.580 | 70.00th=[ 176], 80.00th=[ 201], 90.00th=[ 288], 95.00th=[ 380], 00:29:25.580 | 99.00th=[ 550], 99.50th=[ 567], 99.90th=[ 617], 99.95th=[ 625], 00:29:25.581 | 99.99th=[ 634] 00:29:25.581 bw ( KiB/s): min=28672, max=305564, per=14.64%, avg=123636.75, stdev=74606.47, samples=20 00:29:25.581 iops : min= 112, max= 1193, avg=482.90, stdev=291.38, samples=20 00:29:25.581 lat (usec) : 750=0.04%, 1000=0.16% 00:29:25.581 lat (msec) : 2=0.49%, 4=0.80%, 10=2.29%, 20=4.62%, 50=8.97% 00:29:25.581 lat (msec) : 100=37.96%, 250=31.43%, 500=10.85%, 750=2.39% 00:29:25.581 cpu : usr=1.26%, sys=1.55%, ctx=2457, majf=0, minf=1 00:29:25.581 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.2%, 16=0.3%, 32=0.7%, >=64=98.7% 00:29:25.581 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:29:25.581 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.1%, >=64=0.0% 00:29:25.581 issued rwts: total=0,4894,0,0 short=0,0,0,0 dropped=0,0,0,0 00:29:25.581 latency : target=0, window=0, percentile=100.00%, depth=64 00:29:25.581 job1: (groupid=0, jobs=1): err= 0: pid=312504: Fri Dec 6 17:45:06 2024 00:29:25.581 write: IOPS=247, BW=61.8MiB/s (64.8MB/s)(625MiB/10112msec); 0 zone resets 00:29:25.581 slat (usec): min=19, max=193633, avg=2877.45, stdev=9844.71 00:29:25.581 clat (msec): min=2, max=788, avg=255.78, stdev=170.68 00:29:25.581 lat (msec): min=2, max=795, avg=258.65, stdev=172.76 00:29:25.581 clat percentiles (msec): 00:29:25.581 | 1.00th=[ 6], 5.00th=[ 36], 10.00th=[ 59], 20.00th=[ 101], 00:29:25.581 | 30.00th=[ 161], 40.00th=[ 236], 50.00th=[ 255], 60.00th=[ 266], 00:29:25.581 | 70.00th=[ 284], 80.00th=[ 313], 90.00th=[ 510], 95.00th=[ 684], 00:29:25.581 | 99.00th=[ 751], 99.50th=[ 760], 99.90th=[ 776], 99.95th=[ 785], 00:29:25.581 | 99.99th=[ 793] 00:29:25.581 bw ( KiB/s): min=20480, max=144384, per=7.39%, avg=62419.60, stdev=28560.09, samples=20 00:29:25.581 iops : min= 80, max= 564, avg=243.80, stdev=111.56, samples=20 00:29:25.581 lat (msec) : 4=0.36%, 10=1.32%, 20=1.40%, 50=5.12%, 100=11.84% 00:29:25.581 lat (msec) : 250=26.59%, 500=43.14%, 750=9.04%, 1000=1.20% 00:29:25.581 cpu : usr=0.75%, sys=0.80%, ctx=1467, majf=0, minf=2 00:29:25.581 IO depths : 1=0.1%, 2=0.1%, 4=0.2%, 8=0.3%, 16=0.6%, 32=1.3%, >=64=97.5% 00:29:25.581 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:29:25.581 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.1%, >=64=0.0% 00:29:25.581 issued rwts: total=0,2501,0,0 short=0,0,0,0 dropped=0,0,0,0 00:29:25.581 latency : target=0, window=0, percentile=100.00%, depth=64 00:29:25.581 job2: (groupid=0, jobs=1): err= 0: pid=312505: Fri Dec 6 17:45:06 2024 00:29:25.581 write: IOPS=222, BW=55.6MiB/s (58.3MB/s)(565MiB/10150msec); 0 zone resets 00:29:25.581 slat (usec): min=26, max=133999, avg=4007.83, stdev=9012.17 00:29:25.581 clat (msec): min=25, max=748, avg=283.41, stdev=136.65 00:29:25.581 lat (msec): min=25, max=748, avg=287.42, stdev=138.28 00:29:25.581 clat percentiles (msec): 00:29:25.581 | 1.00th=[ 62], 5.00th=[ 106], 10.00th=[ 140], 20.00th=[ 186], 00:29:25.581 | 30.00th=[ 222], 40.00th=[ 251], 50.00th=[ 259], 60.00th=[ 275], 00:29:25.581 | 70.00th=[ 292], 80.00th=[ 326], 90.00th=[ 510], 95.00th=[ 592], 00:29:25.581 | 99.00th=[ 676], 99.50th=[ 701], 99.90th=[ 743], 99.95th=[ 743], 00:29:25.581 | 99.99th=[ 751] 00:29:25.581 bw ( KiB/s): min=23040, max=106496, per=6.66%, avg=56222.30, stdev=23338.59, samples=20 00:29:25.581 iops : min= 90, max= 416, avg=219.60, stdev=91.17, samples=20 00:29:25.581 lat (msec) : 50=0.35%, 100=4.38%, 250=35.33%, 500=48.83%, 750=11.11% 00:29:25.581 cpu : usr=0.62%, sys=0.76%, ctx=769, majf=0, minf=1 00:29:25.581 IO depths : 1=0.1%, 2=0.1%, 4=0.2%, 8=0.4%, 16=0.7%, 32=1.4%, >=64=97.2% 00:29:25.581 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:29:25.581 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.1%, >=64=0.0% 00:29:25.581 issued rwts: total=0,2259,0,0 short=0,0,0,0 dropped=0,0,0,0 00:29:25.581 latency : target=0, window=0, percentile=100.00%, depth=64 00:29:25.581 job3: (groupid=0, jobs=1): err= 0: pid=312506: Fri Dec 6 17:45:06 2024 00:29:25.581 write: IOPS=224, BW=56.1MiB/s (58.8MB/s)(567MiB/10112msec); 0 zone resets 00:29:25.581 slat (usec): min=24, max=82750, avg=3977.83, stdev=8768.16 00:29:25.581 clat (usec): min=1294, max=729695, avg=281024.23, stdev=130831.28 00:29:25.581 lat (usec): min=1429, max=729737, avg=285002.05, stdev=132687.59 00:29:25.581 clat percentiles (msec): 00:29:25.581 | 1.00th=[ 3], 5.00th=[ 68], 10.00th=[ 165], 20.00th=[ 211], 00:29:25.581 | 30.00th=[ 239], 40.00th=[ 253], 50.00th=[ 262], 60.00th=[ 275], 00:29:25.581 | 70.00th=[ 292], 80.00th=[ 321], 90.00th=[ 481], 95.00th=[ 558], 00:29:25.581 | 99.00th=[ 684], 99.50th=[ 709], 99.90th=[ 726], 99.95th=[ 726], 00:29:25.581 | 99.99th=[ 726] 00:29:25.581 bw ( KiB/s): min=24576, max=96768, per=6.69%, avg=56477.45, stdev=20367.83, samples=20 00:29:25.581 iops : min= 96, max= 378, avg=220.60, stdev=79.58, samples=20 00:29:25.581 lat (msec) : 2=0.97%, 4=2.07%, 10=1.37%, 20=0.09%, 50=0.22% 00:29:25.581 lat (msec) : 100=1.59%, 250=30.37%, 500=54.34%, 750=8.99% 00:29:25.581 cpu : usr=0.74%, sys=0.69%, ctx=808, majf=0, minf=1 00:29:25.581 IO depths : 1=0.1%, 2=0.1%, 4=0.2%, 8=0.4%, 16=0.7%, 32=1.4%, >=64=97.2% 00:29:25.581 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:29:25.581 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.1%, >=64=0.0% 00:29:25.581 issued rwts: total=0,2269,0,0 short=0,0,0,0 dropped=0,0,0,0 00:29:25.581 latency : target=0, window=0, percentile=100.00%, depth=64 00:29:25.581 job4: (groupid=0, jobs=1): err= 0: pid=312507: Fri Dec 6 17:45:06 2024 00:29:25.581 write: IOPS=289, BW=72.4MiB/s (75.9MB/s)(736MiB/10164msec); 0 zone resets 00:29:25.581 slat (usec): min=22, max=115626, avg=2588.48, stdev=7372.79 00:29:25.581 clat (msec): min=8, max=672, avg=218.18, stdev=125.56 00:29:25.581 lat (msec): min=8, max=672, avg=220.76, stdev=127.06 00:29:25.581 clat percentiles (msec): 00:29:25.581 | 1.00th=[ 16], 5.00th=[ 47], 10.00th=[ 78], 20.00th=[ 136], 00:29:25.581 | 30.00th=[ 153], 40.00th=[ 169], 50.00th=[ 190], 60.00th=[ 218], 00:29:25.581 | 70.00th=[ 264], 80.00th=[ 296], 90.00th=[ 363], 95.00th=[ 485], 00:29:25.581 | 99.00th=[ 642], 99.50th=[ 659], 99.90th=[ 676], 99.95th=[ 676], 00:29:25.581 | 99.99th=[ 676] 00:29:25.581 bw ( KiB/s): min=24576, max=139776, per=8.73%, avg=73735.55, stdev=30951.42, samples=20 00:29:25.581 iops : min= 96, max= 546, avg=288.00, stdev=120.90, samples=20 00:29:25.581 lat (msec) : 10=0.44%, 20=1.26%, 50=3.53%, 100=7.91%, 250=54.59% 00:29:25.581 lat (msec) : 500=27.62%, 750=4.65% 00:29:25.581 cpu : usr=0.80%, sys=0.84%, ctx=1369, majf=0, minf=1 00:29:25.581 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.3%, 16=0.5%, 32=1.1%, >=64=97.9% 00:29:25.581 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:29:25.581 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.1%, >=64=0.0% 00:29:25.581 issued rwts: total=0,2944,0,0 short=0,0,0,0 dropped=0,0,0,0 00:29:25.581 latency : target=0, window=0, percentile=100.00%, depth=64 00:29:25.581 job5: (groupid=0, jobs=1): err= 0: pid=312508: Fri Dec 6 17:45:06 2024 00:29:25.581 write: IOPS=306, BW=76.6MiB/s (80.3MB/s)(773MiB/10092msec); 0 zone resets 00:29:25.581 slat (usec): min=17, max=132327, avg=1744.71, stdev=6692.07 00:29:25.581 clat (usec): min=664, max=774236, avg=206085.22, stdev=177128.05 00:29:25.581 lat (usec): min=698, max=774292, avg=207829.93, stdev=178609.92 00:29:25.581 clat percentiles (msec): 00:29:25.581 | 1.00th=[ 3], 5.00th=[ 11], 10.00th=[ 22], 20.00th=[ 52], 00:29:25.581 | 30.00th=[ 77], 40.00th=[ 106], 50.00th=[ 144], 60.00th=[ 205], 00:29:25.581 | 70.00th=[ 262], 80.00th=[ 388], 90.00th=[ 498], 95.00th=[ 542], 00:29:25.581 | 99.00th=[ 634], 99.50th=[ 684], 99.90th=[ 751], 99.95th=[ 768], 00:29:25.581 | 99.99th=[ 776] 00:29:25.581 bw ( KiB/s): min=28672, max=159744, per=9.19%, avg=77580.55, stdev=34436.85, samples=20 00:29:25.581 iops : min= 112, max= 624, avg=303.00, stdev=134.45, samples=20 00:29:25.581 lat (usec) : 750=0.03%, 1000=0.16% 00:29:25.581 lat (msec) : 2=0.29%, 4=2.04%, 10=1.91%, 20=4.91%, 50=10.57% 00:29:25.581 lat (msec) : 100=18.20%, 250=30.23%, 500=21.99%, 750=9.54%, 1000=0.13% 00:29:25.581 cpu : usr=0.86%, sys=1.00%, ctx=2023, majf=0, minf=1 00:29:25.581 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.3%, 16=0.5%, 32=1.0%, >=64=98.0% 00:29:25.581 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:29:25.581 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.1%, >=64=0.0% 00:29:25.581 issued rwts: total=0,3093,0,0 short=0,0,0,0 dropped=0,0,0,0 00:29:25.581 latency : target=0, window=0, percentile=100.00%, depth=64 00:29:25.581 job6: (groupid=0, jobs=1): err= 0: pid=312509: Fri Dec 6 17:45:06 2024 00:29:25.581 write: IOPS=301, BW=75.3MiB/s (78.9MB/s)(765MiB/10164msec); 0 zone resets 00:29:25.581 slat (usec): min=16, max=138250, avg=1835.29, stdev=6790.52 00:29:25.581 clat (usec): min=788, max=675403, avg=210655.89, stdev=156566.02 00:29:25.581 lat (usec): min=844, max=675466, avg=212491.18, stdev=158319.49 00:29:25.581 clat percentiles (msec): 00:29:25.581 | 1.00th=[ 3], 5.00th=[ 6], 10.00th=[ 12], 20.00th=[ 56], 00:29:25.581 | 30.00th=[ 110], 40.00th=[ 150], 50.00th=[ 186], 60.00th=[ 218], 00:29:25.581 | 70.00th=[ 284], 80.00th=[ 351], 90.00th=[ 426], 95.00th=[ 527], 00:29:25.581 | 99.00th=[ 625], 99.50th=[ 634], 99.90th=[ 667], 99.95th=[ 676], 00:29:25.581 | 99.99th=[ 676] 00:29:25.581 bw ( KiB/s): min=25600, max=152576, per=9.08%, avg=76702.10, stdev=33508.99, samples=20 00:29:25.581 iops : min= 100, max= 596, avg=299.60, stdev=130.91, samples=20 00:29:25.581 lat (usec) : 1000=0.07% 00:29:25.581 lat (msec) : 2=0.69%, 4=2.32%, 10=6.11%, 20=1.99%, 50=8.07% 00:29:25.581 lat (msec) : 100=8.86%, 250=36.93%, 500=28.53%, 750=6.44% 00:29:25.581 cpu : usr=0.82%, sys=1.10%, ctx=2184, majf=0, minf=1 00:29:25.581 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.3%, 16=0.5%, 32=1.0%, >=64=97.9% 00:29:25.581 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:29:25.581 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.1%, >=64=0.0% 00:29:25.581 issued rwts: total=0,3060,0,0 short=0,0,0,0 dropped=0,0,0,0 00:29:25.581 latency : target=0, window=0, percentile=100.00%, depth=64 00:29:25.581 job7: (groupid=0, jobs=1): err= 0: pid=312510: Fri Dec 6 17:45:06 2024 00:29:25.581 write: IOPS=212, BW=53.2MiB/s (55.8MB/s)(541MiB/10171msec); 0 zone resets 00:29:25.581 slat (usec): min=18, max=72707, avg=3722.23, stdev=8761.39 00:29:25.581 clat (msec): min=11, max=668, avg=296.80, stdev=134.61 00:29:25.581 lat (msec): min=11, max=678, avg=300.53, stdev=136.48 00:29:25.581 clat percentiles (msec): 00:29:25.581 | 1.00th=[ 47], 5.00th=[ 115], 10.00th=[ 142], 20.00th=[ 169], 00:29:25.581 | 30.00th=[ 197], 40.00th=[ 232], 50.00th=[ 275], 60.00th=[ 338], 00:29:25.581 | 70.00th=[ 380], 80.00th=[ 414], 90.00th=[ 489], 95.00th=[ 535], 00:29:25.581 | 99.00th=[ 609], 99.50th=[ 634], 99.90th=[ 659], 99.95th=[ 667], 00:29:25.581 | 99.99th=[ 667] 00:29:25.581 bw ( KiB/s): min=30720, max=109568, per=6.37%, avg=53790.10, stdev=21255.22, samples=20 00:29:25.581 iops : min= 120, max= 428, avg=210.10, stdev=83.04, samples=20 00:29:25.581 lat (msec) : 20=0.42%, 50=0.69%, 100=0.74%, 250=43.37%, 500=46.19% 00:29:25.581 lat (msec) : 750=8.59% 00:29:25.581 cpu : usr=0.64%, sys=0.59%, ctx=956, majf=0, minf=1 00:29:25.581 IO depths : 1=0.1%, 2=0.1%, 4=0.2%, 8=0.4%, 16=0.7%, 32=1.5%, >=64=97.1% 00:29:25.581 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:29:25.581 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.1%, >=64=0.0% 00:29:25.581 issued rwts: total=0,2165,0,0 short=0,0,0,0 dropped=0,0,0,0 00:29:25.581 latency : target=0, window=0, percentile=100.00%, depth=64 00:29:25.582 job8: (groupid=0, jobs=1): err= 0: pid=312511: Fri Dec 6 17:45:06 2024 00:29:25.582 write: IOPS=327, BW=81.9MiB/s (85.9MB/s)(833MiB/10173msec); 0 zone resets 00:29:25.582 slat (usec): min=17, max=97278, avg=2203.18, stdev=6539.20 00:29:25.582 clat (usec): min=1132, max=724439, avg=193065.81, stdev=158825.47 00:29:25.582 lat (usec): min=1336, max=724475, avg=195269.00, stdev=160420.75 00:29:25.582 clat percentiles (msec): 00:29:25.582 | 1.00th=[ 4], 5.00th=[ 12], 10.00th=[ 32], 20.00th=[ 68], 00:29:25.582 | 30.00th=[ 72], 40.00th=[ 105], 50.00th=[ 120], 60.00th=[ 194], 00:29:25.582 | 70.00th=[ 271], 80.00th=[ 363], 90.00th=[ 409], 95.00th=[ 498], 00:29:25.582 | 99.00th=[ 659], 99.50th=[ 676], 99.90th=[ 709], 99.95th=[ 718], 00:29:25.582 | 99.99th=[ 726] 00:29:25.582 bw ( KiB/s): min=30720, max=207872, per=9.91%, avg=83660.80, stdev=61572.66, samples=20 00:29:25.582 iops : min= 120, max= 812, avg=326.80, stdev=240.52, samples=20 00:29:25.582 lat (msec) : 2=0.21%, 4=1.29%, 10=2.37%, 20=4.35%, 50=5.10% 00:29:25.582 lat (msec) : 100=25.27%, 250=28.57%, 500=28.03%, 750=4.80% 00:29:25.582 cpu : usr=0.85%, sys=1.12%, ctx=1679, majf=0, minf=1 00:29:25.582 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.2%, 16=0.5%, 32=1.0%, >=64=98.1% 00:29:25.582 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:29:25.582 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.1%, >=64=0.0% 00:29:25.582 issued rwts: total=0,3332,0,0 short=0,0,0,0 dropped=0,0,0,0 00:29:25.582 latency : target=0, window=0, percentile=100.00%, depth=64 00:29:25.582 job9: (groupid=0, jobs=1): err= 0: pid=312512: Fri Dec 6 17:45:06 2024 00:29:25.582 write: IOPS=473, BW=118MiB/s (124MB/s)(1203MiB/10159msec); 0 zone resets 00:29:25.582 slat (usec): min=23, max=63719, avg=2031.98, stdev=4678.69 00:29:25.582 clat (msec): min=24, max=498, avg=132.99, stdev=90.81 00:29:25.582 lat (msec): min=24, max=498, avg=135.02, stdev=92.05 00:29:25.582 clat percentiles (msec): 00:29:25.582 | 1.00th=[ 41], 5.00th=[ 44], 10.00th=[ 50], 20.00th=[ 56], 00:29:25.582 | 30.00th=[ 58], 40.00th=[ 92], 50.00th=[ 113], 60.00th=[ 136], 00:29:25.582 | 70.00th=[ 155], 80.00th=[ 190], 90.00th=[ 279], 95.00th=[ 317], 00:29:25.582 | 99.00th=[ 430], 99.50th=[ 464], 99.90th=[ 498], 99.95th=[ 498], 00:29:25.582 | 99.99th=[ 498] 00:29:25.582 bw ( KiB/s): min=34816, max=288256, per=14.40%, avg=121580.10, stdev=72707.07, samples=20 00:29:25.582 iops : min= 136, max= 1126, avg=474.85, stdev=284.08, samples=20 00:29:25.582 lat (msec) : 50=11.26%, 100=31.46%, 250=44.98%, 500=12.30% 00:29:25.582 cpu : usr=1.17%, sys=1.47%, ctx=1289, majf=0, minf=1 00:29:25.582 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.2%, 16=0.3%, 32=0.7%, >=64=98.7% 00:29:25.582 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:29:25.582 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.1%, >=64=0.0% 00:29:25.582 issued rwts: total=0,4813,0,0 short=0,0,0,0 dropped=0,0,0,0 00:29:25.582 latency : target=0, window=0, percentile=100.00%, depth=64 00:29:25.582 job10: (groupid=0, jobs=1): err= 0: pid=312513: Fri Dec 6 17:45:06 2024 00:29:25.582 write: IOPS=220, BW=55.1MiB/s (57.8MB/s)(557MiB/10107msec); 0 zone resets 00:29:25.582 slat (usec): min=25, max=280016, avg=4178.36, stdev=12275.23 00:29:25.582 clat (msec): min=35, max=705, avg=285.49, stdev=126.68 00:29:25.582 lat (msec): min=35, max=705, avg=289.67, stdev=127.98 00:29:25.582 clat percentiles (msec): 00:29:25.582 | 1.00th=[ 66], 5.00th=[ 122], 10.00th=[ 167], 20.00th=[ 213], 00:29:25.582 | 30.00th=[ 236], 40.00th=[ 251], 50.00th=[ 259], 60.00th=[ 268], 00:29:25.582 | 70.00th=[ 284], 80.00th=[ 305], 90.00th=[ 523], 95.00th=[ 592], 00:29:25.582 | 99.00th=[ 667], 99.50th=[ 693], 99.90th=[ 709], 99.95th=[ 709], 00:29:25.582 | 99.99th=[ 709] 00:29:25.582 bw ( KiB/s): min=22528, max=80896, per=6.56%, avg=55391.30, stdev=16626.00, samples=20 00:29:25.582 iops : min= 88, max= 316, avg=216.30, stdev=64.91, samples=20 00:29:25.582 lat (msec) : 50=0.67%, 100=2.33%, 250=36.91%, 500=48.81%, 750=11.27% 00:29:25.582 cpu : usr=0.62%, sys=0.72%, ctx=658, majf=0, minf=1 00:29:25.582 IO depths : 1=0.1%, 2=0.1%, 4=0.2%, 8=0.4%, 16=0.7%, 32=1.4%, >=64=97.2% 00:29:25.582 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:29:25.582 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.1%, >=64=0.0% 00:29:25.582 issued rwts: total=0,2227,0,0 short=0,0,0,0 dropped=0,0,0,0 00:29:25.582 latency : target=0, window=0, percentile=100.00%, depth=64 00:29:25.582 00:29:25.582 Run status group 0 (all jobs): 00:29:25.582 WRITE: bw=825MiB/s (865MB/s), 53.2MiB/s-121MiB/s (55.8MB/s-127MB/s), io=8389MiB (8797MB), run=10092-10173msec 00:29:25.582 00:29:25.582 Disk stats (read/write): 00:29:25.582 nvme0n1: ios=47/9619, merge=0/0, ticks=2353/1203335, in_queue=1205688, util=100.00% 00:29:25.582 nvme10n1: ios=42/4816, merge=0/0, ticks=80/1215341, in_queue=1215421, util=98.04% 00:29:25.582 nvme1n1: ios=0/4332, merge=0/0, ticks=0/1210850, in_queue=1210850, util=97.71% 00:29:25.582 nvme2n1: ios=49/4351, merge=0/0, ticks=1559/1203846, in_queue=1205405, util=100.00% 00:29:25.582 nvme3n1: ios=47/5736, merge=0/0, ticks=2431/1199616, in_queue=1202047, util=100.00% 00:29:25.582 nvme4n1: ios=44/6020, merge=0/0, ticks=2106/1214525, in_queue=1216631, util=100.00% 00:29:25.582 nvme5n1: ios=0/5966, merge=0/0, ticks=0/1218844, in_queue=1218844, util=98.39% 00:29:25.582 nvme6n1: ios=0/4180, merge=0/0, ticks=0/1210121, in_queue=1210121, util=98.49% 00:29:25.582 nvme7n1: ios=40/6513, merge=0/0, ticks=1747/1212215, in_queue=1213962, util=99.84% 00:29:25.582 nvme8n1: ios=0/9473, merge=0/0, ticks=0/1197888, in_queue=1197888, util=98.95% 00:29:25.582 nvme9n1: ios=44/4260, merge=0/0, ticks=4494/1142835, in_queue=1147329, util=100.00% 00:29:25.582 17:45:06 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@36 -- # sync 00:29:25.582 17:45:06 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@37 -- # seq 1 11 00:29:25.582 17:45:06 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@37 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:29:25.582 17:45:06 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@38 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:29:25.582 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:29:25.582 17:45:06 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@39 -- # waitforserial_disconnect SPDK1 00:29:25.582 17:45:06 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1223 -- # local i=0 00:29:25.582 17:45:06 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:29:25.582 17:45:06 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1224 -- # grep -q -w SPDK1 00:29:25.582 17:45:06 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:29:25.582 17:45:06 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1231 -- # grep -q -w SPDK1 00:29:25.582 17:45:06 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1235 -- # return 0 00:29:25.582 17:45:06 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@40 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:29:25.582 17:45:06 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:25.582 17:45:06 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:29:25.582 17:45:06 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:25.582 17:45:06 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@37 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:29:25.582 17:45:06 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@38 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode2 00:29:25.582 NQN:nqn.2016-06.io.spdk:cnode2 disconnected 1 controller(s) 00:29:25.582 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@39 -- # waitforserial_disconnect SPDK2 00:29:25.582 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1223 -- # local i=0 00:29:25.582 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:29:25.582 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1224 -- # grep -q -w SPDK2 00:29:25.582 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:29:25.582 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1231 -- # grep -q -w SPDK2 00:29:25.582 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1235 -- # return 0 00:29:25.582 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@40 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode2 00:29:25.582 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:25.582 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:29:25.582 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:25.582 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@37 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:29:25.582 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@38 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode3 00:29:25.846 NQN:nqn.2016-06.io.spdk:cnode3 disconnected 1 controller(s) 00:29:25.846 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@39 -- # waitforserial_disconnect SPDK3 00:29:25.846 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1223 -- # local i=0 00:29:25.846 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:29:25.846 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1224 -- # grep -q -w SPDK3 00:29:25.846 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:29:25.846 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1231 -- # grep -q -w SPDK3 00:29:25.846 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1235 -- # return 0 00:29:25.846 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@40 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode3 00:29:25.846 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:25.846 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:29:25.846 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:25.846 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@37 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:29:25.846 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@38 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode4 00:29:26.103 NQN:nqn.2016-06.io.spdk:cnode4 disconnected 1 controller(s) 00:29:26.103 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@39 -- # waitforserial_disconnect SPDK4 00:29:26.103 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1223 -- # local i=0 00:29:26.103 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:29:26.103 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1224 -- # grep -q -w SPDK4 00:29:26.103 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:29:26.103 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1231 -- # grep -q -w SPDK4 00:29:26.103 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1235 -- # return 0 00:29:26.103 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@40 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode4 00:29:26.103 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:26.103 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:29:26.103 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:26.103 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@37 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:29:26.103 17:45:07 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@38 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode5 00:29:26.360 NQN:nqn.2016-06.io.spdk:cnode5 disconnected 1 controller(s) 00:29:26.360 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@39 -- # waitforserial_disconnect SPDK5 00:29:26.360 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1223 -- # local i=0 00:29:26.360 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:29:26.360 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1224 -- # grep -q -w SPDK5 00:29:26.360 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:29:26.360 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1231 -- # grep -q -w SPDK5 00:29:26.360 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1235 -- # return 0 00:29:26.360 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@40 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode5 00:29:26.360 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:26.360 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:29:26.360 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:26.360 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@37 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:29:26.360 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@38 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode6 00:29:26.617 NQN:nqn.2016-06.io.spdk:cnode6 disconnected 1 controller(s) 00:29:26.617 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@39 -- # waitforserial_disconnect SPDK6 00:29:26.617 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1223 -- # local i=0 00:29:26.617 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:29:26.617 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1224 -- # grep -q -w SPDK6 00:29:26.617 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:29:26.617 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1231 -- # grep -q -w SPDK6 00:29:26.617 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1235 -- # return 0 00:29:26.617 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@40 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode6 00:29:26.617 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:26.617 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:29:26.617 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:26.617 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@37 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:29:26.617 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@38 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode7 00:29:26.909 NQN:nqn.2016-06.io.spdk:cnode7 disconnected 1 controller(s) 00:29:26.909 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@39 -- # waitforserial_disconnect SPDK7 00:29:26.909 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1223 -- # local i=0 00:29:26.909 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:29:26.910 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1224 -- # grep -q -w SPDK7 00:29:26.910 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:29:26.910 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1231 -- # grep -q -w SPDK7 00:29:26.910 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1235 -- # return 0 00:29:26.910 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@40 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode7 00:29:26.910 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:26.910 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:29:26.910 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:26.910 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@37 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:29:26.910 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@38 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode8 00:29:26.910 NQN:nqn.2016-06.io.spdk:cnode8 disconnected 1 controller(s) 00:29:26.910 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@39 -- # waitforserial_disconnect SPDK8 00:29:26.910 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1223 -- # local i=0 00:29:26.910 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:29:26.910 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1224 -- # grep -q -w SPDK8 00:29:26.910 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:29:26.910 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1231 -- # grep -q -w SPDK8 00:29:26.910 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1235 -- # return 0 00:29:26.910 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@40 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode8 00:29:26.910 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:26.910 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:29:26.910 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:26.910 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@37 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:29:26.910 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@38 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode9 00:29:27.220 NQN:nqn.2016-06.io.spdk:cnode9 disconnected 1 controller(s) 00:29:27.220 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@39 -- # waitforserial_disconnect SPDK9 00:29:27.220 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1223 -- # local i=0 00:29:27.220 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:29:27.220 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1224 -- # grep -q -w SPDK9 00:29:27.220 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:29:27.220 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1231 -- # grep -q -w SPDK9 00:29:27.220 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1235 -- # return 0 00:29:27.220 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@40 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode9 00:29:27.220 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:27.220 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:29:27.220 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:27.220 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@37 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:29:27.220 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@38 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode10 00:29:27.220 NQN:nqn.2016-06.io.spdk:cnode10 disconnected 1 controller(s) 00:29:27.220 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@39 -- # waitforserial_disconnect SPDK10 00:29:27.220 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1223 -- # local i=0 00:29:27.220 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:29:27.220 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1224 -- # grep -q -w SPDK10 00:29:27.220 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:29:27.220 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1231 -- # grep -q -w SPDK10 00:29:27.220 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1235 -- # return 0 00:29:27.220 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@40 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode10 00:29:27.220 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:27.220 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:29:27.220 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:27.220 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@37 -- # for i in $(seq 1 $NVMF_SUBSYS) 00:29:27.220 17:45:08 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@38 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode11 00:29:27.530 NQN:nqn.2016-06.io.spdk:cnode11 disconnected 1 controller(s) 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@39 -- # waitforserial_disconnect SPDK11 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1223 -- # local i=0 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1224 -- # grep -q -w SPDK11 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1231 -- # grep -q -w SPDK11 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1235 -- # return 0 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@40 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode11 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@43 -- # rm -f ./local-job0-0-verify.state 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@45 -- # trap - SIGINT SIGTERM EXIT 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- target/multiconnection.sh@47 -- # nvmftestfini 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@516 -- # nvmfcleanup 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@121 -- # sync 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@124 -- # set +e 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@125 -- # for i in {1..20} 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:29:27.530 rmmod nvme_tcp 00:29:27.530 rmmod nvme_fabrics 00:29:27.530 rmmod nvme_keyring 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@128 -- # set -e 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@129 -- # return 0 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@517 -- # '[' -n 307639 ']' 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@518 -- # killprocess 307639 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@954 -- # '[' -z 307639 ']' 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@958 -- # kill -0 307639 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@959 -- # uname 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 307639 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@972 -- # echo 'killing process with pid 307639' 00:29:27.530 killing process with pid 307639 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@973 -- # kill 307639 00:29:27.530 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@978 -- # wait 307639 00:29:27.870 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:29:27.870 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:29:27.870 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:29:27.870 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@297 -- # iptr 00:29:27.870 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@791 -- # iptables-save 00:29:27.870 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@791 -- # iptables-restore 00:29:27.870 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:29:27.870 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:29:27.870 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@302 -- # remove_spdk_ns 00:29:27.870 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:29:27.870 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:29:27.870 17:45:09 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:29:30.445 00:29:30.445 real 1m0.972s 00:29:30.445 user 3m35.622s 00:29:30.445 sys 0m14.818s 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@1130 -- # xtrace_disable 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_multiconnection -- common/autotest_common.sh@10 -- # set +x 00:29:30.445 ************************************ 00:29:30.445 END TEST nvmf_multiconnection 00:29:30.445 ************************************ 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@50 -- # run_test nvmf_initiator_timeout /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/initiator_timeout.sh --transport=tcp 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:29:30.445 ************************************ 00:29:30.445 START TEST nvmf_initiator_timeout 00:29:30.445 ************************************ 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/initiator_timeout.sh --transport=tcp 00:29:30.445 * Looking for test storage... 00:29:30.445 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@1711 -- # lcov --version 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@333 -- # local ver1 ver1_l 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@334 -- # local ver2 ver2_l 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@336 -- # IFS=.-: 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@336 -- # read -ra ver1 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@337 -- # IFS=.-: 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@337 -- # read -ra ver2 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@338 -- # local 'op=<' 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@340 -- # ver1_l=2 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@341 -- # ver2_l=1 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@344 -- # case "$op" in 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@345 -- # : 1 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@364 -- # (( v = 0 )) 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@365 -- # decimal 1 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@353 -- # local d=1 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@355 -- # echo 1 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@365 -- # ver1[v]=1 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@366 -- # decimal 2 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@353 -- # local d=2 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@355 -- # echo 2 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@366 -- # ver2[v]=2 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@368 -- # return 0 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:29:30.445 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:29:30.445 --rc genhtml_branch_coverage=1 00:29:30.445 --rc genhtml_function_coverage=1 00:29:30.445 --rc genhtml_legend=1 00:29:30.445 --rc geninfo_all_blocks=1 00:29:30.445 --rc geninfo_unexecuted_blocks=1 00:29:30.445 00:29:30.445 ' 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:29:30.445 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:29:30.445 --rc genhtml_branch_coverage=1 00:29:30.445 --rc genhtml_function_coverage=1 00:29:30.445 --rc genhtml_legend=1 00:29:30.445 --rc geninfo_all_blocks=1 00:29:30.445 --rc geninfo_unexecuted_blocks=1 00:29:30.445 00:29:30.445 ' 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:29:30.445 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:29:30.445 --rc genhtml_branch_coverage=1 00:29:30.445 --rc genhtml_function_coverage=1 00:29:30.445 --rc genhtml_legend=1 00:29:30.445 --rc geninfo_all_blocks=1 00:29:30.445 --rc geninfo_unexecuted_blocks=1 00:29:30.445 00:29:30.445 ' 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:29:30.445 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:29:30.445 --rc genhtml_branch_coverage=1 00:29:30.445 --rc genhtml_function_coverage=1 00:29:30.445 --rc genhtml_legend=1 00:29:30.445 --rc geninfo_all_blocks=1 00:29:30.445 --rc geninfo_unexecuted_blocks=1 00:29:30.445 00:29:30.445 ' 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@7 -- # uname -s 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:29:30.445 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@15 -- # shopt -s extglob 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- paths/export.sh@5 -- # export PATH 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@51 -- # : 0 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:29:30.446 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@55 -- # have_pci_nics=0 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@11 -- # MALLOC_BDEV_SIZE=64 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@14 -- # nvmftestinit 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@476 -- # prepare_net_devs 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@438 -- # local -g is_hw=no 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@440 -- # remove_spdk_ns 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@309 -- # xtrace_disable 00:29:30.446 17:45:11 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@10 -- # set +x 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@315 -- # pci_devs=() 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@315 -- # local -a pci_devs 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@316 -- # pci_net_devs=() 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@317 -- # pci_drivers=() 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@317 -- # local -A pci_drivers 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@319 -- # net_devs=() 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@319 -- # local -ga net_devs 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@320 -- # e810=() 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@320 -- # local -ga e810 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@321 -- # x722=() 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@321 -- # local -ga x722 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@322 -- # mlx=() 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@322 -- # local -ga mlx 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:29:32.346 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:29:32.346 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@418 -- # [[ up == up ]] 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:29:32.346 Found net devices under 0000:0a:00.0: cvl_0_0 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@418 -- # [[ up == up ]] 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:29:32.346 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:29:32.346 Found net devices under 0000:0a:00.1: cvl_0_1 00:29:32.347 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:29:32.347 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:29:32.347 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@442 -- # is_hw=yes 00:29:32.347 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:29:32.347 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:29:32.347 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:29:32.347 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:29:32.347 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:29:32.347 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:29:32.347 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:29:32.347 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:29:32.347 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:29:32.347 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:29:32.347 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:29:32.347 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:29:32.347 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:29:32.347 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:29:32.347 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:29:32.347 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:29:32.347 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:29:32.347 17:45:13 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:29:32.347 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:29:32.347 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.188 ms 00:29:32.347 00:29:32.347 --- 10.0.0.2 ping statistics --- 00:29:32.347 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:29:32.347 rtt min/avg/max/mdev = 0.188/0.188/0.188/0.000 ms 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:29:32.347 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:29:32.347 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.123 ms 00:29:32.347 00:29:32.347 --- 10.0.0.1 ping statistics --- 00:29:32.347 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:29:32.347 rtt min/avg/max/mdev = 0.123/0.123/0.123/0.000 ms 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@450 -- # return 0 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@15 -- # nvmfappstart -m 0xF 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@726 -- # xtrace_disable 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@10 -- # set +x 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@509 -- # nvmfpid=316332 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@510 -- # waitforlisten 316332 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@835 -- # '[' -z 316332 ']' 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@840 -- # local max_retries=100 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:32.347 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@844 -- # xtrace_disable 00:29:32.347 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@10 -- # set +x 00:29:32.347 [2024-12-06 17:45:14.171523] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:29:32.347 [2024-12-06 17:45:14.171595] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:29:32.612 [2024-12-06 17:45:14.247976] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:29:32.612 [2024-12-06 17:45:14.294048] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:29:32.612 [2024-12-06 17:45:14.294100] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:29:32.612 [2024-12-06 17:45:14.294131] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:29:32.612 [2024-12-06 17:45:14.294142] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:29:32.612 [2024-12-06 17:45:14.294152] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:29:32.612 [2024-12-06 17:45:14.295756] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:29:32.612 [2024-12-06 17:45:14.295842] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:29:32.612 [2024-12-06 17:45:14.295846] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:29:32.612 [2024-12-06 17:45:14.295780] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:29:32.612 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:29:32.612 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@868 -- # return 0 00:29:32.612 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:29:32.612 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@732 -- # xtrace_disable 00:29:32.612 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@10 -- # set +x 00:29:32.870 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:29:32.870 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@17 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; killprocess $nvmfpid; nvmftestfini $1; exit 1' SIGINT SIGTERM EXIT 00:29:32.870 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:29:32.870 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:32.870 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@10 -- # set +x 00:29:32.870 Malloc0 00:29:32.870 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:32.870 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@22 -- # rpc_cmd bdev_delay_create -b Malloc0 -d Delay0 -r 30 -t 30 -w 30 -n 30 00:29:32.870 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:32.870 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@10 -- # set +x 00:29:32.870 Delay0 00:29:32.870 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:32.870 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@24 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:29:32.870 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:32.870 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@10 -- # set +x 00:29:32.870 [2024-12-06 17:45:14.504448] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:29:32.870 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:32.870 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@25 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:29:32.870 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:32.870 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@10 -- # set +x 00:29:32.870 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:32.870 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@26 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:29:32.870 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:32.870 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@10 -- # set +x 00:29:32.870 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:32.870 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:29:32.870 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:32.870 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@10 -- # set +x 00:29:32.870 [2024-12-06 17:45:14.532730] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:29:32.870 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:32.870 17:45:14 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@29 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:29:33.445 17:45:15 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@31 -- # waitforserial SPDKISFASTANDAWESOME 00:29:33.445 17:45:15 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@1202 -- # local i=0 00:29:33.445 17:45:15 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:29:33.445 17:45:15 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:29:33.445 17:45:15 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@1209 -- # sleep 2 00:29:35.344 17:45:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:29:35.344 17:45:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:29:35.344 17:45:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:29:35.344 17:45:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:29:35.344 17:45:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:29:35.344 17:45:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@1212 -- # return 0 00:29:35.344 17:45:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@35 -- # fio_pid=316753 00:29:35.344 17:45:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t write -r 60 -v 00:29:35.344 17:45:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@37 -- # sleep 3 00:29:35.344 [global] 00:29:35.344 thread=1 00:29:35.344 invalidate=1 00:29:35.344 rw=write 00:29:35.344 time_based=1 00:29:35.344 runtime=60 00:29:35.344 ioengine=libaio 00:29:35.344 direct=1 00:29:35.344 bs=4096 00:29:35.344 iodepth=1 00:29:35.344 norandommap=0 00:29:35.344 numjobs=1 00:29:35.344 00:29:35.344 verify_dump=1 00:29:35.344 verify_backlog=512 00:29:35.344 verify_state_save=0 00:29:35.344 do_verify=1 00:29:35.344 verify=crc32c-intel 00:29:35.344 [job0] 00:29:35.344 filename=/dev/nvme0n1 00:29:35.344 Could not set queue depth (nvme0n1) 00:29:35.600 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:29:35.600 fio-3.35 00:29:35.600 Starting 1 thread 00:29:38.874 17:45:20 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@40 -- # rpc_cmd bdev_delay_update_latency Delay0 avg_read 31000000 00:29:38.874 17:45:20 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:38.874 17:45:20 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@10 -- # set +x 00:29:38.874 true 00:29:38.874 17:45:20 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:38.874 17:45:20 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@41 -- # rpc_cmd bdev_delay_update_latency Delay0 avg_write 31000000 00:29:38.874 17:45:20 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:38.874 17:45:20 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@10 -- # set +x 00:29:38.874 true 00:29:38.874 17:45:20 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:38.874 17:45:20 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@42 -- # rpc_cmd bdev_delay_update_latency Delay0 p99_read 31000000 00:29:38.874 17:45:20 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:38.874 17:45:20 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@10 -- # set +x 00:29:38.874 true 00:29:38.874 17:45:20 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:38.874 17:45:20 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@43 -- # rpc_cmd bdev_delay_update_latency Delay0 p99_write 310000000 00:29:38.874 17:45:20 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:38.874 17:45:20 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@10 -- # set +x 00:29:38.874 true 00:29:38.874 17:45:20 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:38.874 17:45:20 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@45 -- # sleep 3 00:29:41.397 17:45:23 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@48 -- # rpc_cmd bdev_delay_update_latency Delay0 avg_read 30 00:29:41.397 17:45:23 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:41.397 17:45:23 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@10 -- # set +x 00:29:41.397 true 00:29:41.397 17:45:23 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:41.397 17:45:23 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@49 -- # rpc_cmd bdev_delay_update_latency Delay0 avg_write 30 00:29:41.397 17:45:23 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:41.397 17:45:23 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@10 -- # set +x 00:29:41.397 true 00:29:41.397 17:45:23 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:41.397 17:45:23 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@50 -- # rpc_cmd bdev_delay_update_latency Delay0 p99_read 30 00:29:41.397 17:45:23 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:41.397 17:45:23 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@10 -- # set +x 00:29:41.397 true 00:29:41.397 17:45:23 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:41.397 17:45:23 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@51 -- # rpc_cmd bdev_delay_update_latency Delay0 p99_write 30 00:29:41.398 17:45:23 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@563 -- # xtrace_disable 00:29:41.398 17:45:23 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@10 -- # set +x 00:29:41.398 true 00:29:41.398 17:45:23 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:29:41.398 17:45:23 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@53 -- # fio_status=0 00:29:41.398 17:45:23 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@54 -- # wait 316753 00:30:37.596 00:30:37.596 job0: (groupid=0, jobs=1): err= 0: pid=316822: Fri Dec 6 17:46:17 2024 00:30:37.596 read: IOPS=48, BW=196KiB/s (200kB/s)(11.5MiB/60030msec) 00:30:37.596 slat (nsec): min=4589, max=45243, avg=12561.16, stdev=7406.09 00:30:37.596 clat (usec): min=199, max=40968k, avg=20225.24, stdev=756222.42 00:30:37.596 lat (usec): min=206, max=40968k, avg=20237.80, stdev=756222.50 00:30:37.596 clat percentiles (usec): 00:30:37.596 | 1.00th=[ 208], 5.00th=[ 215], 10.00th=[ 219], 00:30:37.596 | 20.00th=[ 225], 30.00th=[ 231], 40.00th=[ 237], 00:30:37.596 | 50.00th=[ 243], 60.00th=[ 265], 70.00th=[ 273], 00:30:37.596 | 80.00th=[ 285], 90.00th=[ 41157], 95.00th=[ 41157], 00:30:37.596 | 99.00th=[ 42206], 99.50th=[ 42206], 99.90th=[ 42206], 00:30:37.596 | 99.95th=[ 43779], 99.99th=[17112761] 00:30:37.596 write: IOPS=51, BW=205KiB/s (210kB/s)(12.0MiB/60030msec); 0 zone resets 00:30:37.596 slat (nsec): min=5970, max=62910, avg=9952.14, stdev=5917.71 00:30:37.596 clat (usec): min=156, max=441, avg=189.24, stdev=22.93 00:30:37.596 lat (usec): min=164, max=456, avg=199.19, stdev=26.70 00:30:37.596 clat percentiles (usec): 00:30:37.596 | 1.00th=[ 163], 5.00th=[ 169], 10.00th=[ 172], 20.00th=[ 176], 00:30:37.596 | 30.00th=[ 180], 40.00th=[ 182], 50.00th=[ 186], 60.00th=[ 188], 00:30:37.596 | 70.00th=[ 192], 80.00th=[ 198], 90.00th=[ 206], 95.00th=[ 219], 00:30:37.596 | 99.00th=[ 306], 99.50th=[ 338], 99.90th=[ 388], 99.95th=[ 412], 00:30:37.596 | 99.99th=[ 441] 00:30:37.596 bw ( KiB/s): min= 4096, max= 9792, per=100.00%, avg=6144.00, stdev=2659.56, samples=4 00:30:37.596 iops : min= 1024, max= 2448, avg=1536.00, stdev=664.89, samples=4 00:30:37.596 lat (usec) : 250=77.39%, 500=15.37%, 750=0.05% 00:30:37.596 lat (msec) : 50=7.17%, >=2000=0.02% 00:30:37.596 cpu : usr=0.07%, sys=0.11%, ctx=6009, majf=0, minf=1 00:30:37.597 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:30:37.597 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:30:37.597 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:30:37.597 issued rwts: total=2935,3072,0,0 short=0,0,0,0 dropped=0,0,0,0 00:30:37.597 latency : target=0, window=0, percentile=100.00%, depth=1 00:30:37.597 00:30:37.597 Run status group 0 (all jobs): 00:30:37.597 READ: bw=196KiB/s (200kB/s), 196KiB/s-196KiB/s (200kB/s-200kB/s), io=11.5MiB (12.0MB), run=60030-60030msec 00:30:37.597 WRITE: bw=205KiB/s (210kB/s), 205KiB/s-205KiB/s (210kB/s-210kB/s), io=12.0MiB (12.6MB), run=60030-60030msec 00:30:37.597 00:30:37.597 Disk stats (read/write): 00:30:37.597 nvme0n1: ios=3030/3072, merge=0/0, ticks=18271/556, in_queue=18827, util=99.91% 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@56 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:30:37.597 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@57 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@1223 -- # local i=0 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@1235 -- # return 0 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@59 -- # '[' 0 -eq 0 ']' 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@60 -- # echo 'nvmf hotplug test: fio successful as expected' 00:30:37.597 nvmf hotplug test: fio successful as expected 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@67 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@10 -- # set +x 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@69 -- # rm -f ./local-job0-0-verify.state 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@71 -- # trap - SIGINT SIGTERM EXIT 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- target/initiator_timeout.sh@73 -- # nvmftestfini 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@516 -- # nvmfcleanup 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@121 -- # sync 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@124 -- # set +e 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@125 -- # for i in {1..20} 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:30:37.597 rmmod nvme_tcp 00:30:37.597 rmmod nvme_fabrics 00:30:37.597 rmmod nvme_keyring 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@128 -- # set -e 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@129 -- # return 0 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@517 -- # '[' -n 316332 ']' 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@518 -- # killprocess 316332 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@954 -- # '[' -z 316332 ']' 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@958 -- # kill -0 316332 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@959 -- # uname 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 316332 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@972 -- # echo 'killing process with pid 316332' 00:30:37.597 killing process with pid 316332 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@973 -- # kill 316332 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@978 -- # wait 316332 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@297 -- # iptr 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@791 -- # iptables-restore 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@791 -- # iptables-save 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@302 -- # remove_spdk_ns 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:30:37.597 17:46:17 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:30:38.531 17:46:20 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:30:38.531 00:30:38.531 real 1m8.277s 00:30:38.531 user 4m11.390s 00:30:38.531 sys 0m6.185s 00:30:38.531 17:46:20 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@1130 -- # xtrace_disable 00:30:38.531 17:46:20 nvmf_tcp.nvmf_target_extra.nvmf_initiator_timeout -- common/autotest_common.sh@10 -- # set +x 00:30:38.531 ************************************ 00:30:38.531 END TEST nvmf_initiator_timeout 00:30:38.531 ************************************ 00:30:38.531 17:46:20 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@53 -- # [[ phy == phy ]] 00:30:38.531 17:46:20 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@54 -- # '[' tcp = tcp ']' 00:30:38.531 17:46:20 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@55 -- # gather_supported_nvmf_pci_devs 00:30:38.531 17:46:20 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@309 -- # xtrace_disable 00:30:38.531 17:46:20 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@315 -- # pci_devs=() 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@315 -- # local -a pci_devs 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@316 -- # pci_net_devs=() 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@317 -- # pci_drivers=() 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@317 -- # local -A pci_drivers 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@319 -- # net_devs=() 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@319 -- # local -ga net_devs 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@320 -- # e810=() 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@320 -- # local -ga e810 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@321 -- # x722=() 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@321 -- # local -ga x722 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@322 -- # mlx=() 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@322 -- # local -ga mlx 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:30:40.434 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:30:40.434 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:30:40.434 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@418 -- # [[ up == up ]] 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:30:40.435 Found net devices under 0000:0a:00.0: cvl_0_0 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@418 -- # [[ up == up ]] 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:30:40.435 Found net devices under 0000:0a:00.1: cvl_0_1 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@56 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@57 -- # (( 2 > 0 )) 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@58 -- # run_test nvmf_perf_adq /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/perf_adq.sh --transport=tcp 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:30:40.435 ************************************ 00:30:40.435 START TEST nvmf_perf_adq 00:30:40.435 ************************************ 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/perf_adq.sh --transport=tcp 00:30:40.435 * Looking for test storage... 00:30:40.435 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1711 -- # lcov --version 00:30:40.435 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@333 -- # local ver1 ver1_l 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@334 -- # local ver2 ver2_l 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@336 -- # IFS=.-: 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@336 -- # read -ra ver1 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@337 -- # IFS=.-: 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@337 -- # read -ra ver2 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@338 -- # local 'op=<' 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@340 -- # ver1_l=2 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@341 -- # ver2_l=1 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@344 -- # case "$op" in 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@345 -- # : 1 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@364 -- # (( v = 0 )) 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@365 -- # decimal 1 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@353 -- # local d=1 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@355 -- # echo 1 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@365 -- # ver1[v]=1 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@366 -- # decimal 2 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@353 -- # local d=2 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@355 -- # echo 2 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@366 -- # ver2[v]=2 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@368 -- # return 0 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:30:40.696 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:40.696 --rc genhtml_branch_coverage=1 00:30:40.696 --rc genhtml_function_coverage=1 00:30:40.696 --rc genhtml_legend=1 00:30:40.696 --rc geninfo_all_blocks=1 00:30:40.696 --rc geninfo_unexecuted_blocks=1 00:30:40.696 00:30:40.696 ' 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:30:40.696 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:40.696 --rc genhtml_branch_coverage=1 00:30:40.696 --rc genhtml_function_coverage=1 00:30:40.696 --rc genhtml_legend=1 00:30:40.696 --rc geninfo_all_blocks=1 00:30:40.696 --rc geninfo_unexecuted_blocks=1 00:30:40.696 00:30:40.696 ' 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:30:40.696 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:40.696 --rc genhtml_branch_coverage=1 00:30:40.696 --rc genhtml_function_coverage=1 00:30:40.696 --rc genhtml_legend=1 00:30:40.696 --rc geninfo_all_blocks=1 00:30:40.696 --rc geninfo_unexecuted_blocks=1 00:30:40.696 00:30:40.696 ' 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:30:40.696 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:40.696 --rc genhtml_branch_coverage=1 00:30:40.696 --rc genhtml_function_coverage=1 00:30:40.696 --rc genhtml_legend=1 00:30:40.696 --rc geninfo_all_blocks=1 00:30:40.696 --rc geninfo_unexecuted_blocks=1 00:30:40.696 00:30:40.696 ' 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@7 -- # uname -s 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:30:40.696 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@15 -- # shopt -s extglob 00:30:40.697 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:30:40.697 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:30:40.697 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:30:40.697 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:40.697 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:40.697 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:40.697 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- paths/export.sh@5 -- # export PATH 00:30:40.697 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:40.697 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@51 -- # : 0 00:30:40.697 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:30:40.697 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:30:40.697 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:30:40.697 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:30:40.697 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:30:40.697 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:30:40.697 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:30:40.697 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:30:40.697 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:30:40.697 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@55 -- # have_pci_nics=0 00:30:40.697 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@11 -- # gather_supported_nvmf_pci_devs 00:30:40.697 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@309 -- # xtrace_disable 00:30:40.697 17:46:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@315 -- # pci_devs=() 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@315 -- # local -a pci_devs 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@316 -- # pci_net_devs=() 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@317 -- # pci_drivers=() 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@317 -- # local -A pci_drivers 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@319 -- # net_devs=() 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@319 -- # local -ga net_devs 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@320 -- # e810=() 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@320 -- # local -ga e810 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@321 -- # x722=() 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@321 -- # local -ga x722 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@322 -- # mlx=() 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@322 -- # local -ga mlx 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:30:43.233 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:30:43.233 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@418 -- # [[ up == up ]] 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:30:43.233 Found net devices under 0000:0a:00.0: cvl_0_0 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:30:43.233 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:30:43.234 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:30:43.234 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:30:43.234 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:30:43.234 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@418 -- # [[ up == up ]] 00:30:43.234 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:30:43.234 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:30:43.234 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:30:43.234 Found net devices under 0000:0a:00.1: cvl_0_1 00:30:43.234 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:30:43.234 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:30:43.234 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@12 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:30:43.234 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@13 -- # (( 2 == 0 )) 00:30:43.234 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@18 -- # perf=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf 00:30:43.234 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@68 -- # adq_reload_driver 00:30:43.234 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@58 -- # modprobe -a sch_mqprio 00:30:43.234 17:46:24 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@61 -- # rmmod ice 00:30:43.493 17:46:25 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@62 -- # modprobe ice 00:30:47.675 17:46:28 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@63 -- # sleep 5 00:30:52.944 17:46:33 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@76 -- # nvmftestinit 00:30:52.944 17:46:33 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:30:52.944 17:46:33 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:30:52.944 17:46:33 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@476 -- # prepare_net_devs 00:30:52.944 17:46:33 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@438 -- # local -g is_hw=no 00:30:52.944 17:46:33 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@440 -- # remove_spdk_ns 00:30:52.944 17:46:33 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:30:52.944 17:46:33 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:30:52.944 17:46:33 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:30:52.944 17:46:33 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:30:52.944 17:46:33 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:30:52.944 17:46:33 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@309 -- # xtrace_disable 00:30:52.944 17:46:33 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:30:52.944 17:46:33 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:30:52.944 17:46:33 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@315 -- # pci_devs=() 00:30:52.944 17:46:33 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@315 -- # local -a pci_devs 00:30:52.944 17:46:33 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@316 -- # pci_net_devs=() 00:30:52.944 17:46:33 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:30:52.944 17:46:33 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@317 -- # pci_drivers=() 00:30:52.944 17:46:33 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@317 -- # local -A pci_drivers 00:30:52.944 17:46:33 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@319 -- # net_devs=() 00:30:52.944 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@319 -- # local -ga net_devs 00:30:52.944 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@320 -- # e810=() 00:30:52.944 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@320 -- # local -ga e810 00:30:52.944 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@321 -- # x722=() 00:30:52.944 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@321 -- # local -ga x722 00:30:52.944 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@322 -- # mlx=() 00:30:52.944 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@322 -- # local -ga mlx 00:30:52.944 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:30:52.944 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:30:52.944 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:30:52.944 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:30:52.944 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:30:52.944 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:30:52.944 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:30:52.944 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:30:52.944 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:30:52.944 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:30:52.944 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:30:52.944 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:30:52.944 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:30:52.944 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:30:52.944 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:30:52.944 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:30:52.944 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:30:52.944 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:30:52.944 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:30:52.944 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:30:52.944 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:30:52.945 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@418 -- # [[ up == up ]] 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:30:52.945 Found net devices under 0000:0a:00.0: cvl_0_0 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@418 -- # [[ up == up ]] 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:30:52.945 Found net devices under 0000:0a:00.1: cvl_0_1 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@442 -- # is_hw=yes 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:30:52.945 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:30:52.945 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.207 ms 00:30:52.945 00:30:52.945 --- 10.0.0.2 ping statistics --- 00:30:52.945 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:30:52.945 rtt min/avg/max/mdev = 0.207/0.207/0.207/0.000 ms 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:30:52.945 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:30:52.945 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.131 ms 00:30:52.945 00:30:52.945 --- 10.0.0.1 ping statistics --- 00:30:52.945 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:30:52.945 rtt min/avg/max/mdev = 0.131/0.131/0.131/0.000 ms 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@450 -- # return 0 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@77 -- # nvmfappstart -m 0xF --wait-for-rpc 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:30:52.945 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@726 -- # xtrace_disable 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@509 -- # nvmfpid=328619 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF --wait-for-rpc 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@510 -- # waitforlisten 328619 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@835 -- # '[' -z 328619 ']' 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@840 -- # local max_retries=100 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:30:52.946 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@844 -- # xtrace_disable 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:30:52.946 [2024-12-06 17:46:34.240722] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:30:52.946 [2024-12-06 17:46:34.240805] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:30:52.946 [2024-12-06 17:46:34.314084] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:30:52.946 [2024-12-06 17:46:34.361134] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:30:52.946 [2024-12-06 17:46:34.361192] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:30:52.946 [2024-12-06 17:46:34.361222] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:30:52.946 [2024-12-06 17:46:34.361233] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:30:52.946 [2024-12-06 17:46:34.361243] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:30:52.946 [2024-12-06 17:46:34.362846] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:30:52.946 [2024-12-06 17:46:34.362874] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:30:52.946 [2024-12-06 17:46:34.362898] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:30:52.946 [2024-12-06 17:46:34.362901] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@868 -- # return 0 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@732 -- # xtrace_disable 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@78 -- # adq_configure_nvmf_target 0 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@42 -- # rpc_cmd sock_get_default_impl 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@42 -- # jq -r .impl_name 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@42 -- # socket_impl=posix 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@43 -- # rpc_cmd sock_impl_set_options --enable-placement-id 0 --enable-zerocopy-send-server -i posix 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@44 -- # rpc_cmd framework_start_init 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@45 -- # rpc_cmd nvmf_create_transport -t tcp -o --io-unit-size 8192 --sock-priority 0 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:30:52.946 [2024-12-06 17:46:34.650456] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@46 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:30:52.946 Malloc1 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@47 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@48 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@49 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:30:52.946 [2024-12-06 17:46:34.714279] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@82 -- # perfpid=328719 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@83 -- # sleep 2 00:30:52.946 17:46:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 64 -o 4096 -w randread -t 10 -c 0xF0 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:30:55.477 17:46:36 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@85 -- # rpc_cmd nvmf_get_stats 00:30:55.477 17:46:36 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:30:55.477 17:46:36 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:30:55.477 17:46:36 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:30:55.477 17:46:36 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@85 -- # nvmf_stats='{ 00:30:55.477 "tick_rate": 2700000000, 00:30:55.477 "poll_groups": [ 00:30:55.477 { 00:30:55.477 "name": "nvmf_tgt_poll_group_000", 00:30:55.477 "admin_qpairs": 1, 00:30:55.477 "io_qpairs": 1, 00:30:55.477 "current_admin_qpairs": 1, 00:30:55.477 "current_io_qpairs": 1, 00:30:55.477 "pending_bdev_io": 0, 00:30:55.477 "completed_nvme_io": 18613, 00:30:55.477 "transports": [ 00:30:55.477 { 00:30:55.477 "trtype": "TCP" 00:30:55.477 } 00:30:55.477 ] 00:30:55.477 }, 00:30:55.477 { 00:30:55.477 "name": "nvmf_tgt_poll_group_001", 00:30:55.477 "admin_qpairs": 0, 00:30:55.477 "io_qpairs": 1, 00:30:55.477 "current_admin_qpairs": 0, 00:30:55.477 "current_io_qpairs": 1, 00:30:55.477 "pending_bdev_io": 0, 00:30:55.477 "completed_nvme_io": 20301, 00:30:55.477 "transports": [ 00:30:55.477 { 00:30:55.477 "trtype": "TCP" 00:30:55.477 } 00:30:55.477 ] 00:30:55.477 }, 00:30:55.477 { 00:30:55.477 "name": "nvmf_tgt_poll_group_002", 00:30:55.477 "admin_qpairs": 0, 00:30:55.477 "io_qpairs": 1, 00:30:55.477 "current_admin_qpairs": 0, 00:30:55.477 "current_io_qpairs": 1, 00:30:55.477 "pending_bdev_io": 0, 00:30:55.477 "completed_nvme_io": 20225, 00:30:55.477 "transports": [ 00:30:55.477 { 00:30:55.477 "trtype": "TCP" 00:30:55.477 } 00:30:55.477 ] 00:30:55.477 }, 00:30:55.477 { 00:30:55.477 "name": "nvmf_tgt_poll_group_003", 00:30:55.477 "admin_qpairs": 0, 00:30:55.477 "io_qpairs": 1, 00:30:55.477 "current_admin_qpairs": 0, 00:30:55.477 "current_io_qpairs": 1, 00:30:55.477 "pending_bdev_io": 0, 00:30:55.477 "completed_nvme_io": 20390, 00:30:55.477 "transports": [ 00:30:55.477 { 00:30:55.477 "trtype": "TCP" 00:30:55.477 } 00:30:55.478 ] 00:30:55.478 } 00:30:55.478 ] 00:30:55.478 }' 00:30:55.478 17:46:36 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@86 -- # jq -r '.poll_groups[] | select(.current_io_qpairs == 1) | length' 00:30:55.478 17:46:36 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@86 -- # wc -l 00:30:55.478 17:46:36 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@86 -- # count=4 00:30:55.478 17:46:36 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@87 -- # [[ 4 -ne 4 ]] 00:30:55.478 17:46:36 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@91 -- # wait 328719 00:31:03.579 Initializing NVMe Controllers 00:31:03.579 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:31:03.579 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 4 00:31:03.579 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 5 00:31:03.579 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 6 00:31:03.579 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 7 00:31:03.579 Initialization complete. Launching workers. 00:31:03.579 ======================================================== 00:31:03.579 Latency(us) 00:31:03.579 Device Information : IOPS MiB/s Average min max 00:31:03.579 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 4: 10596.10 41.39 6040.84 2599.97 10244.59 00:31:03.579 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 5: 10768.20 42.06 5943.89 2655.42 9531.88 00:31:03.579 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 6: 10647.50 41.59 6012.03 2538.68 9621.74 00:31:03.579 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 7: 9789.90 38.24 6539.89 2424.01 11053.40 00:31:03.579 ======================================================== 00:31:03.579 Total : 41801.69 163.29 6125.41 2424.01 11053.40 00:31:03.579 00:31:03.579 17:46:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@92 -- # nvmftestfini 00:31:03.579 17:46:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@516 -- # nvmfcleanup 00:31:03.579 17:46:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@121 -- # sync 00:31:03.579 17:46:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:31:03.579 17:46:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@124 -- # set +e 00:31:03.579 17:46:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@125 -- # for i in {1..20} 00:31:03.579 17:46:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:31:03.579 rmmod nvme_tcp 00:31:03.579 rmmod nvme_fabrics 00:31:03.579 rmmod nvme_keyring 00:31:03.580 17:46:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:31:03.580 17:46:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@128 -- # set -e 00:31:03.580 17:46:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@129 -- # return 0 00:31:03.580 17:46:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@517 -- # '[' -n 328619 ']' 00:31:03.580 17:46:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@518 -- # killprocess 328619 00:31:03.580 17:46:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@954 -- # '[' -z 328619 ']' 00:31:03.580 17:46:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@958 -- # kill -0 328619 00:31:03.580 17:46:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@959 -- # uname 00:31:03.580 17:46:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:31:03.580 17:46:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 328619 00:31:03.580 17:46:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:31:03.580 17:46:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:31:03.580 17:46:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@972 -- # echo 'killing process with pid 328619' 00:31:03.580 killing process with pid 328619 00:31:03.580 17:46:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@973 -- # kill 328619 00:31:03.580 17:46:44 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@978 -- # wait 328619 00:31:03.580 17:46:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:31:03.580 17:46:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:31:03.580 17:46:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:31:03.580 17:46:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@297 -- # iptr 00:31:03.580 17:46:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@791 -- # iptables-save 00:31:03.580 17:46:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:31:03.580 17:46:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@791 -- # iptables-restore 00:31:03.580 17:46:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:31:03.580 17:46:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@302 -- # remove_spdk_ns 00:31:03.580 17:46:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:31:03.580 17:46:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:31:03.580 17:46:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:31:05.484 17:46:47 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:31:05.484 17:46:47 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@94 -- # adq_reload_driver 00:31:05.484 17:46:47 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@58 -- # modprobe -a sch_mqprio 00:31:05.484 17:46:47 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@61 -- # rmmod ice 00:31:06.051 17:46:47 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@62 -- # modprobe ice 00:31:08.576 17:46:50 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@63 -- # sleep 5 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@97 -- # nvmftestinit 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@476 -- # prepare_net_devs 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@438 -- # local -g is_hw=no 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@440 -- # remove_spdk_ns 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@309 -- # xtrace_disable 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@315 -- # pci_devs=() 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@315 -- # local -a pci_devs 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@316 -- # pci_net_devs=() 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@317 -- # pci_drivers=() 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@317 -- # local -A pci_drivers 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@319 -- # net_devs=() 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@319 -- # local -ga net_devs 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@320 -- # e810=() 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@320 -- # local -ga e810 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@321 -- # x722=() 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@321 -- # local -ga x722 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@322 -- # mlx=() 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@322 -- # local -ga mlx 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:31:13.848 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:31:13.848 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:31:13.848 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@418 -- # [[ up == up ]] 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:31:13.849 Found net devices under 0000:0a:00.0: cvl_0_0 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@418 -- # [[ up == up ]] 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:31:13.849 Found net devices under 0000:0a:00.1: cvl_0_1 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@442 -- # is_hw=yes 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:31:13.849 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:31:13.849 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.152 ms 00:31:13.849 00:31:13.849 --- 10.0.0.2 ping statistics --- 00:31:13.849 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:31:13.849 rtt min/avg/max/mdev = 0.152/0.152/0.152/0.000 ms 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:31:13.849 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:31:13.849 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.104 ms 00:31:13.849 00:31:13.849 --- 10.0.0.1 ping statistics --- 00:31:13.849 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:31:13.849 rtt min/avg/max/mdev = 0.104/0.104/0.104/0.000 ms 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@450 -- # return 0 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@98 -- # adq_configure_driver 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@22 -- # ip netns exec cvl_0_0_ns_spdk ethtool --offload cvl_0_0 hw-tc-offload on 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@24 -- # ip netns exec cvl_0_0_ns_spdk ethtool --set-priv-flags cvl_0_0 channel-pkt-inspect-optimize off 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@26 -- # sysctl -w net.core.busy_poll=1 00:31:13.849 net.core.busy_poll = 1 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@27 -- # sysctl -w net.core.busy_read=1 00:31:13.849 net.core.busy_read = 1 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@29 -- # tc=/usr/sbin/tc 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@31 -- # ip netns exec cvl_0_0_ns_spdk /usr/sbin/tc qdisc add dev cvl_0_0 root mqprio num_tc 2 map 0 1 queues 2@0 2@2 hw 1 mode channel 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@33 -- # ip netns exec cvl_0_0_ns_spdk /usr/sbin/tc qdisc add dev cvl_0_0 ingress 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@35 -- # ip netns exec cvl_0_0_ns_spdk /usr/sbin/tc filter add dev cvl_0_0 protocol ip parent ffff: prio 1 flower dst_ip 10.0.0.2/32 ip_proto tcp dst_port 4420 skip_sw hw_tc 1 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@38 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/nvmf/set_xps_rxqs cvl_0_0 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@99 -- # nvmfappstart -m 0xF --wait-for-rpc 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@726 -- # xtrace_disable 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@509 -- # nvmfpid=331390 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF --wait-for-rpc 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@510 -- # waitforlisten 331390 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@835 -- # '[' -z 331390 ']' 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@840 -- # local max_retries=100 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:31:13.849 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@844 -- # xtrace_disable 00:31:13.849 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:31:13.849 [2024-12-06 17:46:55.641289] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:31:13.849 [2024-12-06 17:46:55.641390] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:31:14.107 [2024-12-06 17:46:55.718746] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:31:14.107 [2024-12-06 17:46:55.768807] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:31:14.107 [2024-12-06 17:46:55.768869] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:31:14.107 [2024-12-06 17:46:55.768899] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:31:14.107 [2024-12-06 17:46:55.768912] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:31:14.107 [2024-12-06 17:46:55.768922] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:31:14.107 [2024-12-06 17:46:55.770464] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:31:14.107 [2024-12-06 17:46:55.770494] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:31:14.107 [2024-12-06 17:46:55.770554] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:31:14.107 [2024-12-06 17:46:55.770558] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:31:14.107 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:31:14.107 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@868 -- # return 0 00:31:14.107 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:31:14.107 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@732 -- # xtrace_disable 00:31:14.107 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:31:14.107 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:31:14.107 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@100 -- # adq_configure_nvmf_target 1 00:31:14.107 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@42 -- # rpc_cmd sock_get_default_impl 00:31:14.107 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@42 -- # jq -r .impl_name 00:31:14.107 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:14.107 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:31:14.107 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:14.107 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@42 -- # socket_impl=posix 00:31:14.107 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@43 -- # rpc_cmd sock_impl_set_options --enable-placement-id 1 --enable-zerocopy-send-server -i posix 00:31:14.107 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:14.107 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:31:14.365 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:14.365 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@44 -- # rpc_cmd framework_start_init 00:31:14.365 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:14.365 17:46:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:31:14.365 17:46:56 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:14.365 17:46:56 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@45 -- # rpc_cmd nvmf_create_transport -t tcp -o --io-unit-size 8192 --sock-priority 1 00:31:14.365 17:46:56 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:14.365 17:46:56 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:31:14.365 [2024-12-06 17:46:56.060942] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:31:14.365 17:46:56 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:14.365 17:46:56 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@46 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:31:14.365 17:46:56 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:14.365 17:46:56 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:31:14.365 Malloc1 00:31:14.365 17:46:56 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:14.365 17:46:56 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@47 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:31:14.365 17:46:56 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:14.365 17:46:56 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:31:14.365 17:46:56 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:14.365 17:46:56 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@48 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:31:14.365 17:46:56 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:14.365 17:46:56 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:31:14.365 17:46:56 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:14.365 17:46:56 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@49 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:31:14.365 17:46:56 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:14.365 17:46:56 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:31:14.365 [2024-12-06 17:46:56.125969] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:31:14.365 17:46:56 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:14.365 17:46:56 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@104 -- # perfpid=331430 00:31:14.365 17:46:56 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@105 -- # sleep 2 00:31:14.365 17:46:56 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@101 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 64 -o 4096 -w randread -t 10 -c 0xF0 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:31:16.891 17:46:58 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@107 -- # rpc_cmd nvmf_get_stats 00:31:16.891 17:46:58 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:16.891 17:46:58 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:31:16.891 17:46:58 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:16.891 17:46:58 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@107 -- # nvmf_stats='{ 00:31:16.891 "tick_rate": 2700000000, 00:31:16.891 "poll_groups": [ 00:31:16.891 { 00:31:16.891 "name": "nvmf_tgt_poll_group_000", 00:31:16.891 "admin_qpairs": 1, 00:31:16.891 "io_qpairs": 2, 00:31:16.891 "current_admin_qpairs": 1, 00:31:16.891 "current_io_qpairs": 2, 00:31:16.891 "pending_bdev_io": 0, 00:31:16.891 "completed_nvme_io": 26081, 00:31:16.891 "transports": [ 00:31:16.891 { 00:31:16.891 "trtype": "TCP" 00:31:16.891 } 00:31:16.891 ] 00:31:16.891 }, 00:31:16.891 { 00:31:16.891 "name": "nvmf_tgt_poll_group_001", 00:31:16.891 "admin_qpairs": 0, 00:31:16.891 "io_qpairs": 2, 00:31:16.891 "current_admin_qpairs": 0, 00:31:16.891 "current_io_qpairs": 2, 00:31:16.891 "pending_bdev_io": 0, 00:31:16.891 "completed_nvme_io": 25973, 00:31:16.891 "transports": [ 00:31:16.891 { 00:31:16.891 "trtype": "TCP" 00:31:16.891 } 00:31:16.891 ] 00:31:16.891 }, 00:31:16.891 { 00:31:16.891 "name": "nvmf_tgt_poll_group_002", 00:31:16.891 "admin_qpairs": 0, 00:31:16.891 "io_qpairs": 0, 00:31:16.891 "current_admin_qpairs": 0, 00:31:16.891 "current_io_qpairs": 0, 00:31:16.891 "pending_bdev_io": 0, 00:31:16.891 "completed_nvme_io": 0, 00:31:16.891 "transports": [ 00:31:16.891 { 00:31:16.891 "trtype": "TCP" 00:31:16.891 } 00:31:16.891 ] 00:31:16.891 }, 00:31:16.891 { 00:31:16.891 "name": "nvmf_tgt_poll_group_003", 00:31:16.891 "admin_qpairs": 0, 00:31:16.891 "io_qpairs": 0, 00:31:16.891 "current_admin_qpairs": 0, 00:31:16.891 "current_io_qpairs": 0, 00:31:16.891 "pending_bdev_io": 0, 00:31:16.891 "completed_nvme_io": 0, 00:31:16.891 "transports": [ 00:31:16.891 { 00:31:16.891 "trtype": "TCP" 00:31:16.891 } 00:31:16.891 ] 00:31:16.891 } 00:31:16.891 ] 00:31:16.891 }' 00:31:16.891 17:46:58 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@108 -- # jq -r '.poll_groups[] | select(.current_io_qpairs == 0) | length' 00:31:16.891 17:46:58 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@108 -- # wc -l 00:31:16.891 17:46:58 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@108 -- # count=2 00:31:16.891 17:46:58 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@109 -- # [[ 2 -lt 2 ]] 00:31:16.891 17:46:58 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@114 -- # wait 331430 00:31:25.012 Initializing NVMe Controllers 00:31:25.012 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:31:25.012 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 4 00:31:25.012 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 5 00:31:25.012 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 6 00:31:25.012 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 7 00:31:25.012 Initialization complete. Launching workers. 00:31:25.012 ======================================================== 00:31:25.012 Latency(us) 00:31:25.012 Device Information : IOPS MiB/s Average min max 00:31:25.012 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 4: 7771.80 30.36 8239.01 1800.58 52893.49 00:31:25.012 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 5: 7215.10 28.18 8887.73 1644.53 53583.13 00:31:25.012 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 6: 6079.50 23.75 10528.57 1859.97 54399.49 00:31:25.012 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 7: 5858.40 22.88 10930.27 1551.59 53947.76 00:31:25.012 ======================================================== 00:31:25.012 Total : 26924.79 105.17 9515.39 1551.59 54399.49 00:31:25.012 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@115 -- # nvmftestfini 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@516 -- # nvmfcleanup 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@121 -- # sync 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@124 -- # set +e 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@125 -- # for i in {1..20} 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:31:25.012 rmmod nvme_tcp 00:31:25.012 rmmod nvme_fabrics 00:31:25.012 rmmod nvme_keyring 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@128 -- # set -e 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@129 -- # return 0 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@517 -- # '[' -n 331390 ']' 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@518 -- # killprocess 331390 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@954 -- # '[' -z 331390 ']' 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@958 -- # kill -0 331390 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@959 -- # uname 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 331390 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@972 -- # echo 'killing process with pid 331390' 00:31:25.012 killing process with pid 331390 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@973 -- # kill 331390 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@978 -- # wait 331390 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@297 -- # iptr 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@791 -- # iptables-save 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@791 -- # iptables-restore 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@302 -- # remove_spdk_ns 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:31:25.012 17:47:06 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:31:28.300 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:31:28.300 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@117 -- # trap - SIGINT SIGTERM EXIT 00:31:28.300 00:31:28.300 real 0m47.448s 00:31:28.300 user 2m38.926s 00:31:28.300 sys 0m11.343s 00:31:28.300 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1130 -- # xtrace_disable 00:31:28.300 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:31:28.300 ************************************ 00:31:28.300 END TEST nvmf_perf_adq 00:31:28.300 ************************************ 00:31:28.300 17:47:09 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@65 -- # run_test nvmf_shutdown /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/shutdown.sh --transport=tcp 00:31:28.300 17:47:09 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:31:28.300 17:47:09 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:31:28.300 17:47:09 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:31:28.300 ************************************ 00:31:28.300 START TEST nvmf_shutdown 00:31:28.300 ************************************ 00:31:28.300 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/shutdown.sh --transport=tcp 00:31:28.300 * Looking for test storage... 00:31:28.301 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1711 -- # lcov --version 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@333 -- # local ver1 ver1_l 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@334 -- # local ver2 ver2_l 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@336 -- # IFS=.-: 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@336 -- # read -ra ver1 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@337 -- # IFS=.-: 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@337 -- # read -ra ver2 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@338 -- # local 'op=<' 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@340 -- # ver1_l=2 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@341 -- # ver2_l=1 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@344 -- # case "$op" in 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@345 -- # : 1 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@364 -- # (( v = 0 )) 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@365 -- # decimal 1 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@353 -- # local d=1 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@355 -- # echo 1 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@365 -- # ver1[v]=1 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@366 -- # decimal 2 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@353 -- # local d=2 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@355 -- # echo 2 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@366 -- # ver2[v]=2 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@368 -- # return 0 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:31:28.301 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:31:28.301 --rc genhtml_branch_coverage=1 00:31:28.301 --rc genhtml_function_coverage=1 00:31:28.301 --rc genhtml_legend=1 00:31:28.301 --rc geninfo_all_blocks=1 00:31:28.301 --rc geninfo_unexecuted_blocks=1 00:31:28.301 00:31:28.301 ' 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:31:28.301 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:31:28.301 --rc genhtml_branch_coverage=1 00:31:28.301 --rc genhtml_function_coverage=1 00:31:28.301 --rc genhtml_legend=1 00:31:28.301 --rc geninfo_all_blocks=1 00:31:28.301 --rc geninfo_unexecuted_blocks=1 00:31:28.301 00:31:28.301 ' 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:31:28.301 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:31:28.301 --rc genhtml_branch_coverage=1 00:31:28.301 --rc genhtml_function_coverage=1 00:31:28.301 --rc genhtml_legend=1 00:31:28.301 --rc geninfo_all_blocks=1 00:31:28.301 --rc geninfo_unexecuted_blocks=1 00:31:28.301 00:31:28.301 ' 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:31:28.301 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:31:28.301 --rc genhtml_branch_coverage=1 00:31:28.301 --rc genhtml_function_coverage=1 00:31:28.301 --rc genhtml_legend=1 00:31:28.301 --rc geninfo_all_blocks=1 00:31:28.301 --rc geninfo_unexecuted_blocks=1 00:31:28.301 00:31:28.301 ' 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@7 -- # uname -s 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@15 -- # shopt -s extglob 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- paths/export.sh@5 -- # export PATH 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@51 -- # : 0 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:31:28.301 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@55 -- # have_pci_nics=0 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@12 -- # MALLOC_BDEV_SIZE=64 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@162 -- # run_test nvmf_shutdown_tc1 nvmf_shutdown_tc1 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1111 -- # xtrace_disable 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@10 -- # set +x 00:31:28.301 ************************************ 00:31:28.301 START TEST nvmf_shutdown_tc1 00:31:28.301 ************************************ 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@1129 -- # nvmf_shutdown_tc1 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@75 -- # starttarget 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@16 -- # nvmftestinit 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@476 -- # prepare_net_devs 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@438 -- # local -g is_hw=no 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@440 -- # remove_spdk_ns 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:31:28.301 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:31:28.302 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:31:28.302 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:31:28.302 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@309 -- # xtrace_disable 00:31:28.302 17:47:09 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@315 -- # pci_devs=() 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@315 -- # local -a pci_devs 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@316 -- # pci_net_devs=() 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@317 -- # pci_drivers=() 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@317 -- # local -A pci_drivers 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@319 -- # net_devs=() 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@319 -- # local -ga net_devs 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@320 -- # e810=() 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@320 -- # local -ga e810 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@321 -- # x722=() 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@321 -- # local -ga x722 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@322 -- # mlx=() 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@322 -- # local -ga mlx 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:31:30.202 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:31:30.202 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@418 -- # [[ up == up ]] 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:31:30.202 Found net devices under 0000:0a:00.0: cvl_0_0 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:31:30.202 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:31:30.203 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@418 -- # [[ up == up ]] 00:31:30.203 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:31:30.203 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:31:30.203 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:31:30.203 Found net devices under 0000:0a:00.1: cvl_0_1 00:31:30.203 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:31:30.203 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:31:30.203 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@442 -- # is_hw=yes 00:31:30.203 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:31:30.203 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:31:30.203 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:31:30.203 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:31:30.203 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:31:30.203 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:31:30.203 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:31:30.203 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:31:30.203 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:31:30.203 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:31:30.203 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:31:30.203 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:31:30.203 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:31:30.203 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:31:30.203 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:31:30.203 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:31:30.203 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:31:30.203 17:47:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:31:30.203 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:31:30.203 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:31:30.461 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:31:30.461 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:31:30.461 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:31:30.461 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:31:30.461 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:31:30.461 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:31:30.461 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:31:30.461 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.142 ms 00:31:30.461 00:31:30.461 --- 10.0.0.2 ping statistics --- 00:31:30.461 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:31:30.461 rtt min/avg/max/mdev = 0.142/0.142/0.142/0.000 ms 00:31:30.461 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:31:30.461 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:31:30.461 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.101 ms 00:31:30.462 00:31:30.462 --- 10.0.0.1 ping statistics --- 00:31:30.462 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:31:30.462 rtt min/avg/max/mdev = 0.101/0.101/0.101/0.000 ms 00:31:30.462 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:31:30.462 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@450 -- # return 0 00:31:30.462 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:31:30.462 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:31:30.462 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:31:30.462 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:31:30.462 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:31:30.462 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:31:30.462 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:31:30.462 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@19 -- # nvmfappstart -m 0x1E 00:31:30.462 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:31:30.462 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@726 -- # xtrace_disable 00:31:30.462 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:31:30.462 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@509 -- # nvmfpid=334725 00:31:30.462 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1E 00:31:30.462 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@510 -- # waitforlisten 334725 00:31:30.462 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@835 -- # '[' -z 334725 ']' 00:31:30.462 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:31:30.462 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@840 -- # local max_retries=100 00:31:30.462 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:31:30.462 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:31:30.462 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@844 -- # xtrace_disable 00:31:30.462 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:31:30.462 [2024-12-06 17:47:12.177290] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:31:30.462 [2024-12-06 17:47:12.177365] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:31:30.462 [2024-12-06 17:47:12.251906] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:31:30.720 [2024-12-06 17:47:12.300882] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:31:30.720 [2024-12-06 17:47:12.300952] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:31:30.720 [2024-12-06 17:47:12.300974] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:31:30.720 [2024-12-06 17:47:12.301012] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:31:30.720 [2024-12-06 17:47:12.301022] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:31:30.720 [2024-12-06 17:47:12.302717] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:31:30.720 [2024-12-06 17:47:12.302780] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:31:30.720 [2024-12-06 17:47:12.302823] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:31:30.720 [2024-12-06 17:47:12.302827] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@868 -- # return 0 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@732 -- # xtrace_disable 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@21 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:31:30.720 [2024-12-06 17:47:12.453567] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@23 -- # num_subsystems=({1..10}) 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@25 -- # timing_enter create_subsystems 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@726 -- # xtrace_disable 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@27 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:30.720 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:31:30.721 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@36 -- # rpc_cmd 00:31:30.721 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:30.721 17:47:12 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:31:30.721 Malloc1 00:31:30.721 [2024-12-06 17:47:12.555035] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:31:30.978 Malloc2 00:31:30.978 Malloc3 00:31:30.978 Malloc4 00:31:30.978 Malloc5 00:31:30.978 Malloc6 00:31:31.237 Malloc7 00:31:31.237 Malloc8 00:31:31.237 Malloc9 00:31:31.237 Malloc10 00:31:31.237 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:31.237 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@37 -- # timing_exit create_subsystems 00:31:31.237 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@732 -- # xtrace_disable 00:31:31.237 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:31:31.237 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@79 -- # perfpid=334898 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@80 -- # waitforlisten 334898 /var/tmp/bdevperf.sock 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@835 -- # '[' -z 334898 ']' 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@78 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -m 0x1 -i 1 -r /var/tmp/bdevperf.sock --json /dev/fd/63 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@78 -- # gen_nvmf_target_json 1 2 3 4 5 6 7 8 9 10 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@840 -- # local max_retries=100 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # config=() 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:31:31.238 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # local subsystem config 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@844 -- # xtrace_disable 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:31.238 { 00:31:31.238 "params": { 00:31:31.238 "name": "Nvme$subsystem", 00:31:31.238 "trtype": "$TEST_TRANSPORT", 00:31:31.238 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:31.238 "adrfam": "ipv4", 00:31:31.238 "trsvcid": "$NVMF_PORT", 00:31:31.238 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:31.238 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:31.238 "hdgst": ${hdgst:-false}, 00:31:31.238 "ddgst": ${ddgst:-false} 00:31:31.238 }, 00:31:31.238 "method": "bdev_nvme_attach_controller" 00:31:31.238 } 00:31:31.238 EOF 00:31:31.238 )") 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:31.238 { 00:31:31.238 "params": { 00:31:31.238 "name": "Nvme$subsystem", 00:31:31.238 "trtype": "$TEST_TRANSPORT", 00:31:31.238 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:31.238 "adrfam": "ipv4", 00:31:31.238 "trsvcid": "$NVMF_PORT", 00:31:31.238 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:31.238 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:31.238 "hdgst": ${hdgst:-false}, 00:31:31.238 "ddgst": ${ddgst:-false} 00:31:31.238 }, 00:31:31.238 "method": "bdev_nvme_attach_controller" 00:31:31.238 } 00:31:31.238 EOF 00:31:31.238 )") 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:31.238 { 00:31:31.238 "params": { 00:31:31.238 "name": "Nvme$subsystem", 00:31:31.238 "trtype": "$TEST_TRANSPORT", 00:31:31.238 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:31.238 "adrfam": "ipv4", 00:31:31.238 "trsvcid": "$NVMF_PORT", 00:31:31.238 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:31.238 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:31.238 "hdgst": ${hdgst:-false}, 00:31:31.238 "ddgst": ${ddgst:-false} 00:31:31.238 }, 00:31:31.238 "method": "bdev_nvme_attach_controller" 00:31:31.238 } 00:31:31.238 EOF 00:31:31.238 )") 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:31.238 { 00:31:31.238 "params": { 00:31:31.238 "name": "Nvme$subsystem", 00:31:31.238 "trtype": "$TEST_TRANSPORT", 00:31:31.238 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:31.238 "adrfam": "ipv4", 00:31:31.238 "trsvcid": "$NVMF_PORT", 00:31:31.238 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:31.238 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:31.238 "hdgst": ${hdgst:-false}, 00:31:31.238 "ddgst": ${ddgst:-false} 00:31:31.238 }, 00:31:31.238 "method": "bdev_nvme_attach_controller" 00:31:31.238 } 00:31:31.238 EOF 00:31:31.238 )") 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:31.238 { 00:31:31.238 "params": { 00:31:31.238 "name": "Nvme$subsystem", 00:31:31.238 "trtype": "$TEST_TRANSPORT", 00:31:31.238 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:31.238 "adrfam": "ipv4", 00:31:31.238 "trsvcid": "$NVMF_PORT", 00:31:31.238 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:31.238 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:31.238 "hdgst": ${hdgst:-false}, 00:31:31.238 "ddgst": ${ddgst:-false} 00:31:31.238 }, 00:31:31.238 "method": "bdev_nvme_attach_controller" 00:31:31.238 } 00:31:31.238 EOF 00:31:31.238 )") 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:31.238 { 00:31:31.238 "params": { 00:31:31.238 "name": "Nvme$subsystem", 00:31:31.238 "trtype": "$TEST_TRANSPORT", 00:31:31.238 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:31.238 "adrfam": "ipv4", 00:31:31.238 "trsvcid": "$NVMF_PORT", 00:31:31.238 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:31.238 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:31.238 "hdgst": ${hdgst:-false}, 00:31:31.238 "ddgst": ${ddgst:-false} 00:31:31.238 }, 00:31:31.238 "method": "bdev_nvme_attach_controller" 00:31:31.238 } 00:31:31.238 EOF 00:31:31.238 )") 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:31.238 { 00:31:31.238 "params": { 00:31:31.238 "name": "Nvme$subsystem", 00:31:31.238 "trtype": "$TEST_TRANSPORT", 00:31:31.238 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:31.238 "adrfam": "ipv4", 00:31:31.238 "trsvcid": "$NVMF_PORT", 00:31:31.238 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:31.238 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:31.238 "hdgst": ${hdgst:-false}, 00:31:31.238 "ddgst": ${ddgst:-false} 00:31:31.238 }, 00:31:31.238 "method": "bdev_nvme_attach_controller" 00:31:31.238 } 00:31:31.238 EOF 00:31:31.238 )") 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:31.238 { 00:31:31.238 "params": { 00:31:31.238 "name": "Nvme$subsystem", 00:31:31.238 "trtype": "$TEST_TRANSPORT", 00:31:31.238 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:31.238 "adrfam": "ipv4", 00:31:31.238 "trsvcid": "$NVMF_PORT", 00:31:31.238 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:31.238 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:31.238 "hdgst": ${hdgst:-false}, 00:31:31.238 "ddgst": ${ddgst:-false} 00:31:31.238 }, 00:31:31.238 "method": "bdev_nvme_attach_controller" 00:31:31.238 } 00:31:31.238 EOF 00:31:31.238 )") 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:31.238 { 00:31:31.238 "params": { 00:31:31.238 "name": "Nvme$subsystem", 00:31:31.238 "trtype": "$TEST_TRANSPORT", 00:31:31.238 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:31.238 "adrfam": "ipv4", 00:31:31.238 "trsvcid": "$NVMF_PORT", 00:31:31.238 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:31.238 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:31.238 "hdgst": ${hdgst:-false}, 00:31:31.238 "ddgst": ${ddgst:-false} 00:31:31.238 }, 00:31:31.238 "method": "bdev_nvme_attach_controller" 00:31:31.238 } 00:31:31.238 EOF 00:31:31.238 )") 00:31:31.238 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:31:31.239 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:31.239 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:31.239 { 00:31:31.239 "params": { 00:31:31.239 "name": "Nvme$subsystem", 00:31:31.239 "trtype": "$TEST_TRANSPORT", 00:31:31.239 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:31.239 "adrfam": "ipv4", 00:31:31.239 "trsvcid": "$NVMF_PORT", 00:31:31.239 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:31.239 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:31.239 "hdgst": ${hdgst:-false}, 00:31:31.239 "ddgst": ${ddgst:-false} 00:31:31.239 }, 00:31:31.239 "method": "bdev_nvme_attach_controller" 00:31:31.239 } 00:31:31.239 EOF 00:31:31.239 )") 00:31:31.239 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:31:31.239 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@584 -- # jq . 00:31:31.239 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@585 -- # IFS=, 00:31:31.239 17:47:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:31:31.239 "params": { 00:31:31.239 "name": "Nvme1", 00:31:31.239 "trtype": "tcp", 00:31:31.239 "traddr": "10.0.0.2", 00:31:31.239 "adrfam": "ipv4", 00:31:31.239 "trsvcid": "4420", 00:31:31.239 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:31:31.239 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:31:31.239 "hdgst": false, 00:31:31.239 "ddgst": false 00:31:31.239 }, 00:31:31.239 "method": "bdev_nvme_attach_controller" 00:31:31.239 },{ 00:31:31.239 "params": { 00:31:31.239 "name": "Nvme2", 00:31:31.239 "trtype": "tcp", 00:31:31.239 "traddr": "10.0.0.2", 00:31:31.239 "adrfam": "ipv4", 00:31:31.239 "trsvcid": "4420", 00:31:31.239 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:31:31.239 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:31:31.239 "hdgst": false, 00:31:31.239 "ddgst": false 00:31:31.239 }, 00:31:31.239 "method": "bdev_nvme_attach_controller" 00:31:31.239 },{ 00:31:31.239 "params": { 00:31:31.239 "name": "Nvme3", 00:31:31.239 "trtype": "tcp", 00:31:31.239 "traddr": "10.0.0.2", 00:31:31.239 "adrfam": "ipv4", 00:31:31.239 "trsvcid": "4420", 00:31:31.239 "subnqn": "nqn.2016-06.io.spdk:cnode3", 00:31:31.239 "hostnqn": "nqn.2016-06.io.spdk:host3", 00:31:31.239 "hdgst": false, 00:31:31.239 "ddgst": false 00:31:31.239 }, 00:31:31.239 "method": "bdev_nvme_attach_controller" 00:31:31.239 },{ 00:31:31.239 "params": { 00:31:31.239 "name": "Nvme4", 00:31:31.239 "trtype": "tcp", 00:31:31.239 "traddr": "10.0.0.2", 00:31:31.239 "adrfam": "ipv4", 00:31:31.239 "trsvcid": "4420", 00:31:31.239 "subnqn": "nqn.2016-06.io.spdk:cnode4", 00:31:31.239 "hostnqn": "nqn.2016-06.io.spdk:host4", 00:31:31.239 "hdgst": false, 00:31:31.239 "ddgst": false 00:31:31.239 }, 00:31:31.239 "method": "bdev_nvme_attach_controller" 00:31:31.239 },{ 00:31:31.239 "params": { 00:31:31.239 "name": "Nvme5", 00:31:31.239 "trtype": "tcp", 00:31:31.239 "traddr": "10.0.0.2", 00:31:31.239 "adrfam": "ipv4", 00:31:31.239 "trsvcid": "4420", 00:31:31.239 "subnqn": "nqn.2016-06.io.spdk:cnode5", 00:31:31.239 "hostnqn": "nqn.2016-06.io.spdk:host5", 00:31:31.239 "hdgst": false, 00:31:31.239 "ddgst": false 00:31:31.239 }, 00:31:31.239 "method": "bdev_nvme_attach_controller" 00:31:31.239 },{ 00:31:31.239 "params": { 00:31:31.239 "name": "Nvme6", 00:31:31.239 "trtype": "tcp", 00:31:31.239 "traddr": "10.0.0.2", 00:31:31.239 "adrfam": "ipv4", 00:31:31.239 "trsvcid": "4420", 00:31:31.239 "subnqn": "nqn.2016-06.io.spdk:cnode6", 00:31:31.239 "hostnqn": "nqn.2016-06.io.spdk:host6", 00:31:31.239 "hdgst": false, 00:31:31.239 "ddgst": false 00:31:31.239 }, 00:31:31.239 "method": "bdev_nvme_attach_controller" 00:31:31.239 },{ 00:31:31.239 "params": { 00:31:31.239 "name": "Nvme7", 00:31:31.239 "trtype": "tcp", 00:31:31.239 "traddr": "10.0.0.2", 00:31:31.239 "adrfam": "ipv4", 00:31:31.239 "trsvcid": "4420", 00:31:31.239 "subnqn": "nqn.2016-06.io.spdk:cnode7", 00:31:31.239 "hostnqn": "nqn.2016-06.io.spdk:host7", 00:31:31.239 "hdgst": false, 00:31:31.239 "ddgst": false 00:31:31.239 }, 00:31:31.239 "method": "bdev_nvme_attach_controller" 00:31:31.239 },{ 00:31:31.239 "params": { 00:31:31.239 "name": "Nvme8", 00:31:31.239 "trtype": "tcp", 00:31:31.239 "traddr": "10.0.0.2", 00:31:31.239 "adrfam": "ipv4", 00:31:31.239 "trsvcid": "4420", 00:31:31.239 "subnqn": "nqn.2016-06.io.spdk:cnode8", 00:31:31.239 "hostnqn": "nqn.2016-06.io.spdk:host8", 00:31:31.239 "hdgst": false, 00:31:31.239 "ddgst": false 00:31:31.239 }, 00:31:31.239 "method": "bdev_nvme_attach_controller" 00:31:31.239 },{ 00:31:31.239 "params": { 00:31:31.239 "name": "Nvme9", 00:31:31.239 "trtype": "tcp", 00:31:31.239 "traddr": "10.0.0.2", 00:31:31.239 "adrfam": "ipv4", 00:31:31.239 "trsvcid": "4420", 00:31:31.239 "subnqn": "nqn.2016-06.io.spdk:cnode9", 00:31:31.239 "hostnqn": "nqn.2016-06.io.spdk:host9", 00:31:31.239 "hdgst": false, 00:31:31.239 "ddgst": false 00:31:31.239 }, 00:31:31.239 "method": "bdev_nvme_attach_controller" 00:31:31.239 },{ 00:31:31.239 "params": { 00:31:31.239 "name": "Nvme10", 00:31:31.239 "trtype": "tcp", 00:31:31.239 "traddr": "10.0.0.2", 00:31:31.239 "adrfam": "ipv4", 00:31:31.239 "trsvcid": "4420", 00:31:31.239 "subnqn": "nqn.2016-06.io.spdk:cnode10", 00:31:31.239 "hostnqn": "nqn.2016-06.io.spdk:host10", 00:31:31.239 "hdgst": false, 00:31:31.239 "ddgst": false 00:31:31.239 }, 00:31:31.239 "method": "bdev_nvme_attach_controller" 00:31:31.239 }' 00:31:31.239 [2024-12-06 17:47:13.070353] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:31:31.239 [2024-12-06 17:47:13.070444] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk1 --proc-type=auto ] 00:31:31.498 [2024-12-06 17:47:13.144596] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:31:31.498 [2024-12-06 17:47:13.191836] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:31:33.405 17:47:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:31:33.405 17:47:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@868 -- # return 0 00:31:33.406 17:47:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@81 -- # rpc_cmd -s /var/tmp/bdevperf.sock framework_wait_init 00:31:33.406 17:47:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:33.406 17:47:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:31:33.406 17:47:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:33.406 17:47:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@84 -- # kill -9 334898 00:31:33.406 17:47:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@85 -- # rm -f /var/run/spdk_bdev1 00:31:33.406 17:47:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@88 -- # sleep 1 00:31:34.341 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/shutdown.sh: line 74: 334898 Killed $rootdir/test/app/bdev_svc/bdev_svc -m 0x1 -i 1 -r /var/tmp/bdevperf.sock --json <(gen_nvmf_target_json "${num_subsystems[@]}") 00:31:34.341 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@89 -- # kill -0 334725 00:31:34.341 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@92 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -q 64 -o 65536 -w verify -t 1 00:31:34.341 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@92 -- # gen_nvmf_target_json 1 2 3 4 5 6 7 8 9 10 00:31:34.341 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # config=() 00:31:34.341 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # local subsystem config 00:31:34.341 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:34.341 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:34.341 { 00:31:34.341 "params": { 00:31:34.341 "name": "Nvme$subsystem", 00:31:34.341 "trtype": "$TEST_TRANSPORT", 00:31:34.341 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:34.341 "adrfam": "ipv4", 00:31:34.341 "trsvcid": "$NVMF_PORT", 00:31:34.341 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:34.341 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:34.341 "hdgst": ${hdgst:-false}, 00:31:34.341 "ddgst": ${ddgst:-false} 00:31:34.341 }, 00:31:34.341 "method": "bdev_nvme_attach_controller" 00:31:34.341 } 00:31:34.341 EOF 00:31:34.341 )") 00:31:34.341 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:31:34.341 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:34.341 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:34.341 { 00:31:34.341 "params": { 00:31:34.341 "name": "Nvme$subsystem", 00:31:34.341 "trtype": "$TEST_TRANSPORT", 00:31:34.341 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:34.341 "adrfam": "ipv4", 00:31:34.341 "trsvcid": "$NVMF_PORT", 00:31:34.341 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:34.341 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:34.341 "hdgst": ${hdgst:-false}, 00:31:34.341 "ddgst": ${ddgst:-false} 00:31:34.341 }, 00:31:34.341 "method": "bdev_nvme_attach_controller" 00:31:34.341 } 00:31:34.341 EOF 00:31:34.341 )") 00:31:34.341 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:31:34.341 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:34.341 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:34.341 { 00:31:34.341 "params": { 00:31:34.341 "name": "Nvme$subsystem", 00:31:34.341 "trtype": "$TEST_TRANSPORT", 00:31:34.341 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:34.341 "adrfam": "ipv4", 00:31:34.341 "trsvcid": "$NVMF_PORT", 00:31:34.341 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:34.341 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:34.341 "hdgst": ${hdgst:-false}, 00:31:34.341 "ddgst": ${ddgst:-false} 00:31:34.341 }, 00:31:34.341 "method": "bdev_nvme_attach_controller" 00:31:34.341 } 00:31:34.341 EOF 00:31:34.341 )") 00:31:34.341 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:31:34.341 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:34.341 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:34.341 { 00:31:34.341 "params": { 00:31:34.341 "name": "Nvme$subsystem", 00:31:34.341 "trtype": "$TEST_TRANSPORT", 00:31:34.341 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:34.341 "adrfam": "ipv4", 00:31:34.341 "trsvcid": "$NVMF_PORT", 00:31:34.341 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:34.341 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:34.341 "hdgst": ${hdgst:-false}, 00:31:34.341 "ddgst": ${ddgst:-false} 00:31:34.341 }, 00:31:34.341 "method": "bdev_nvme_attach_controller" 00:31:34.341 } 00:31:34.341 EOF 00:31:34.341 )") 00:31:34.341 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:31:34.341 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:34.341 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:34.341 { 00:31:34.341 "params": { 00:31:34.341 "name": "Nvme$subsystem", 00:31:34.341 "trtype": "$TEST_TRANSPORT", 00:31:34.341 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:34.341 "adrfam": "ipv4", 00:31:34.341 "trsvcid": "$NVMF_PORT", 00:31:34.341 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:34.341 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:34.341 "hdgst": ${hdgst:-false}, 00:31:34.341 "ddgst": ${ddgst:-false} 00:31:34.341 }, 00:31:34.341 "method": "bdev_nvme_attach_controller" 00:31:34.341 } 00:31:34.341 EOF 00:31:34.341 )") 00:31:34.341 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:31:34.341 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:34.341 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:34.341 { 00:31:34.341 "params": { 00:31:34.341 "name": "Nvme$subsystem", 00:31:34.341 "trtype": "$TEST_TRANSPORT", 00:31:34.341 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:34.341 "adrfam": "ipv4", 00:31:34.341 "trsvcid": "$NVMF_PORT", 00:31:34.341 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:34.341 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:34.341 "hdgst": ${hdgst:-false}, 00:31:34.341 "ddgst": ${ddgst:-false} 00:31:34.341 }, 00:31:34.341 "method": "bdev_nvme_attach_controller" 00:31:34.341 } 00:31:34.341 EOF 00:31:34.341 )") 00:31:34.341 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:31:34.341 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:34.341 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:34.341 { 00:31:34.341 "params": { 00:31:34.341 "name": "Nvme$subsystem", 00:31:34.341 "trtype": "$TEST_TRANSPORT", 00:31:34.341 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:34.341 "adrfam": "ipv4", 00:31:34.341 "trsvcid": "$NVMF_PORT", 00:31:34.341 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:34.341 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:34.341 "hdgst": ${hdgst:-false}, 00:31:34.341 "ddgst": ${ddgst:-false} 00:31:34.341 }, 00:31:34.342 "method": "bdev_nvme_attach_controller" 00:31:34.342 } 00:31:34.342 EOF 00:31:34.342 )") 00:31:34.342 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:31:34.342 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:34.342 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:34.342 { 00:31:34.342 "params": { 00:31:34.342 "name": "Nvme$subsystem", 00:31:34.342 "trtype": "$TEST_TRANSPORT", 00:31:34.342 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:34.342 "adrfam": "ipv4", 00:31:34.342 "trsvcid": "$NVMF_PORT", 00:31:34.342 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:34.342 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:34.342 "hdgst": ${hdgst:-false}, 00:31:34.342 "ddgst": ${ddgst:-false} 00:31:34.342 }, 00:31:34.342 "method": "bdev_nvme_attach_controller" 00:31:34.342 } 00:31:34.342 EOF 00:31:34.342 )") 00:31:34.342 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:31:34.342 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:34.342 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:34.342 { 00:31:34.342 "params": { 00:31:34.342 "name": "Nvme$subsystem", 00:31:34.342 "trtype": "$TEST_TRANSPORT", 00:31:34.342 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:34.342 "adrfam": "ipv4", 00:31:34.342 "trsvcid": "$NVMF_PORT", 00:31:34.342 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:34.342 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:34.342 "hdgst": ${hdgst:-false}, 00:31:34.342 "ddgst": ${ddgst:-false} 00:31:34.342 }, 00:31:34.342 "method": "bdev_nvme_attach_controller" 00:31:34.342 } 00:31:34.342 EOF 00:31:34.342 )") 00:31:34.342 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:31:34.342 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:34.342 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:34.342 { 00:31:34.342 "params": { 00:31:34.342 "name": "Nvme$subsystem", 00:31:34.342 "trtype": "$TEST_TRANSPORT", 00:31:34.342 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:34.342 "adrfam": "ipv4", 00:31:34.342 "trsvcid": "$NVMF_PORT", 00:31:34.342 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:34.342 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:34.342 "hdgst": ${hdgst:-false}, 00:31:34.342 "ddgst": ${ddgst:-false} 00:31:34.342 }, 00:31:34.342 "method": "bdev_nvme_attach_controller" 00:31:34.342 } 00:31:34.342 EOF 00:31:34.342 )") 00:31:34.342 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # cat 00:31:34.342 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@584 -- # jq . 00:31:34.342 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@585 -- # IFS=, 00:31:34.342 17:47:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:31:34.342 "params": { 00:31:34.342 "name": "Nvme1", 00:31:34.342 "trtype": "tcp", 00:31:34.342 "traddr": "10.0.0.2", 00:31:34.342 "adrfam": "ipv4", 00:31:34.342 "trsvcid": "4420", 00:31:34.342 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:31:34.342 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:31:34.342 "hdgst": false, 00:31:34.342 "ddgst": false 00:31:34.342 }, 00:31:34.342 "method": "bdev_nvme_attach_controller" 00:31:34.342 },{ 00:31:34.342 "params": { 00:31:34.342 "name": "Nvme2", 00:31:34.342 "trtype": "tcp", 00:31:34.342 "traddr": "10.0.0.2", 00:31:34.342 "adrfam": "ipv4", 00:31:34.342 "trsvcid": "4420", 00:31:34.342 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:31:34.342 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:31:34.342 "hdgst": false, 00:31:34.342 "ddgst": false 00:31:34.342 }, 00:31:34.342 "method": "bdev_nvme_attach_controller" 00:31:34.342 },{ 00:31:34.342 "params": { 00:31:34.342 "name": "Nvme3", 00:31:34.342 "trtype": "tcp", 00:31:34.342 "traddr": "10.0.0.2", 00:31:34.342 "adrfam": "ipv4", 00:31:34.342 "trsvcid": "4420", 00:31:34.342 "subnqn": "nqn.2016-06.io.spdk:cnode3", 00:31:34.342 "hostnqn": "nqn.2016-06.io.spdk:host3", 00:31:34.342 "hdgst": false, 00:31:34.342 "ddgst": false 00:31:34.342 }, 00:31:34.342 "method": "bdev_nvme_attach_controller" 00:31:34.342 },{ 00:31:34.342 "params": { 00:31:34.342 "name": "Nvme4", 00:31:34.342 "trtype": "tcp", 00:31:34.342 "traddr": "10.0.0.2", 00:31:34.342 "adrfam": "ipv4", 00:31:34.342 "trsvcid": "4420", 00:31:34.342 "subnqn": "nqn.2016-06.io.spdk:cnode4", 00:31:34.342 "hostnqn": "nqn.2016-06.io.spdk:host4", 00:31:34.342 "hdgst": false, 00:31:34.342 "ddgst": false 00:31:34.342 }, 00:31:34.342 "method": "bdev_nvme_attach_controller" 00:31:34.342 },{ 00:31:34.342 "params": { 00:31:34.342 "name": "Nvme5", 00:31:34.342 "trtype": "tcp", 00:31:34.342 "traddr": "10.0.0.2", 00:31:34.342 "adrfam": "ipv4", 00:31:34.342 "trsvcid": "4420", 00:31:34.342 "subnqn": "nqn.2016-06.io.spdk:cnode5", 00:31:34.342 "hostnqn": "nqn.2016-06.io.spdk:host5", 00:31:34.342 "hdgst": false, 00:31:34.342 "ddgst": false 00:31:34.342 }, 00:31:34.342 "method": "bdev_nvme_attach_controller" 00:31:34.342 },{ 00:31:34.342 "params": { 00:31:34.342 "name": "Nvme6", 00:31:34.342 "trtype": "tcp", 00:31:34.342 "traddr": "10.0.0.2", 00:31:34.342 "adrfam": "ipv4", 00:31:34.342 "trsvcid": "4420", 00:31:34.342 "subnqn": "nqn.2016-06.io.spdk:cnode6", 00:31:34.342 "hostnqn": "nqn.2016-06.io.spdk:host6", 00:31:34.342 "hdgst": false, 00:31:34.342 "ddgst": false 00:31:34.342 }, 00:31:34.342 "method": "bdev_nvme_attach_controller" 00:31:34.342 },{ 00:31:34.342 "params": { 00:31:34.342 "name": "Nvme7", 00:31:34.342 "trtype": "tcp", 00:31:34.342 "traddr": "10.0.0.2", 00:31:34.342 "adrfam": "ipv4", 00:31:34.342 "trsvcid": "4420", 00:31:34.342 "subnqn": "nqn.2016-06.io.spdk:cnode7", 00:31:34.342 "hostnqn": "nqn.2016-06.io.spdk:host7", 00:31:34.342 "hdgst": false, 00:31:34.342 "ddgst": false 00:31:34.342 }, 00:31:34.342 "method": "bdev_nvme_attach_controller" 00:31:34.342 },{ 00:31:34.342 "params": { 00:31:34.342 "name": "Nvme8", 00:31:34.342 "trtype": "tcp", 00:31:34.342 "traddr": "10.0.0.2", 00:31:34.342 "adrfam": "ipv4", 00:31:34.342 "trsvcid": "4420", 00:31:34.342 "subnqn": "nqn.2016-06.io.spdk:cnode8", 00:31:34.342 "hostnqn": "nqn.2016-06.io.spdk:host8", 00:31:34.342 "hdgst": false, 00:31:34.342 "ddgst": false 00:31:34.342 }, 00:31:34.342 "method": "bdev_nvme_attach_controller" 00:31:34.342 },{ 00:31:34.342 "params": { 00:31:34.342 "name": "Nvme9", 00:31:34.342 "trtype": "tcp", 00:31:34.342 "traddr": "10.0.0.2", 00:31:34.342 "adrfam": "ipv4", 00:31:34.342 "trsvcid": "4420", 00:31:34.342 "subnqn": "nqn.2016-06.io.spdk:cnode9", 00:31:34.342 "hostnqn": "nqn.2016-06.io.spdk:host9", 00:31:34.342 "hdgst": false, 00:31:34.342 "ddgst": false 00:31:34.342 }, 00:31:34.342 "method": "bdev_nvme_attach_controller" 00:31:34.342 },{ 00:31:34.342 "params": { 00:31:34.342 "name": "Nvme10", 00:31:34.342 "trtype": "tcp", 00:31:34.342 "traddr": "10.0.0.2", 00:31:34.342 "adrfam": "ipv4", 00:31:34.342 "trsvcid": "4420", 00:31:34.342 "subnqn": "nqn.2016-06.io.spdk:cnode10", 00:31:34.342 "hostnqn": "nqn.2016-06.io.spdk:host10", 00:31:34.342 "hdgst": false, 00:31:34.342 "ddgst": false 00:31:34.342 }, 00:31:34.342 "method": "bdev_nvme_attach_controller" 00:31:34.342 }' 00:31:34.342 [2024-12-06 17:47:16.143806] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:31:34.342 [2024-12-06 17:47:16.143892] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid335316 ] 00:31:34.600 [2024-12-06 17:47:16.219141] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:31:34.600 [2024-12-06 17:47:16.268522] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:31:35.974 Running I/O for 1 seconds... 00:31:37.166 1813.00 IOPS, 113.31 MiB/s 00:31:37.166 Latency(us) 00:31:37.166 [2024-12-06T16:47:19.005Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:31:37.166 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:37.166 Verification LBA range: start 0x0 length 0x400 00:31:37.166 Nvme1n1 : 1.14 223.81 13.99 0.00 0.00 283213.56 17961.72 256318.58 00:31:37.166 Job: Nvme2n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:37.166 Verification LBA range: start 0x0 length 0x400 00:31:37.166 Nvme2n1 : 1.13 234.56 14.66 0.00 0.00 258676.96 11359.57 239230.67 00:31:37.166 Job: Nvme3n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:37.166 Verification LBA range: start 0x0 length 0x400 00:31:37.166 Nvme3n1 : 1.14 224.97 14.06 0.00 0.00 272626.16 21942.42 254765.13 00:31:37.166 Job: Nvme4n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:37.166 Verification LBA range: start 0x0 length 0x400 00:31:37.166 Nvme4n1 : 1.16 275.25 17.20 0.00 0.00 218346.31 12233.39 253211.69 00:31:37.166 Job: Nvme5n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:37.166 Verification LBA range: start 0x0 length 0x400 00:31:37.166 Nvme5n1 : 1.15 221.69 13.86 0.00 0.00 267722.15 20971.52 253211.69 00:31:37.166 Job: Nvme6n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:37.166 Verification LBA range: start 0x0 length 0x400 00:31:37.166 Nvme6n1 : 1.16 220.96 13.81 0.00 0.00 264054.14 21068.61 256318.58 00:31:37.166 Job: Nvme7n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:37.166 Verification LBA range: start 0x0 length 0x400 00:31:37.166 Nvme7n1 : 1.15 225.85 14.12 0.00 0.00 253131.03 2560.76 248551.35 00:31:37.166 Job: Nvme8n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:37.166 Verification LBA range: start 0x0 length 0x400 00:31:37.166 Nvme8n1 : 1.17 276.01 17.25 0.00 0.00 204410.89 10097.40 245444.46 00:31:37.166 Job: Nvme9n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:37.166 Verification LBA range: start 0x0 length 0x400 00:31:37.166 Nvme9n1 : 1.17 218.79 13.67 0.00 0.00 253309.53 22427.88 279620.27 00:31:37.166 Job: Nvme10n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:37.166 Verification LBA range: start 0x0 length 0x400 00:31:37.166 Nvme10n1 : 1.17 219.32 13.71 0.00 0.00 248214.19 38641.97 265639.25 00:31:37.166 [2024-12-06T16:47:19.005Z] =================================================================================================================== 00:31:37.166 [2024-12-06T16:47:19.005Z] Total : 2341.20 146.33 0.00 0.00 250381.95 2560.76 279620.27 00:31:37.424 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@95 -- # stoptarget 00:31:37.424 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@42 -- # rm -f ./local-job0-0-verify.state 00:31:37.424 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@43 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:31:37.424 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@44 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:31:37.424 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@46 -- # nvmftestfini 00:31:37.424 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@516 -- # nvmfcleanup 00:31:37.424 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@121 -- # sync 00:31:37.424 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:31:37.425 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@124 -- # set +e 00:31:37.425 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@125 -- # for i in {1..20} 00:31:37.425 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:31:37.425 rmmod nvme_tcp 00:31:37.425 rmmod nvme_fabrics 00:31:37.425 rmmod nvme_keyring 00:31:37.425 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:31:37.425 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@128 -- # set -e 00:31:37.425 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@129 -- # return 0 00:31:37.425 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@517 -- # '[' -n 334725 ']' 00:31:37.425 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@518 -- # killprocess 334725 00:31:37.425 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@954 -- # '[' -z 334725 ']' 00:31:37.425 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@958 -- # kill -0 334725 00:31:37.425 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@959 -- # uname 00:31:37.425 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:31:37.425 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 334725 00:31:37.425 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:31:37.425 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:31:37.425 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@972 -- # echo 'killing process with pid 334725' 00:31:37.425 killing process with pid 334725 00:31:37.425 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@973 -- # kill 334725 00:31:37.425 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@978 -- # wait 334725 00:31:37.993 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:31:37.993 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:31:37.993 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:31:37.993 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@297 -- # iptr 00:31:37.993 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@791 -- # iptables-save 00:31:37.993 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:31:37.993 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@791 -- # iptables-restore 00:31:37.993 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:31:37.993 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@302 -- # remove_spdk_ns 00:31:37.993 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:31:37.993 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:31:37.993 17:47:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:31:39.907 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:31:39.907 00:31:39.907 real 0m11.735s 00:31:39.907 user 0m33.855s 00:31:39.907 sys 0m3.193s 00:31:39.907 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@1130 -- # xtrace_disable 00:31:39.907 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:31:39.907 ************************************ 00:31:39.907 END TEST nvmf_shutdown_tc1 00:31:39.907 ************************************ 00:31:39.907 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@163 -- # run_test nvmf_shutdown_tc2 nvmf_shutdown_tc2 00:31:39.907 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:31:39.907 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1111 -- # xtrace_disable 00:31:39.907 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@10 -- # set +x 00:31:39.907 ************************************ 00:31:39.907 START TEST nvmf_shutdown_tc2 00:31:39.907 ************************************ 00:31:39.907 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@1129 -- # nvmf_shutdown_tc2 00:31:39.907 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@100 -- # starttarget 00:31:39.907 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@16 -- # nvmftestinit 00:31:39.907 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:31:39.907 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:31:39.907 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@476 -- # prepare_net_devs 00:31:39.907 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@438 -- # local -g is_hw=no 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@440 -- # remove_spdk_ns 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@309 -- # xtrace_disable 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@315 -- # pci_devs=() 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@315 -- # local -a pci_devs 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@316 -- # pci_net_devs=() 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@317 -- # pci_drivers=() 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@317 -- # local -A pci_drivers 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@319 -- # net_devs=() 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@319 -- # local -ga net_devs 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@320 -- # e810=() 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@320 -- # local -ga e810 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@321 -- # x722=() 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@321 -- # local -ga x722 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@322 -- # mlx=() 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@322 -- # local -ga mlx 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:31:39.908 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:31:39.908 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@418 -- # [[ up == up ]] 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:31:39.908 Found net devices under 0000:0a:00.0: cvl_0_0 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@418 -- # [[ up == up ]] 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:31:39.908 Found net devices under 0000:0a:00.1: cvl_0_1 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@442 -- # is_hw=yes 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:31:39.908 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:31:40.166 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:31:40.166 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:31:40.166 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:31:40.166 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:31:40.166 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:31:40.166 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:31:40.166 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:31:40.166 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:31:40.166 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:31:40.166 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.168 ms 00:31:40.166 00:31:40.166 --- 10.0.0.2 ping statistics --- 00:31:40.166 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:31:40.166 rtt min/avg/max/mdev = 0.168/0.168/0.168/0.000 ms 00:31:40.166 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:31:40.166 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:31:40.166 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.184 ms 00:31:40.166 00:31:40.166 --- 10.0.0.1 ping statistics --- 00:31:40.166 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:31:40.166 rtt min/avg/max/mdev = 0.184/0.184/0.184/0.000 ms 00:31:40.167 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:31:40.167 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@450 -- # return 0 00:31:40.167 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:31:40.167 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:31:40.167 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:31:40.167 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:31:40.167 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:31:40.167 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:31:40.167 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:31:40.167 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@19 -- # nvmfappstart -m 0x1E 00:31:40.167 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:31:40.167 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@726 -- # xtrace_disable 00:31:40.167 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:31:40.167 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@509 -- # nvmfpid=336080 00:31:40.167 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1E 00:31:40.167 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@510 -- # waitforlisten 336080 00:31:40.167 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@835 -- # '[' -z 336080 ']' 00:31:40.167 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:31:40.167 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@840 -- # local max_retries=100 00:31:40.167 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:31:40.167 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:31:40.167 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@844 -- # xtrace_disable 00:31:40.167 17:47:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:31:40.167 [2024-12-06 17:47:21.910518] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:31:40.167 [2024-12-06 17:47:21.910588] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:31:40.167 [2024-12-06 17:47:21.980026] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:31:40.424 [2024-12-06 17:47:22.025325] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:31:40.424 [2024-12-06 17:47:22.025380] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:31:40.424 [2024-12-06 17:47:22.025393] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:31:40.424 [2024-12-06 17:47:22.025405] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:31:40.424 [2024-12-06 17:47:22.025414] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:31:40.424 [2024-12-06 17:47:22.026868] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:31:40.424 [2024-12-06 17:47:22.026932] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:31:40.424 [2024-12-06 17:47:22.027066] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:31:40.424 [2024-12-06 17:47:22.027070] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@868 -- # return 0 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@732 -- # xtrace_disable 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@21 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:31:40.424 [2024-12-06 17:47:22.221290] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@23 -- # num_subsystems=({1..10}) 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@25 -- # timing_enter create_subsystems 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@726 -- # xtrace_disable 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@27 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:40.424 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:31:40.681 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@36 -- # rpc_cmd 00:31:40.681 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:40.681 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:31:40.681 Malloc1 00:31:40.681 [2024-12-06 17:47:22.328161] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:31:40.681 Malloc2 00:31:40.681 Malloc3 00:31:40.681 Malloc4 00:31:40.681 Malloc5 00:31:40.938 Malloc6 00:31:40.938 Malloc7 00:31:40.938 Malloc8 00:31:40.938 Malloc9 00:31:40.938 Malloc10 00:31:40.938 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:40.938 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@37 -- # timing_exit create_subsystems 00:31:40.938 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@732 -- # xtrace_disable 00:31:40.938 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:31:41.194 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@104 -- # perfpid=336260 00:31:41.194 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@105 -- # waitforlisten 336260 /var/tmp/bdevperf.sock 00:31:41.194 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@835 -- # '[' -z 336260 ']' 00:31:41.194 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@103 -- # gen_nvmf_target_json 1 2 3 4 5 6 7 8 9 10 00:31:41.194 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@103 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock --json /dev/fd/63 -q 64 -o 65536 -w verify -t 10 00:31:41.194 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:31:41.194 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@840 -- # local max_retries=100 00:31:41.194 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@560 -- # config=() 00:31:41.194 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:31:41.194 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:31:41.194 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@560 -- # local subsystem config 00:31:41.194 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@844 -- # xtrace_disable 00:31:41.194 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:41.194 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:31:41.194 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:41.194 { 00:31:41.194 "params": { 00:31:41.194 "name": "Nvme$subsystem", 00:31:41.194 "trtype": "$TEST_TRANSPORT", 00:31:41.194 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:41.194 "adrfam": "ipv4", 00:31:41.194 "trsvcid": "$NVMF_PORT", 00:31:41.194 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:41.194 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:41.194 "hdgst": ${hdgst:-false}, 00:31:41.194 "ddgst": ${ddgst:-false} 00:31:41.194 }, 00:31:41.195 "method": "bdev_nvme_attach_controller" 00:31:41.195 } 00:31:41.195 EOF 00:31:41.195 )") 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # cat 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:41.195 { 00:31:41.195 "params": { 00:31:41.195 "name": "Nvme$subsystem", 00:31:41.195 "trtype": "$TEST_TRANSPORT", 00:31:41.195 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:41.195 "adrfam": "ipv4", 00:31:41.195 "trsvcid": "$NVMF_PORT", 00:31:41.195 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:41.195 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:41.195 "hdgst": ${hdgst:-false}, 00:31:41.195 "ddgst": ${ddgst:-false} 00:31:41.195 }, 00:31:41.195 "method": "bdev_nvme_attach_controller" 00:31:41.195 } 00:31:41.195 EOF 00:31:41.195 )") 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # cat 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:41.195 { 00:31:41.195 "params": { 00:31:41.195 "name": "Nvme$subsystem", 00:31:41.195 "trtype": "$TEST_TRANSPORT", 00:31:41.195 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:41.195 "adrfam": "ipv4", 00:31:41.195 "trsvcid": "$NVMF_PORT", 00:31:41.195 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:41.195 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:41.195 "hdgst": ${hdgst:-false}, 00:31:41.195 "ddgst": ${ddgst:-false} 00:31:41.195 }, 00:31:41.195 "method": "bdev_nvme_attach_controller" 00:31:41.195 } 00:31:41.195 EOF 00:31:41.195 )") 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # cat 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:41.195 { 00:31:41.195 "params": { 00:31:41.195 "name": "Nvme$subsystem", 00:31:41.195 "trtype": "$TEST_TRANSPORT", 00:31:41.195 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:41.195 "adrfam": "ipv4", 00:31:41.195 "trsvcid": "$NVMF_PORT", 00:31:41.195 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:41.195 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:41.195 "hdgst": ${hdgst:-false}, 00:31:41.195 "ddgst": ${ddgst:-false} 00:31:41.195 }, 00:31:41.195 "method": "bdev_nvme_attach_controller" 00:31:41.195 } 00:31:41.195 EOF 00:31:41.195 )") 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # cat 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:41.195 { 00:31:41.195 "params": { 00:31:41.195 "name": "Nvme$subsystem", 00:31:41.195 "trtype": "$TEST_TRANSPORT", 00:31:41.195 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:41.195 "adrfam": "ipv4", 00:31:41.195 "trsvcid": "$NVMF_PORT", 00:31:41.195 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:41.195 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:41.195 "hdgst": ${hdgst:-false}, 00:31:41.195 "ddgst": ${ddgst:-false} 00:31:41.195 }, 00:31:41.195 "method": "bdev_nvme_attach_controller" 00:31:41.195 } 00:31:41.195 EOF 00:31:41.195 )") 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # cat 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:41.195 { 00:31:41.195 "params": { 00:31:41.195 "name": "Nvme$subsystem", 00:31:41.195 "trtype": "$TEST_TRANSPORT", 00:31:41.195 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:41.195 "adrfam": "ipv4", 00:31:41.195 "trsvcid": "$NVMF_PORT", 00:31:41.195 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:41.195 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:41.195 "hdgst": ${hdgst:-false}, 00:31:41.195 "ddgst": ${ddgst:-false} 00:31:41.195 }, 00:31:41.195 "method": "bdev_nvme_attach_controller" 00:31:41.195 } 00:31:41.195 EOF 00:31:41.195 )") 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # cat 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:41.195 { 00:31:41.195 "params": { 00:31:41.195 "name": "Nvme$subsystem", 00:31:41.195 "trtype": "$TEST_TRANSPORT", 00:31:41.195 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:41.195 "adrfam": "ipv4", 00:31:41.195 "trsvcid": "$NVMF_PORT", 00:31:41.195 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:41.195 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:41.195 "hdgst": ${hdgst:-false}, 00:31:41.195 "ddgst": ${ddgst:-false} 00:31:41.195 }, 00:31:41.195 "method": "bdev_nvme_attach_controller" 00:31:41.195 } 00:31:41.195 EOF 00:31:41.195 )") 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # cat 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:41.195 { 00:31:41.195 "params": { 00:31:41.195 "name": "Nvme$subsystem", 00:31:41.195 "trtype": "$TEST_TRANSPORT", 00:31:41.195 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:41.195 "adrfam": "ipv4", 00:31:41.195 "trsvcid": "$NVMF_PORT", 00:31:41.195 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:41.195 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:41.195 "hdgst": ${hdgst:-false}, 00:31:41.195 "ddgst": ${ddgst:-false} 00:31:41.195 }, 00:31:41.195 "method": "bdev_nvme_attach_controller" 00:31:41.195 } 00:31:41.195 EOF 00:31:41.195 )") 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # cat 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:41.195 { 00:31:41.195 "params": { 00:31:41.195 "name": "Nvme$subsystem", 00:31:41.195 "trtype": "$TEST_TRANSPORT", 00:31:41.195 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:41.195 "adrfam": "ipv4", 00:31:41.195 "trsvcid": "$NVMF_PORT", 00:31:41.195 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:41.195 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:41.195 "hdgst": ${hdgst:-false}, 00:31:41.195 "ddgst": ${ddgst:-false} 00:31:41.195 }, 00:31:41.195 "method": "bdev_nvme_attach_controller" 00:31:41.195 } 00:31:41.195 EOF 00:31:41.195 )") 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # cat 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:41.195 { 00:31:41.195 "params": { 00:31:41.195 "name": "Nvme$subsystem", 00:31:41.195 "trtype": "$TEST_TRANSPORT", 00:31:41.195 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:41.195 "adrfam": "ipv4", 00:31:41.195 "trsvcid": "$NVMF_PORT", 00:31:41.195 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:41.195 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:41.195 "hdgst": ${hdgst:-false}, 00:31:41.195 "ddgst": ${ddgst:-false} 00:31:41.195 }, 00:31:41.195 "method": "bdev_nvme_attach_controller" 00:31:41.195 } 00:31:41.195 EOF 00:31:41.195 )") 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # cat 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@584 -- # jq . 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@585 -- # IFS=, 00:31:41.195 17:47:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:31:41.195 "params": { 00:31:41.195 "name": "Nvme1", 00:31:41.195 "trtype": "tcp", 00:31:41.195 "traddr": "10.0.0.2", 00:31:41.195 "adrfam": "ipv4", 00:31:41.195 "trsvcid": "4420", 00:31:41.195 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:31:41.195 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:31:41.195 "hdgst": false, 00:31:41.195 "ddgst": false 00:31:41.195 }, 00:31:41.195 "method": "bdev_nvme_attach_controller" 00:31:41.195 },{ 00:31:41.195 "params": { 00:31:41.195 "name": "Nvme2", 00:31:41.195 "trtype": "tcp", 00:31:41.195 "traddr": "10.0.0.2", 00:31:41.195 "adrfam": "ipv4", 00:31:41.195 "trsvcid": "4420", 00:31:41.195 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:31:41.195 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:31:41.195 "hdgst": false, 00:31:41.195 "ddgst": false 00:31:41.195 }, 00:31:41.195 "method": "bdev_nvme_attach_controller" 00:31:41.195 },{ 00:31:41.195 "params": { 00:31:41.195 "name": "Nvme3", 00:31:41.195 "trtype": "tcp", 00:31:41.195 "traddr": "10.0.0.2", 00:31:41.195 "adrfam": "ipv4", 00:31:41.195 "trsvcid": "4420", 00:31:41.195 "subnqn": "nqn.2016-06.io.spdk:cnode3", 00:31:41.195 "hostnqn": "nqn.2016-06.io.spdk:host3", 00:31:41.195 "hdgst": false, 00:31:41.195 "ddgst": false 00:31:41.195 }, 00:31:41.195 "method": "bdev_nvme_attach_controller" 00:31:41.195 },{ 00:31:41.195 "params": { 00:31:41.195 "name": "Nvme4", 00:31:41.195 "trtype": "tcp", 00:31:41.195 "traddr": "10.0.0.2", 00:31:41.195 "adrfam": "ipv4", 00:31:41.195 "trsvcid": "4420", 00:31:41.195 "subnqn": "nqn.2016-06.io.spdk:cnode4", 00:31:41.195 "hostnqn": "nqn.2016-06.io.spdk:host4", 00:31:41.195 "hdgst": false, 00:31:41.195 "ddgst": false 00:31:41.195 }, 00:31:41.195 "method": "bdev_nvme_attach_controller" 00:31:41.195 },{ 00:31:41.195 "params": { 00:31:41.195 "name": "Nvme5", 00:31:41.195 "trtype": "tcp", 00:31:41.195 "traddr": "10.0.0.2", 00:31:41.195 "adrfam": "ipv4", 00:31:41.195 "trsvcid": "4420", 00:31:41.195 "subnqn": "nqn.2016-06.io.spdk:cnode5", 00:31:41.195 "hostnqn": "nqn.2016-06.io.spdk:host5", 00:31:41.195 "hdgst": false, 00:31:41.195 "ddgst": false 00:31:41.195 }, 00:31:41.195 "method": "bdev_nvme_attach_controller" 00:31:41.195 },{ 00:31:41.195 "params": { 00:31:41.195 "name": "Nvme6", 00:31:41.195 "trtype": "tcp", 00:31:41.195 "traddr": "10.0.0.2", 00:31:41.195 "adrfam": "ipv4", 00:31:41.195 "trsvcid": "4420", 00:31:41.195 "subnqn": "nqn.2016-06.io.spdk:cnode6", 00:31:41.195 "hostnqn": "nqn.2016-06.io.spdk:host6", 00:31:41.195 "hdgst": false, 00:31:41.195 "ddgst": false 00:31:41.195 }, 00:31:41.195 "method": "bdev_nvme_attach_controller" 00:31:41.195 },{ 00:31:41.195 "params": { 00:31:41.195 "name": "Nvme7", 00:31:41.195 "trtype": "tcp", 00:31:41.195 "traddr": "10.0.0.2", 00:31:41.195 "adrfam": "ipv4", 00:31:41.195 "trsvcid": "4420", 00:31:41.195 "subnqn": "nqn.2016-06.io.spdk:cnode7", 00:31:41.195 "hostnqn": "nqn.2016-06.io.spdk:host7", 00:31:41.195 "hdgst": false, 00:31:41.195 "ddgst": false 00:31:41.195 }, 00:31:41.195 "method": "bdev_nvme_attach_controller" 00:31:41.195 },{ 00:31:41.195 "params": { 00:31:41.195 "name": "Nvme8", 00:31:41.195 "trtype": "tcp", 00:31:41.195 "traddr": "10.0.0.2", 00:31:41.195 "adrfam": "ipv4", 00:31:41.195 "trsvcid": "4420", 00:31:41.195 "subnqn": "nqn.2016-06.io.spdk:cnode8", 00:31:41.195 "hostnqn": "nqn.2016-06.io.spdk:host8", 00:31:41.195 "hdgst": false, 00:31:41.195 "ddgst": false 00:31:41.195 }, 00:31:41.195 "method": "bdev_nvme_attach_controller" 00:31:41.195 },{ 00:31:41.195 "params": { 00:31:41.195 "name": "Nvme9", 00:31:41.195 "trtype": "tcp", 00:31:41.195 "traddr": "10.0.0.2", 00:31:41.195 "adrfam": "ipv4", 00:31:41.195 "trsvcid": "4420", 00:31:41.195 "subnqn": "nqn.2016-06.io.spdk:cnode9", 00:31:41.195 "hostnqn": "nqn.2016-06.io.spdk:host9", 00:31:41.195 "hdgst": false, 00:31:41.195 "ddgst": false 00:31:41.195 }, 00:31:41.195 "method": "bdev_nvme_attach_controller" 00:31:41.195 },{ 00:31:41.195 "params": { 00:31:41.195 "name": "Nvme10", 00:31:41.195 "trtype": "tcp", 00:31:41.195 "traddr": "10.0.0.2", 00:31:41.195 "adrfam": "ipv4", 00:31:41.195 "trsvcid": "4420", 00:31:41.195 "subnqn": "nqn.2016-06.io.spdk:cnode10", 00:31:41.195 "hostnqn": "nqn.2016-06.io.spdk:host10", 00:31:41.195 "hdgst": false, 00:31:41.195 "ddgst": false 00:31:41.195 }, 00:31:41.195 "method": "bdev_nvme_attach_controller" 00:31:41.195 }' 00:31:41.195 [2024-12-06 17:47:22.823988] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:31:41.195 [2024-12-06 17:47:22.824060] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid336260 ] 00:31:41.195 [2024-12-06 17:47:22.896569] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:31:41.195 [2024-12-06 17:47:22.943650] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:31:43.087 Running I/O for 10 seconds... 00:31:43.087 17:47:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:31:43.087 17:47:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@868 -- # return 0 00:31:43.087 17:47:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@106 -- # rpc_cmd -s /var/tmp/bdevperf.sock framework_wait_init 00:31:43.087 17:47:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:43.087 17:47:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:31:43.087 17:47:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:43.087 17:47:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@108 -- # waitforio /var/tmp/bdevperf.sock Nvme1n1 00:31:43.087 17:47:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@51 -- # '[' -z /var/tmp/bdevperf.sock ']' 00:31:43.087 17:47:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@55 -- # '[' -z Nvme1n1 ']' 00:31:43.087 17:47:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@58 -- # local ret=1 00:31:43.087 17:47:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@59 -- # local i 00:31:43.087 17:47:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@60 -- # (( i = 10 )) 00:31:43.087 17:47:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@60 -- # (( i != 0 )) 00:31:43.087 17:47:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:31:43.087 17:47:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # jq -r '.bdevs[0].num_read_ops' 00:31:43.087 17:47:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:43.087 17:47:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:31:43.087 17:47:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:43.087 17:47:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # read_io_count=3 00:31:43.087 17:47:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@64 -- # '[' 3 -ge 100 ']' 00:31:43.087 17:47:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@68 -- # sleep 0.25 00:31:43.343 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@60 -- # (( i-- )) 00:31:43.343 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@60 -- # (( i != 0 )) 00:31:43.343 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:31:43.343 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # jq -r '.bdevs[0].num_read_ops' 00:31:43.343 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:43.343 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:31:43.600 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:43.600 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # read_io_count=67 00:31:43.600 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@64 -- # '[' 67 -ge 100 ']' 00:31:43.600 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@68 -- # sleep 0.25 00:31:43.858 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@60 -- # (( i-- )) 00:31:43.858 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@60 -- # (( i != 0 )) 00:31:43.858 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:31:43.858 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # jq -r '.bdevs[0].num_read_ops' 00:31:43.858 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:43.858 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:31:43.858 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:43.858 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # read_io_count=131 00:31:43.858 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@64 -- # '[' 131 -ge 100 ']' 00:31:43.858 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@65 -- # ret=0 00:31:43.858 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@66 -- # break 00:31:43.858 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@70 -- # return 0 00:31:43.858 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@111 -- # killprocess 336260 00:31:43.858 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@954 -- # '[' -z 336260 ']' 00:31:43.858 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@958 -- # kill -0 336260 00:31:43.858 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@959 -- # uname 00:31:43.858 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:31:43.858 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 336260 00:31:43.858 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:31:43.858 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:31:43.858 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@972 -- # echo 'killing process with pid 336260' 00:31:43.858 killing process with pid 336260 00:31:43.858 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@973 -- # kill 336260 00:31:43.858 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@978 -- # wait 336260 00:31:43.858 Received shutdown signal, test time was about 0.948375 seconds 00:31:43.858 00:31:43.858 Latency(us) 00:31:43.858 [2024-12-06T16:47:25.697Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:31:43.858 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:43.858 Verification LBA range: start 0x0 length 0x400 00:31:43.858 Nvme1n1 : 0.94 271.16 16.95 0.00 0.00 232967.21 29127.11 239230.67 00:31:43.858 Job: Nvme2n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:43.858 Verification LBA range: start 0x0 length 0x400 00:31:43.858 Nvme2n1 : 0.95 270.16 16.89 0.00 0.00 228811.09 18350.08 234570.33 00:31:43.858 Job: Nvme3n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:43.858 Verification LBA range: start 0x0 length 0x400 00:31:43.858 Nvme3n1 : 0.94 271.90 16.99 0.00 0.00 223243.00 18350.08 248551.35 00:31:43.858 Job: Nvme4n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:43.858 Verification LBA range: start 0x0 length 0x400 00:31:43.858 Nvme4n1 : 0.94 273.26 17.08 0.00 0.00 217355.38 27767.85 242337.56 00:31:43.858 Job: Nvme5n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:43.858 Verification LBA range: start 0x0 length 0x400 00:31:43.858 Nvme5n1 : 0.92 209.58 13.10 0.00 0.00 276725.00 21651.15 234570.33 00:31:43.858 Job: Nvme6n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:43.858 Verification LBA range: start 0x0 length 0x400 00:31:43.858 Nvme6n1 : 0.93 206.81 12.93 0.00 0.00 274720.49 21262.79 265639.25 00:31:43.858 Job: Nvme7n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:43.858 Verification LBA range: start 0x0 length 0x400 00:31:43.858 Nvme7n1 : 0.91 211.52 13.22 0.00 0.00 261638.57 21942.42 246997.90 00:31:43.858 Job: Nvme8n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:43.858 Verification LBA range: start 0x0 length 0x400 00:31:43.858 Nvme8n1 : 0.89 214.81 13.43 0.00 0.00 251164.95 16990.81 253211.69 00:31:43.858 Job: Nvme9n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:43.858 Verification LBA range: start 0x0 length 0x400 00:31:43.858 Nvme9n1 : 0.93 205.91 12.87 0.00 0.00 258196.42 18932.62 284280.60 00:31:43.858 Job: Nvme10n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:43.858 Verification LBA range: start 0x0 length 0x400 00:31:43.858 Nvme10n1 : 0.92 208.53 13.03 0.00 0.00 248311.40 19029.71 253211.69 00:31:43.858 [2024-12-06T16:47:25.697Z] =================================================================================================================== 00:31:43.858 [2024-12-06T16:47:25.697Z] Total : 2343.64 146.48 0.00 0.00 244758.15 16990.81 284280.60 00:31:44.114 17:47:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@114 -- # sleep 1 00:31:45.061 17:47:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@115 -- # kill -0 336080 00:31:45.061 17:47:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@117 -- # stoptarget 00:31:45.061 17:47:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@42 -- # rm -f ./local-job0-0-verify.state 00:31:45.061 17:47:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@43 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:31:45.061 17:47:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@44 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:31:45.061 17:47:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@46 -- # nvmftestfini 00:31:45.061 17:47:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@516 -- # nvmfcleanup 00:31:45.061 17:47:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@121 -- # sync 00:31:45.061 17:47:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:31:45.061 17:47:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@124 -- # set +e 00:31:45.061 17:47:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@125 -- # for i in {1..20} 00:31:45.061 17:47:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:31:45.061 rmmod nvme_tcp 00:31:45.061 rmmod nvme_fabrics 00:31:45.061 rmmod nvme_keyring 00:31:45.061 17:47:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:31:45.061 17:47:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@128 -- # set -e 00:31:45.061 17:47:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@129 -- # return 0 00:31:45.061 17:47:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@517 -- # '[' -n 336080 ']' 00:31:45.061 17:47:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@518 -- # killprocess 336080 00:31:45.061 17:47:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@954 -- # '[' -z 336080 ']' 00:31:45.061 17:47:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@958 -- # kill -0 336080 00:31:45.061 17:47:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@959 -- # uname 00:31:45.319 17:47:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:31:45.319 17:47:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 336080 00:31:45.319 17:47:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:31:45.319 17:47:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:31:45.319 17:47:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@972 -- # echo 'killing process with pid 336080' 00:31:45.319 killing process with pid 336080 00:31:45.319 17:47:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@973 -- # kill 336080 00:31:45.319 17:47:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@978 -- # wait 336080 00:31:45.578 17:47:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:31:45.578 17:47:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:31:45.578 17:47:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:31:45.578 17:47:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@297 -- # iptr 00:31:45.578 17:47:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@791 -- # iptables-save 00:31:45.578 17:47:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:31:45.578 17:47:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@791 -- # iptables-restore 00:31:45.578 17:47:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:31:45.578 17:47:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@302 -- # remove_spdk_ns 00:31:45.578 17:47:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:31:45.578 17:47:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:31:45.578 17:47:27 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:31:48.124 00:31:48.124 real 0m7.780s 00:31:48.124 user 0m24.142s 00:31:48.124 sys 0m1.463s 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@1130 -- # xtrace_disable 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:31:48.124 ************************************ 00:31:48.124 END TEST nvmf_shutdown_tc2 00:31:48.124 ************************************ 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@164 -- # run_test nvmf_shutdown_tc3 nvmf_shutdown_tc3 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1111 -- # xtrace_disable 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@10 -- # set +x 00:31:48.124 ************************************ 00:31:48.124 START TEST nvmf_shutdown_tc3 00:31:48.124 ************************************ 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@1129 -- # nvmf_shutdown_tc3 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@122 -- # starttarget 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@16 -- # nvmftestinit 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@476 -- # prepare_net_devs 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@438 -- # local -g is_hw=no 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@440 -- # remove_spdk_ns 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@309 -- # xtrace_disable 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@315 -- # pci_devs=() 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@315 -- # local -a pci_devs 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@316 -- # pci_net_devs=() 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@317 -- # pci_drivers=() 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@317 -- # local -A pci_drivers 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@319 -- # net_devs=() 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@319 -- # local -ga net_devs 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@320 -- # e810=() 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@320 -- # local -ga e810 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@321 -- # x722=() 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@321 -- # local -ga x722 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@322 -- # mlx=() 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@322 -- # local -ga mlx 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:31:48.124 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:31:48.124 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:31:48.124 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@418 -- # [[ up == up ]] 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:31:48.125 Found net devices under 0000:0a:00.0: cvl_0_0 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@418 -- # [[ up == up ]] 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:31:48.125 Found net devices under 0000:0a:00.1: cvl_0_1 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@442 -- # is_hw=yes 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:31:48.125 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:31:48.125 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.274 ms 00:31:48.125 00:31:48.125 --- 10.0.0.2 ping statistics --- 00:31:48.125 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:31:48.125 rtt min/avg/max/mdev = 0.274/0.274/0.274/0.000 ms 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:31:48.125 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:31:48.125 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.162 ms 00:31:48.125 00:31:48.125 --- 10.0.0.1 ping statistics --- 00:31:48.125 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:31:48.125 rtt min/avg/max/mdev = 0.162/0.162/0.162/0.000 ms 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@450 -- # return 0 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@19 -- # nvmfappstart -m 0x1E 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@726 -- # xtrace_disable 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@509 -- # nvmfpid=337174 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1E 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@510 -- # waitforlisten 337174 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@835 -- # '[' -z 337174 ']' 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@840 -- # local max_retries=100 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:31:48.125 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@844 -- # xtrace_disable 00:31:48.125 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:31:48.125 [2024-12-06 17:47:29.730364] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:31:48.125 [2024-12-06 17:47:29.730452] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:31:48.125 [2024-12-06 17:47:29.806042] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:31:48.125 [2024-12-06 17:47:29.855016] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:31:48.125 [2024-12-06 17:47:29.855076] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:31:48.126 [2024-12-06 17:47:29.855089] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:31:48.126 [2024-12-06 17:47:29.855101] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:31:48.126 [2024-12-06 17:47:29.855111] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:31:48.126 [2024-12-06 17:47:29.856714] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:31:48.126 [2024-12-06 17:47:29.856844] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:31:48.126 [2024-12-06 17:47:29.856867] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:31:48.126 [2024-12-06 17:47:29.856870] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:31:48.384 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:31:48.384 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@868 -- # return 0 00:31:48.384 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:31:48.384 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@732 -- # xtrace_disable 00:31:48.384 17:47:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@21 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:31:48.384 [2024-12-06 17:47:30.009228] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@23 -- # num_subsystems=({1..10}) 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@25 -- # timing_enter create_subsystems 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@726 -- # xtrace_disable 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@27 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@36 -- # rpc_cmd 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:48.384 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:31:48.384 Malloc1 00:31:48.384 [2024-12-06 17:47:30.109045] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:31:48.384 Malloc2 00:31:48.384 Malloc3 00:31:48.642 Malloc4 00:31:48.642 Malloc5 00:31:48.642 Malloc6 00:31:48.642 Malloc7 00:31:48.642 Malloc8 00:31:48.900 Malloc9 00:31:48.900 Malloc10 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@37 -- # timing_exit create_subsystems 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@732 -- # xtrace_disable 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@126 -- # perfpid=337355 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@127 -- # waitforlisten 337355 /var/tmp/bdevperf.sock 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@835 -- # '[' -z 337355 ']' 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@125 -- # gen_nvmf_target_json 1 2 3 4 5 6 7 8 9 10 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock --json /dev/fd/63 -q 64 -o 65536 -w verify -t 10 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@840 -- # local max_retries=100 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@560 -- # config=() 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:31:48.900 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@560 -- # local subsystem config 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@844 -- # xtrace_disable 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:48.900 { 00:31:48.900 "params": { 00:31:48.900 "name": "Nvme$subsystem", 00:31:48.900 "trtype": "$TEST_TRANSPORT", 00:31:48.900 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:48.900 "adrfam": "ipv4", 00:31:48.900 "trsvcid": "$NVMF_PORT", 00:31:48.900 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:48.900 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:48.900 "hdgst": ${hdgst:-false}, 00:31:48.900 "ddgst": ${ddgst:-false} 00:31:48.900 }, 00:31:48.900 "method": "bdev_nvme_attach_controller" 00:31:48.900 } 00:31:48.900 EOF 00:31:48.900 )") 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # cat 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:48.900 { 00:31:48.900 "params": { 00:31:48.900 "name": "Nvme$subsystem", 00:31:48.900 "trtype": "$TEST_TRANSPORT", 00:31:48.900 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:48.900 "adrfam": "ipv4", 00:31:48.900 "trsvcid": "$NVMF_PORT", 00:31:48.900 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:48.900 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:48.900 "hdgst": ${hdgst:-false}, 00:31:48.900 "ddgst": ${ddgst:-false} 00:31:48.900 }, 00:31:48.900 "method": "bdev_nvme_attach_controller" 00:31:48.900 } 00:31:48.900 EOF 00:31:48.900 )") 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # cat 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:48.900 { 00:31:48.900 "params": { 00:31:48.900 "name": "Nvme$subsystem", 00:31:48.900 "trtype": "$TEST_TRANSPORT", 00:31:48.900 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:48.900 "adrfam": "ipv4", 00:31:48.900 "trsvcid": "$NVMF_PORT", 00:31:48.900 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:48.900 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:48.900 "hdgst": ${hdgst:-false}, 00:31:48.900 "ddgst": ${ddgst:-false} 00:31:48.900 }, 00:31:48.900 "method": "bdev_nvme_attach_controller" 00:31:48.900 } 00:31:48.900 EOF 00:31:48.900 )") 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # cat 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:48.900 { 00:31:48.900 "params": { 00:31:48.900 "name": "Nvme$subsystem", 00:31:48.900 "trtype": "$TEST_TRANSPORT", 00:31:48.900 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:48.900 "adrfam": "ipv4", 00:31:48.900 "trsvcid": "$NVMF_PORT", 00:31:48.900 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:48.900 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:48.900 "hdgst": ${hdgst:-false}, 00:31:48.900 "ddgst": ${ddgst:-false} 00:31:48.900 }, 00:31:48.900 "method": "bdev_nvme_attach_controller" 00:31:48.900 } 00:31:48.900 EOF 00:31:48.900 )") 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # cat 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:48.900 { 00:31:48.900 "params": { 00:31:48.900 "name": "Nvme$subsystem", 00:31:48.900 "trtype": "$TEST_TRANSPORT", 00:31:48.900 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:48.900 "adrfam": "ipv4", 00:31:48.900 "trsvcid": "$NVMF_PORT", 00:31:48.900 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:48.900 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:48.900 "hdgst": ${hdgst:-false}, 00:31:48.900 "ddgst": ${ddgst:-false} 00:31:48.900 }, 00:31:48.900 "method": "bdev_nvme_attach_controller" 00:31:48.900 } 00:31:48.900 EOF 00:31:48.900 )") 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # cat 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:48.900 { 00:31:48.900 "params": { 00:31:48.900 "name": "Nvme$subsystem", 00:31:48.900 "trtype": "$TEST_TRANSPORT", 00:31:48.900 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:48.900 "adrfam": "ipv4", 00:31:48.900 "trsvcid": "$NVMF_PORT", 00:31:48.900 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:48.900 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:48.900 "hdgst": ${hdgst:-false}, 00:31:48.900 "ddgst": ${ddgst:-false} 00:31:48.900 }, 00:31:48.900 "method": "bdev_nvme_attach_controller" 00:31:48.900 } 00:31:48.900 EOF 00:31:48.900 )") 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # cat 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:48.900 { 00:31:48.900 "params": { 00:31:48.900 "name": "Nvme$subsystem", 00:31:48.900 "trtype": "$TEST_TRANSPORT", 00:31:48.900 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:48.900 "adrfam": "ipv4", 00:31:48.900 "trsvcid": "$NVMF_PORT", 00:31:48.900 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:48.900 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:48.900 "hdgst": ${hdgst:-false}, 00:31:48.900 "ddgst": ${ddgst:-false} 00:31:48.900 }, 00:31:48.900 "method": "bdev_nvme_attach_controller" 00:31:48.900 } 00:31:48.900 EOF 00:31:48.900 )") 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # cat 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:48.900 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:48.900 { 00:31:48.900 "params": { 00:31:48.900 "name": "Nvme$subsystem", 00:31:48.900 "trtype": "$TEST_TRANSPORT", 00:31:48.900 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:48.900 "adrfam": "ipv4", 00:31:48.900 "trsvcid": "$NVMF_PORT", 00:31:48.900 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:48.900 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:48.900 "hdgst": ${hdgst:-false}, 00:31:48.900 "ddgst": ${ddgst:-false} 00:31:48.901 }, 00:31:48.901 "method": "bdev_nvme_attach_controller" 00:31:48.901 } 00:31:48.901 EOF 00:31:48.901 )") 00:31:48.901 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # cat 00:31:48.901 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:48.901 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:48.901 { 00:31:48.901 "params": { 00:31:48.901 "name": "Nvme$subsystem", 00:31:48.901 "trtype": "$TEST_TRANSPORT", 00:31:48.901 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:48.901 "adrfam": "ipv4", 00:31:48.901 "trsvcid": "$NVMF_PORT", 00:31:48.901 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:48.901 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:48.901 "hdgst": ${hdgst:-false}, 00:31:48.901 "ddgst": ${ddgst:-false} 00:31:48.901 }, 00:31:48.901 "method": "bdev_nvme_attach_controller" 00:31:48.901 } 00:31:48.901 EOF 00:31:48.901 )") 00:31:48.901 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # cat 00:31:48.901 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:31:48.901 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:31:48.901 { 00:31:48.901 "params": { 00:31:48.901 "name": "Nvme$subsystem", 00:31:48.901 "trtype": "$TEST_TRANSPORT", 00:31:48.901 "traddr": "$NVMF_FIRST_TARGET_IP", 00:31:48.901 "adrfam": "ipv4", 00:31:48.901 "trsvcid": "$NVMF_PORT", 00:31:48.901 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:31:48.901 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:31:48.901 "hdgst": ${hdgst:-false}, 00:31:48.901 "ddgst": ${ddgst:-false} 00:31:48.901 }, 00:31:48.901 "method": "bdev_nvme_attach_controller" 00:31:48.901 } 00:31:48.901 EOF 00:31:48.901 )") 00:31:48.901 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # cat 00:31:48.901 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@584 -- # jq . 00:31:48.901 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@585 -- # IFS=, 00:31:48.901 17:47:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:31:48.901 "params": { 00:31:48.901 "name": "Nvme1", 00:31:48.901 "trtype": "tcp", 00:31:48.901 "traddr": "10.0.0.2", 00:31:48.901 "adrfam": "ipv4", 00:31:48.901 "trsvcid": "4420", 00:31:48.901 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:31:48.901 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:31:48.901 "hdgst": false, 00:31:48.901 "ddgst": false 00:31:48.901 }, 00:31:48.901 "method": "bdev_nvme_attach_controller" 00:31:48.901 },{ 00:31:48.901 "params": { 00:31:48.901 "name": "Nvme2", 00:31:48.901 "trtype": "tcp", 00:31:48.901 "traddr": "10.0.0.2", 00:31:48.901 "adrfam": "ipv4", 00:31:48.901 "trsvcid": "4420", 00:31:48.901 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:31:48.901 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:31:48.901 "hdgst": false, 00:31:48.901 "ddgst": false 00:31:48.901 }, 00:31:48.901 "method": "bdev_nvme_attach_controller" 00:31:48.901 },{ 00:31:48.901 "params": { 00:31:48.901 "name": "Nvme3", 00:31:48.901 "trtype": "tcp", 00:31:48.901 "traddr": "10.0.0.2", 00:31:48.901 "adrfam": "ipv4", 00:31:48.901 "trsvcid": "4420", 00:31:48.901 "subnqn": "nqn.2016-06.io.spdk:cnode3", 00:31:48.901 "hostnqn": "nqn.2016-06.io.spdk:host3", 00:31:48.901 "hdgst": false, 00:31:48.901 "ddgst": false 00:31:48.901 }, 00:31:48.901 "method": "bdev_nvme_attach_controller" 00:31:48.901 },{ 00:31:48.901 "params": { 00:31:48.901 "name": "Nvme4", 00:31:48.901 "trtype": "tcp", 00:31:48.901 "traddr": "10.0.0.2", 00:31:48.901 "adrfam": "ipv4", 00:31:48.901 "trsvcid": "4420", 00:31:48.901 "subnqn": "nqn.2016-06.io.spdk:cnode4", 00:31:48.901 "hostnqn": "nqn.2016-06.io.spdk:host4", 00:31:48.901 "hdgst": false, 00:31:48.901 "ddgst": false 00:31:48.901 }, 00:31:48.901 "method": "bdev_nvme_attach_controller" 00:31:48.901 },{ 00:31:48.901 "params": { 00:31:48.901 "name": "Nvme5", 00:31:48.901 "trtype": "tcp", 00:31:48.901 "traddr": "10.0.0.2", 00:31:48.901 "adrfam": "ipv4", 00:31:48.901 "trsvcid": "4420", 00:31:48.901 "subnqn": "nqn.2016-06.io.spdk:cnode5", 00:31:48.901 "hostnqn": "nqn.2016-06.io.spdk:host5", 00:31:48.901 "hdgst": false, 00:31:48.901 "ddgst": false 00:31:48.901 }, 00:31:48.901 "method": "bdev_nvme_attach_controller" 00:31:48.901 },{ 00:31:48.901 "params": { 00:31:48.901 "name": "Nvme6", 00:31:48.901 "trtype": "tcp", 00:31:48.901 "traddr": "10.0.0.2", 00:31:48.901 "adrfam": "ipv4", 00:31:48.901 "trsvcid": "4420", 00:31:48.901 "subnqn": "nqn.2016-06.io.spdk:cnode6", 00:31:48.901 "hostnqn": "nqn.2016-06.io.spdk:host6", 00:31:48.901 "hdgst": false, 00:31:48.901 "ddgst": false 00:31:48.901 }, 00:31:48.901 "method": "bdev_nvme_attach_controller" 00:31:48.901 },{ 00:31:48.901 "params": { 00:31:48.901 "name": "Nvme7", 00:31:48.901 "trtype": "tcp", 00:31:48.901 "traddr": "10.0.0.2", 00:31:48.901 "adrfam": "ipv4", 00:31:48.901 "trsvcid": "4420", 00:31:48.901 "subnqn": "nqn.2016-06.io.spdk:cnode7", 00:31:48.901 "hostnqn": "nqn.2016-06.io.spdk:host7", 00:31:48.901 "hdgst": false, 00:31:48.901 "ddgst": false 00:31:48.901 }, 00:31:48.901 "method": "bdev_nvme_attach_controller" 00:31:48.901 },{ 00:31:48.901 "params": { 00:31:48.901 "name": "Nvme8", 00:31:48.901 "trtype": "tcp", 00:31:48.901 "traddr": "10.0.0.2", 00:31:48.901 "adrfam": "ipv4", 00:31:48.901 "trsvcid": "4420", 00:31:48.901 "subnqn": "nqn.2016-06.io.spdk:cnode8", 00:31:48.901 "hostnqn": "nqn.2016-06.io.spdk:host8", 00:31:48.901 "hdgst": false, 00:31:48.901 "ddgst": false 00:31:48.901 }, 00:31:48.901 "method": "bdev_nvme_attach_controller" 00:31:48.901 },{ 00:31:48.901 "params": { 00:31:48.901 "name": "Nvme9", 00:31:48.901 "trtype": "tcp", 00:31:48.901 "traddr": "10.0.0.2", 00:31:48.901 "adrfam": "ipv4", 00:31:48.901 "trsvcid": "4420", 00:31:48.901 "subnqn": "nqn.2016-06.io.spdk:cnode9", 00:31:48.901 "hostnqn": "nqn.2016-06.io.spdk:host9", 00:31:48.901 "hdgst": false, 00:31:48.901 "ddgst": false 00:31:48.901 }, 00:31:48.901 "method": "bdev_nvme_attach_controller" 00:31:48.901 },{ 00:31:48.901 "params": { 00:31:48.901 "name": "Nvme10", 00:31:48.901 "trtype": "tcp", 00:31:48.901 "traddr": "10.0.0.2", 00:31:48.901 "adrfam": "ipv4", 00:31:48.901 "trsvcid": "4420", 00:31:48.901 "subnqn": "nqn.2016-06.io.spdk:cnode10", 00:31:48.901 "hostnqn": "nqn.2016-06.io.spdk:host10", 00:31:48.901 "hdgst": false, 00:31:48.901 "ddgst": false 00:31:48.901 }, 00:31:48.901 "method": "bdev_nvme_attach_controller" 00:31:48.901 }' 00:31:48.901 [2024-12-06 17:47:30.618896] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:31:48.901 [2024-12-06 17:47:30.618998] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid337355 ] 00:31:48.901 [2024-12-06 17:47:30.691914] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:31:49.159 [2024-12-06 17:47:30.739105] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:31:51.053 Running I/O for 10 seconds... 00:31:51.053 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:31:51.053 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@868 -- # return 0 00:31:51.053 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@128 -- # rpc_cmd -s /var/tmp/bdevperf.sock framework_wait_init 00:31:51.053 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:51.053 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:31:51.053 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:51.053 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@131 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; kill -9 $perfpid || true; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:31:51.053 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@133 -- # waitforio /var/tmp/bdevperf.sock Nvme1n1 00:31:51.053 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@51 -- # '[' -z /var/tmp/bdevperf.sock ']' 00:31:51.053 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@55 -- # '[' -z Nvme1n1 ']' 00:31:51.053 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@58 -- # local ret=1 00:31:51.053 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@59 -- # local i 00:31:51.053 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # (( i = 10 )) 00:31:51.053 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # (( i != 0 )) 00:31:51.053 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:31:51.053 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # jq -r '.bdevs[0].num_read_ops' 00:31:51.053 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:51.053 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:31:51.053 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:51.053 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # read_io_count=3 00:31:51.053 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@64 -- # '[' 3 -ge 100 ']' 00:31:51.053 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@68 -- # sleep 0.25 00:31:51.311 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # (( i-- )) 00:31:51.311 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # (( i != 0 )) 00:31:51.311 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:31:51.311 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # jq -r '.bdevs[0].num_read_ops' 00:31:51.312 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:51.312 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:31:51.312 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:51.312 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # read_io_count=67 00:31:51.312 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@64 -- # '[' 67 -ge 100 ']' 00:31:51.312 17:47:32 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@68 -- # sleep 0.25 00:31:51.588 17:47:33 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # (( i-- )) 00:31:51.588 17:47:33 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # (( i != 0 )) 00:31:51.588 17:47:33 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:31:51.588 17:47:33 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # jq -r '.bdevs[0].num_read_ops' 00:31:51.588 17:47:33 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:51.588 17:47:33 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:31:51.588 17:47:33 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:51.588 17:47:33 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # read_io_count=136 00:31:51.588 17:47:33 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@64 -- # '[' 136 -ge 100 ']' 00:31:51.588 17:47:33 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@65 -- # ret=0 00:31:51.588 17:47:33 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@66 -- # break 00:31:51.588 17:47:33 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@70 -- # return 0 00:31:51.588 17:47:33 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@136 -- # killprocess 337174 00:31:51.588 17:47:33 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@954 -- # '[' -z 337174 ']' 00:31:51.588 17:47:33 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@958 -- # kill -0 337174 00:31:51.588 17:47:33 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@959 -- # uname 00:31:51.588 17:47:33 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:31:51.588 17:47:33 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 337174 00:31:51.588 17:47:33 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:31:51.588 17:47:33 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:31:51.588 17:47:33 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@972 -- # echo 'killing process with pid 337174' 00:31:51.588 killing process with pid 337174 00:31:51.588 17:47:33 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@973 -- # kill 337174 00:31:51.588 17:47:33 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@978 -- # wait 337174 00:31:51.588 [2024-12-06 17:47:33.332775] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.588 [2024-12-06 17:47:33.332860] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.588 [2024-12-06 17:47:33.332901] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.588 [2024-12-06 17:47:33.332915] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.588 [2024-12-06 17:47:33.332927] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.588 [2024-12-06 17:47:33.332939] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.588 [2024-12-06 17:47:33.332951] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.588 [2024-12-06 17:47:33.332962] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.588 [2024-12-06 17:47:33.332993] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.588 [2024-12-06 17:47:33.333004] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.588 [2024-12-06 17:47:33.333016] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.588 [2024-12-06 17:47:33.333027] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.588 [2024-12-06 17:47:33.333039] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.588 [2024-12-06 17:47:33.333057] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.588 [2024-12-06 17:47:33.333068] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.588 [2024-12-06 17:47:33.333079] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.588 [2024-12-06 17:47:33.333091] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.588 [2024-12-06 17:47:33.333102] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.588 [2024-12-06 17:47:33.333114] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.588 [2024-12-06 17:47:33.333125] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.588 [2024-12-06 17:47:33.333136] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.588 [2024-12-06 17:47:33.333148] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.588 [2024-12-06 17:47:33.333160] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.588 [2024-12-06 17:47:33.333172] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.588 [2024-12-06 17:47:33.333183] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.588 [2024-12-06 17:47:33.333195] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.588 [2024-12-06 17:47:33.333207] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333218] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333230] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333241] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333256] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333269] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333281] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333298] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333310] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333322] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333333] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333345] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333356] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333367] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333378] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333389] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333399] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333411] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333422] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333432] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333443] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333454] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333465] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333476] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333487] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333498] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333508] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333520] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333531] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333542] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333554] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333568] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333580] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333591] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333602] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333613] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333624] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.333635] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9750 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335190] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335223] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335240] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335252] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335264] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335275] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335289] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335301] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335313] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335324] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335337] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335349] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335361] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335373] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335387] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335398] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335410] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335422] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335435] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335447] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335473] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335488] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335500] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335512] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335524] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335537] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335549] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335560] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335572] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335584] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335598] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335610] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335625] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335640] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335652] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335672] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335728] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335744] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335757] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335769] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335781] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335794] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335807] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335818] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335830] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335842] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335856] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.589 [2024-12-06 17:47:33.335868] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.335884] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.335899] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.335912] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.335923] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.335935] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.335947] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.335961] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.335980] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.335992] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.336021] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.336033] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.336044] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.336055] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.336066] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.336079] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbc320 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.337671] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdb9c40 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.340650] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.340698] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.340722] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.340735] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.340748] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.340759] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.340773] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.340785] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.340798] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.340810] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.340822] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.340847] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.340860] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.340873] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.340886] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.340898] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.340910] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.340923] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.340935] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.340974] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341090] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341111] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341124] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341137] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341150] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341162] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341174] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341186] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341198] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341210] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341222] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341234] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341246] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341259] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341273] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341285] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341298] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341311] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341328] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341342] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341355] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341368] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341380] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341393] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341405] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341418] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341430] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341442] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341455] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341467] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341479] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341491] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341503] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341515] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341543] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341554] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341566] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341578] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341590] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.590 [2024-12-06 17:47:33.341601] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.341613] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.341625] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.341636] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdba600 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.342601] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.342628] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.342647] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.342661] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.342697] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.342715] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.342727] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.342740] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.342753] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.342765] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.342777] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.342788] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.342801] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.342813] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.342825] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.342837] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.342849] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.342861] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.342873] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.342885] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.342897] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.342909] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.342921] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.342933] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.342945] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.342957] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.342980] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343007] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343020] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343032] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343048] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343060] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343073] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343091] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343104] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343116] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343127] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343138] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343150] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343162] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343174] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343186] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343198] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343209] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343221] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343248] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343261] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343273] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343285] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343297] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343310] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343322] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343334] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343346] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343358] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343371] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343382] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343398] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343411] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343422] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343434] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with t[2024-12-06 17:47:33.343416] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:26240 len:1he state(6) to be set 00:31:51.591 28 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.591 [2024-12-06 17:47:33.343451] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343463] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbaad0 is same with the state(6) to be set 00:31:51.591 [2024-12-06 17:47:33.343463] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.591 [2024-12-06 17:47:33.343496] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.591 [2024-12-06 17:47:33.343512] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.591 [2024-12-06 17:47:33.343528] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.591 [2024-12-06 17:47:33.343543] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.591 [2024-12-06 17:47:33.343559] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.591 [2024-12-06 17:47:33.343573] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.591 [2024-12-06 17:47:33.343588] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.591 [2024-12-06 17:47:33.343602] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.591 [2024-12-06 17:47:33.343617] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.591 [2024-12-06 17:47:33.343631] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.591 [2024-12-06 17:47:33.343647] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.591 [2024-12-06 17:47:33.343661] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.591 [2024-12-06 17:47:33.343687] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.591 [2024-12-06 17:47:33.343702] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.591 [2024-12-06 17:47:33.343721] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.343736] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.343752] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.343765] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.343789] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.343804] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.343819] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.343834] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.343849] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.343863] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.343878] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.343892] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.343907] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.343920] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.343936] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.343949] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.343965] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.343980] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.343995] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.344009] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.344025] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.344040] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.344055] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.344069] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.344085] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.344098] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.344115] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.344129] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.344144] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.344163] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.344179] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.344194] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.344210] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.344241] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.344257] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.344271] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.344286] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.344301] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.344316] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.344329] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.344345] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.344359] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.344374] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.344388] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.344403] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.344416] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.344432] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.344445] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.344461] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.344475] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.344491] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.344504] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.344520] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.344534] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.344552] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.344567] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.344582] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.344596] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.344612] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.344626] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.344641] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.344654] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.344693] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.344720] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.344725] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.592 [2024-12-06 17:47:33.344736] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.344751] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 c[2024-12-06 17:47:33.344750] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with tdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 he state(6) to be set 00:31:51.592 [2024-12-06 17:47:33.344767] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.592 [2024-12-06 17:47:33.344768] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.592 [2024-12-06 17:47:33.344780] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.592 [2024-12-06 17:47:33.344784] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.592 [2024-12-06 17:47:33.344792] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.592 [2024-12-06 17:47:33.344800] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.593 [2024-12-06 17:47:33.344805] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.344815] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.593 [2024-12-06 17:47:33.344817] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.344830] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.344831] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.593 [2024-12-06 17:47:33.344842] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.344850] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.593 [2024-12-06 17:47:33.344854] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.344867] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with t[2024-12-06 17:47:33.344867] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:31872 len:1he state(6) to be set 00:31:51.593 28 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.593 [2024-12-06 17:47:33.344881] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.344883] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.593 [2024-12-06 17:47:33.344893] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.344899] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.593 [2024-12-06 17:47:33.344906] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.344914] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.593 [2024-12-06 17:47:33.344918] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.344930] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:32128 len:1[2024-12-06 17:47:33.344931] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with t28 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.593 he state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.344946] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 c[2024-12-06 17:47:33.344946] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with tdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.593 he state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.344962] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.344964] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.593 [2024-12-06 17:47:33.345000] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with t[2024-12-06 17:47:33.345001] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 che state(6) to be set 00:31:51.593 dw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.593 [2024-12-06 17:47:33.345014] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.345019] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.593 [2024-12-06 17:47:33.345026] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.345033] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.593 [2024-12-06 17:47:33.345038] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.345049] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:32512 len:1[2024-12-06 17:47:33.345050] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with t28 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.593 he state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.345069] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 c[2024-12-06 17:47:33.345070] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with tdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.593 he state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.345084] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.345087] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.593 [2024-12-06 17:47:33.345096] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.345101] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.593 [2024-12-06 17:47:33.345107] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.345117] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24576 len:128[2024-12-06 17:47:33.345119] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with t SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.593 he state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.345133] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 c[2024-12-06 17:47:33.345133] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with tdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.593 he state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.345148] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.345151] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.593 [2024-12-06 17:47:33.345160] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.345166] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.593 [2024-12-06 17:47:33.345172] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.345183] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24832 len:128[2024-12-06 17:47:33.345185] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with t SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.593 he state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.345199] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.593 [2024-12-06 17:47:33.345209] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.345215] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.593 [2024-12-06 17:47:33.345221] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.345229] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.593 [2024-12-06 17:47:33.345233] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.345246] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with t[2024-12-06 17:47:33.345246] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25088 len:128he state(6) to be set 00:31:51.593 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.593 [2024-12-06 17:47:33.345263] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with t[2024-12-06 17:47:33.345265] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 che state(6) to be set 00:31:51.593 dw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.593 [2024-12-06 17:47:33.345278] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.345283] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.593 [2024-12-06 17:47:33.345291] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.345297] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.593 [2024-12-06 17:47:33.345303] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.345313] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128[2024-12-06 17:47:33.345315] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with t SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.593 he state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.345328] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with t[2024-12-06 17:47:33.345328] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 che state(6) to be set 00:31:51.593 dw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.593 [2024-12-06 17:47:33.345342] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.345347] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.593 [2024-12-06 17:47:33.345354] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.345361] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.593 [2024-12-06 17:47:33.345366] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.345377] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.593 [2024-12-06 17:47:33.345379] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.345392] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.593 [2024-12-06 17:47:33.345398] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 [2024-12-06 17:47:33.345403] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.594 [2024-12-06 17:47:33.345416] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with t[2024-12-06 17:47:33.345415] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128he state(6) to be set 00:31:51.594 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.594 [2024-12-06 17:47:33.345430] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.594 [2024-12-06 17:47:33.345433] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 [2024-12-06 17:47:33.345442] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.594 [2024-12-06 17:47:33.345449] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.594 [2024-12-06 17:47:33.345458] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.594 [2024-12-06 17:47:33.345463] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 [2024-12-06 17:47:33.345470] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.594 [2024-12-06 17:47:33.345479] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.594 [2024-12-06 17:47:33.345482] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.594 [2024-12-06 17:47:33.345493] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 c[2024-12-06 17:47:33.345494] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with tdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 he state(6) to be set 00:31:51.594 [2024-12-06 17:47:33.345507] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.594 [2024-12-06 17:47:33.345511] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.594 [2024-12-06 17:47:33.345519] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.594 [2024-12-06 17:47:33.345525] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 [2024-12-06 17:47:33.345531] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.594 [2024-12-06 17:47:33.345543] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.594 [2024-12-06 17:47:33.345554] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.594 [2024-12-06 17:47:33.345565] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with the state(6) to be set 00:31:51.594 [2024-12-06 17:47:33.345572] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] CQ [2024-12-06 17:47:33.345577] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbafa0 is same with ttransport error -6 (No such device or address) on qpair id 1 00:31:51.594 he state(6) to be set 00:31:51.594 [2024-12-06 17:47:33.346221] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.594 [2024-12-06 17:47:33.346247] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 [2024-12-06 17:47:33.346263] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.594 [2024-12-06 17:47:33.346277] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 [2024-12-06 17:47:33.346292] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.594 [2024-12-06 17:47:33.346306] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 [2024-12-06 17:47:33.346321] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.594 [2024-12-06 17:47:33.346334] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 [2024-12-06 17:47:33.346355] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x10f3710 is same with the state(6) to be set 00:31:51.594 [2024-12-06 17:47:33.346419] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.594 [2024-12-06 17:47:33.346440] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 [2024-12-06 17:47:33.346457] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.594 [2024-12-06 17:47:33.346471] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 [2024-12-06 17:47:33.346485] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.594 [2024-12-06 17:47:33.346498] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 [2024-12-06 17:47:33.346513] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.594 [2024-12-06 17:47:33.346526] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 [2024-12-06 17:47:33.346539] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc95950 is same with the state(6) to be set 00:31:51.594 [2024-12-06 17:47:33.346581] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.594 [2024-12-06 17:47:33.346608] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 [2024-12-06 17:47:33.346624] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.594 [2024-12-06 17:47:33.346638] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 [2024-12-06 17:47:33.346653] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.594 [2024-12-06 17:47:33.346676] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 [2024-12-06 17:47:33.346692] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.594 [2024-12-06 17:47:33.346706] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 [2024-12-06 17:47:33.346721] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc96080 is same with the state(6) to be set 00:31:51.594 [2024-12-06 17:47:33.346767] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.594 [2024-12-06 17:47:33.346787] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 [2024-12-06 17:47:33.346803] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.594 [2024-12-06 17:47:33.346822] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 [2024-12-06 17:47:33.346836] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.594 [2024-12-06 17:47:33.346849] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 [2024-12-06 17:47:33.346867] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.594 [2024-12-06 17:47:33.346882] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 [2024-12-06 17:47:33.346894] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x10c4680 is same with the state(6) to be set 00:31:51.594 [2024-12-06 17:47:33.346942] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.594 [2024-12-06 17:47:33.346962] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 [2024-12-06 17:47:33.346978] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.594 [2024-12-06 17:47:33.346991] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 [2024-12-06 17:47:33.347011] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.594 [2024-12-06 17:47:33.347026] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 [2024-12-06 17:47:33.347039] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.594 [2024-12-06 17:47:33.347053] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 [2024-12-06 17:47:33.347066] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x10eb9e0 is same with the state(6) to be set 00:31:51.594 [2024-12-06 17:47:33.347117] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.594 [2024-12-06 17:47:33.347137] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 [2024-12-06 17:47:33.347152] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.594 [2024-12-06 17:47:33.347165] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 [2024-12-06 17:47:33.347180] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.594 [2024-12-06 17:47:33.347196] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 [2024-12-06 17:47:33.347192] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.594 [2024-12-06 17:47:33.347212] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.594 [2024-12-06 17:47:33.347226] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 c[2024-12-06 17:47:33.347224] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with tdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.594 he state(6) to be set 00:31:51.594 [2024-12-06 17:47:33.347242] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xba4610 is same w[2024-12-06 17:47:33.347243] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with tith the state(6) to be set 00:31:51.594 he state(6) to be set 00:31:51.594 [2024-12-06 17:47:33.347258] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347270] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347289] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347291] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.595 [2024-12-06 17:47:33.347302] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347313] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 c[2024-12-06 17:47:33.347315] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with tdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.595 he state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347330] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347331] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.595 [2024-12-06 17:47:33.347342] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347346] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.595 [2024-12-06 17:47:33.347354] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347360] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.595 [2024-12-06 17:47:33.347366] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347373] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.595 [2024-12-06 17:47:33.347379] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347387] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.595 [2024-12-06 17:47:33.347391] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347401] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.595 [2024-12-06 17:47:33.347403] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347415] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347419] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc99630 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347428] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347440] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347453] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347465] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347464] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.595 [2024-12-06 17:47:33.347477] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347485] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 c[2024-12-06 17:47:33.347489] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with tdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.595 he state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347504] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347506] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.595 [2024-12-06 17:47:33.347516] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347520] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.595 [2024-12-06 17:47:33.347528] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347535] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.595 [2024-12-06 17:47:33.347541] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347548] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.595 [2024-12-06 17:47:33.347553] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347563] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.595 [2024-12-06 17:47:33.347565] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347576] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.595 [2024-12-06 17:47:33.347578] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347589] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8d630 is same w[2024-12-06 17:47:33.347590] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with tith the state(6) to be set 00:31:51.595 he state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347604] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347617] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347629] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347641] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347653] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347673] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347688] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347701] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347725] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347736] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24576 len:12[2024-12-06 17:47:33.347743] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with t8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.595 he state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347763] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347764] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.595 [2024-12-06 17:47:33.347776] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347786] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24704 len:12[2024-12-06 17:47:33.347788] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with t8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.595 he state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347802] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with t[2024-12-06 17:47:33.347802] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 che state(6) to be set 00:31:51.595 dw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.595 [2024-12-06 17:47:33.347816] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347820] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.595 [2024-12-06 17:47:33.347828] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347835] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.595 [2024-12-06 17:47:33.347841] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.595 [2024-12-06 17:47:33.347859] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347865] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.595 [2024-12-06 17:47:33.347872] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347880] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.595 [2024-12-06 17:47:33.347884] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347895] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 c[2024-12-06 17:47:33.347896] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with tdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.595 he state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347910] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347913] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.595 [2024-12-06 17:47:33.347922] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347927] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.595 [2024-12-06 17:47:33.347934] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347943] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:25344 len:12[2024-12-06 17:47:33.347946] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with t8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.595 he state(6) to be set 00:31:51.595 [2024-12-06 17:47:33.347973] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.596 [2024-12-06 17:47:33.347974] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.347984] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.596 [2024-12-06 17:47:33.347992] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.347997] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with the state(6) to be set 00:31:51.596 [2024-12-06 17:47:33.348007] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 c[2024-12-06 17:47:33.348009] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb470 is same with tdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 he state(6) to be set 00:31:51.596 [2024-12-06 17:47:33.348025] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348039] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348055] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348068] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348100] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348114] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348135] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348149] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348164] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348178] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348192] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348206] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348221] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348234] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348256] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348270] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348285] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348303] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348319] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348332] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348348] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348362] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348377] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348390] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348405] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348419] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348448] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348463] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348476] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348491] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348504] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348519] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348533] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348548] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348561] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348576] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348590] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348610] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348624] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348639] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348653] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348680] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348697] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348717] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348730] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348751] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348765] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348780] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348794] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348809] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348823] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348854] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348869] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348884] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348899] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348915] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348929] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348944] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348969] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.348984] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.348999] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.349013] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.349028] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.349044] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.349058] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.349073] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.349077] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.596 [2024-12-06 17:47:33.349090] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.349103] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.596 [2024-12-06 17:47:33.349107] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.596 [2024-12-06 17:47:33.349116] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.596 [2024-12-06 17:47:33.349121] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.596 [2024-12-06 17:47:33.349129] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349141] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349142] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.597 [2024-12-06 17:47:33.349154] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349157] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.597 [2024-12-06 17:47:33.349168] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349174] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.597 [2024-12-06 17:47:33.349180] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349188] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.597 [2024-12-06 17:47:33.349192] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349205] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with t[2024-12-06 17:47:33.349204] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:30336 len:1he state(6) to be set 00:31:51.597 28 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.597 [2024-12-06 17:47:33.349219] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349221] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.597 [2024-12-06 17:47:33.349232] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349238] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.597 [2024-12-06 17:47:33.349245] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349252] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.597 [2024-12-06 17:47:33.349257] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349271] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349273] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.597 [2024-12-06 17:47:33.349283] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349291] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.597 [2024-12-06 17:47:33.349297] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349308] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:30720 len:1[2024-12-06 17:47:33.349310] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with t28 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.597 he state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349324] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 c[2024-12-06 17:47:33.349324] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with tdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.597 he state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349340] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349342] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.597 [2024-12-06 17:47:33.349353] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349356] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.597 [2024-12-06 17:47:33.349366] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349372] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.597 [2024-12-06 17:47:33.349394] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349402] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.597 [2024-12-06 17:47:33.349407] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349417] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:31104 len:1[2024-12-06 17:47:33.349420] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with t28 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.597 he state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349433] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 c[2024-12-06 17:47:33.349434] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with tdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.597 he state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349449] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349450] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.597 [2024-12-06 17:47:33.349461] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349465] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.597 [2024-12-06 17:47:33.349474] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349480] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.597 [2024-12-06 17:47:33.349486] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349497] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 c[2024-12-06 17:47:33.349499] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with tdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.597 he state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349513] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349514] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.597 [2024-12-06 17:47:33.349525] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349529] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.597 [2024-12-06 17:47:33.349543] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with t[2024-12-06 17:47:33.349544] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:31616 len:1he state(6) to be set 00:31:51.597 28 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.597 [2024-12-06 17:47:33.349557] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with t[2024-12-06 17:47:33.349559] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 che state(6) to be set 00:31:51.597 dw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.597 [2024-12-06 17:47:33.349571] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349575] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.597 [2024-12-06 17:47:33.349583] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349589] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.597 [2024-12-06 17:47:33.349596] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349608] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349620] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349632] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349658] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349679] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349693] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349717] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349728] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349740] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349752] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349770] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.597 [2024-12-06 17:47:33.349783] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.349795] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.349807] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.349819] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.349831] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.349843] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.349855] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.349867] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.349879] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.349891] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.349903] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.349915] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.349927] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.349938] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbb960 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.350705] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.350743] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.350757] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.350770] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.350783] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.350795] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.350807] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.350819] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.350832] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.350844] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.350856] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.350868] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.350880] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.350897] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.350910] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.350922] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.350935] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.350947] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.350959] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.350990] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351002] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351014] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351026] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351038] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351050] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351061] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351073] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351086] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351097] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351109] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351121] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351133] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351144] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351163] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351175] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351187] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351198] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351209] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351221] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351232] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351247] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351260] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351271] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351282] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351294] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351305] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351317] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351328] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351340] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351351] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351363] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351374] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351386] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351398] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351409] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351421] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351432] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351444] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351461] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351473] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351484] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351495] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.351506] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdbbe30 is same with the state(6) to be set 00:31:51.598 [2024-12-06 17:47:33.368218] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.598 [2024-12-06 17:47:33.368304] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.598 [2024-12-06 17:47:33.368325] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.598 [2024-12-06 17:47:33.368340] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.598 [2024-12-06 17:47:33.368375] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.598 [2024-12-06 17:47:33.368391] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.598 [2024-12-06 17:47:33.368409] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.598 [2024-12-06 17:47:33.368424] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.599 [2024-12-06 17:47:33.368440] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.599 [2024-12-06 17:47:33.368454] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.599 [2024-12-06 17:47:33.368470] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.599 [2024-12-06 17:47:33.368485] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.599 [2024-12-06 17:47:33.368503] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.599 [2024-12-06 17:47:33.368517] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.599 [2024-12-06 17:47:33.370457] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode3, 1] resetting controller 00:31:51.599 [2024-12-06 17:47:33.370515] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xc95950 (9): Bad file descriptor 00:31:51.599 [2024-12-06 17:47:33.370607] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.599 [2024-12-06 17:47:33.370629] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.599 [2024-12-06 17:47:33.370645] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.599 [2024-12-06 17:47:33.370661] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.599 [2024-12-06 17:47:33.370691] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.599 [2024-12-06 17:47:33.370706] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.599 [2024-12-06 17:47:33.370721] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.599 [2024-12-06 17:47:33.370734] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.599 [2024-12-06 17:47:33.370747] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8ea40 is same with the state(6) to be set 00:31:51.599 [2024-12-06 17:47:33.370782] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x10f3710 (9): Bad file descriptor 00:31:51.599 [2024-12-06 17:47:33.370836] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.599 [2024-12-06 17:47:33.370857] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.599 [2024-12-06 17:47:33.370872] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.599 [2024-12-06 17:47:33.370885] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.599 [2024-12-06 17:47:33.370905] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.599 [2024-12-06 17:47:33.370920] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.599 [2024-12-06 17:47:33.370934] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:31:51.599 [2024-12-06 17:47:33.370948] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.599 [2024-12-06 17:47:33.370960] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8e500 is same with the state(6) to be set 00:31:51.599 [2024-12-06 17:47:33.371003] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xc96080 (9): Bad file descriptor 00:31:51.599 [2024-12-06 17:47:33.371035] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x10c4680 (9): Bad file descriptor 00:31:51.599 [2024-12-06 17:47:33.371066] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x10eb9e0 (9): Bad file descriptor 00:31:51.599 [2024-12-06 17:47:33.371098] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xba4610 (9): Bad file descriptor 00:31:51.599 [2024-12-06 17:47:33.371124] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xc99630 (9): Bad file descriptor 00:31:51.599 [2024-12-06 17:47:33.371154] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xc8d630 (9): Bad file descriptor 00:31:51.599 [2024-12-06 17:47:33.372796] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode2, 1] resetting controller 00:31:51.599 [2024-12-06 17:47:33.373776] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:31:51.599 [2024-12-06 17:47:33.373808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xc95950 with addr=10.0.0.2, port=4420 00:31:51.599 [2024-12-06 17:47:33.373826] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc95950 is same with the state(6) to be set 00:31:51.599 [2024-12-06 17:47:33.373906] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:31:51.599 [2024-12-06 17:47:33.373932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xc8d630 with addr=10.0.0.2, port=4420 00:31:51.599 [2024-12-06 17:47:33.373948] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8d630 is same with the state(6) to be set 00:31:51.599 [2024-12-06 17:47:33.374340] nvme_tcp.c:1184:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:31:51.599 [2024-12-06 17:47:33.374415] nvme_tcp.c:1184:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:31:51.599 [2024-12-06 17:47:33.374730] nvme_tcp.c:1184:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:31:51.599 [2024-12-06 17:47:33.374764] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xc95950 (9): Bad file descriptor 00:31:51.599 [2024-12-06 17:47:33.374789] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xc8d630 (9): Bad file descriptor 00:31:51.599 [2024-12-06 17:47:33.374872] nvme_tcp.c:1184:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:31:51.599 [2024-12-06 17:47:33.375003] nvme_tcp.c:1184:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:31:51.599 [2024-12-06 17:47:33.375072] nvme_tcp.c:1184:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:31:51.599 [2024-12-06 17:47:33.375140] nvme_tcp.c:1184:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:31:51.599 [2024-12-06 17:47:33.375207] nvme_tcp.c:1184:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:31:51.599 [2024-12-06 17:47:33.375250] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] Ctrlr is in error state 00:31:51.599 [2024-12-06 17:47:33.375271] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] controller reinitialization failed 00:31:51.599 [2024-12-06 17:47:33.375294] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] in failed state. 00:31:51.599 [2024-12-06 17:47:33.375313] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] Resetting controller failed. 00:31:51.599 [2024-12-06 17:47:33.375329] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] Ctrlr is in error state 00:31:51.599 [2024-12-06 17:47:33.375341] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] controller reinitialization failed 00:31:51.599 [2024-12-06 17:47:33.375354] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] in failed state. 00:31:51.599 [2024-12-06 17:47:33.375366] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] Resetting controller failed. 00:31:51.599 [2024-12-06 17:47:33.380470] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xc8ea40 (9): Bad file descriptor 00:31:51.599 [2024-12-06 17:47:33.380556] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xc8e500 (9): Bad file descriptor 00:31:51.599 [2024-12-06 17:47:33.380763] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.599 [2024-12-06 17:47:33.380789] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.599 [2024-12-06 17:47:33.380822] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.599 [2024-12-06 17:47:33.380838] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.599 [2024-12-06 17:47:33.380856] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.599 [2024-12-06 17:47:33.380871] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.599 [2024-12-06 17:47:33.380888] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.599 [2024-12-06 17:47:33.380903] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.599 [2024-12-06 17:47:33.380919] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.599 [2024-12-06 17:47:33.380934] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.599 [2024-12-06 17:47:33.380950] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.599 [2024-12-06 17:47:33.380965] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.599 [2024-12-06 17:47:33.380981] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.599 [2024-12-06 17:47:33.380996] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.599 [2024-12-06 17:47:33.381013] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.599 [2024-12-06 17:47:33.381028] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.599 [2024-12-06 17:47:33.381045] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.599 [2024-12-06 17:47:33.381059] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.599 [2024-12-06 17:47:33.381085] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.599 [2024-12-06 17:47:33.381102] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.599 [2024-12-06 17:47:33.381119] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.599 [2024-12-06 17:47:33.381133] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.599 [2024-12-06 17:47:33.381149] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.599 [2024-12-06 17:47:33.381165] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.599 [2024-12-06 17:47:33.381181] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.599 [2024-12-06 17:47:33.381196] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.381212] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.381226] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.381243] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.381258] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.381274] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.381288] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.381304] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.381319] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.381335] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.381349] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.381366] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.381380] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.381396] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.381411] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.381427] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.381441] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.381458] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.381476] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.381493] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.381507] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.381524] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.381538] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.381554] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.381568] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.381585] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.381600] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.381617] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.381632] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.381648] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.381671] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.381690] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.381705] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.381721] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.381735] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.381752] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.381766] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.381783] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.381797] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.381813] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.381828] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.381844] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.381858] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.381883] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.381899] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.381916] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.381930] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.381946] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.381960] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.381976] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.381990] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.382007] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:32768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.382021] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.382038] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:32896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.382052] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.382068] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:33024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.382082] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.382098] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:33152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.382112] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.382128] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:33280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.382142] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.382161] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.382175] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.382192] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.382206] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.382223] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.382238] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.382254] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.382272] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.382289] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.382303] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.382319] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.382334] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.382351] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.382366] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.382382] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.382396] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.382412] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.600 [2024-12-06 17:47:33.382427] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.600 [2024-12-06 17:47:33.382443] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.382458] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.382474] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.382488] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.382504] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.382519] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.382535] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.382549] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.382565] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.382580] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.382596] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.382610] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.382626] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.382640] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.382660] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.382683] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.382710] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.382724] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.382742] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.382756] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.382773] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.382787] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.382803] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.382817] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.382833] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x11bbc00 is same with the state(6) to be set 00:31:51.601 [2024-12-06 17:47:33.384093] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.384117] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.384137] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.384153] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.384169] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.384184] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.384201] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.384215] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.384232] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.384246] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.384262] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.384276] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.384292] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.384306] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.384327] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.384342] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.384358] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.384372] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.384389] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.384404] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.384420] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.384434] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.384450] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.384464] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.384482] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.384496] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.384513] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.384527] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.384544] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.384558] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.384574] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.384588] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.384604] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.384618] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.384635] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.384649] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.384671] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.384688] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.384711] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.384729] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.384747] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.384760] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.384776] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.384791] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.384808] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.384823] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.384839] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.601 [2024-12-06 17:47:33.384853] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.601 [2024-12-06 17:47:33.384869] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.384882] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.384899] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.384914] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.384930] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.384944] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.384960] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.384974] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.384990] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385004] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385020] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385034] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385050] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385064] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385081] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385095] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385115] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385130] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385146] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385160] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385176] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385190] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385206] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385221] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385237] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385252] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385268] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385282] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385298] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385312] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385328] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385342] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385358] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385372] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385389] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385403] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385420] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385434] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385450] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385463] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385480] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385497] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385514] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385529] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385546] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385561] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385577] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385591] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385607] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385621] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385638] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385652] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385678] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385695] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385711] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385725] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385741] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385756] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385772] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385786] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385801] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385815] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385832] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385847] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385863] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385887] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385904] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385922] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385939] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385953] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385969] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.385983] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.385999] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.386013] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.386029] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.386043] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.386058] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.386071] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.602 [2024-12-06 17:47:33.386088] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.602 [2024-12-06 17:47:33.386101] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.386116] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x11bef80 is same with the state(6) to be set 00:31:51.603 [2024-12-06 17:47:33.387347] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:16384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.387371] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.387391] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:16512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.387407] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.387423] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:16640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.387437] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.387453] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:16768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.387467] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.387483] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.387497] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.387513] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:17024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.387533] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.387550] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:17152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.387565] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.387580] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:17280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.387595] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.387612] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:17408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.387627] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.387643] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.387658] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.387682] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:17664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.387697] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.387714] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:17792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.387729] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.387744] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:17920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.387759] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.387775] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:18048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.387789] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.387805] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:18176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.387819] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.387835] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.387849] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.387866] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:18432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.387880] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.387896] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:18560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.387911] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.387931] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:18688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.387947] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.387964] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:18816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.387979] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.387995] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:18944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.388009] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.388026] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:19072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.388040] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.388056] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:19200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.388071] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.388086] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:19328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.388101] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.388117] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:19456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.388131] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.388147] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:19584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.388161] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.388178] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:19712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.388193] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.388209] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:19840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.388223] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.388239] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:19968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.388253] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.388270] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:20096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.388284] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.388299] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:20224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.388317] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.388335] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:20352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.388349] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.388366] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:20480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.388380] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.388396] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:20608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.388411] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.388427] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:20736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.388441] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.388458] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:20864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.388472] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.388488] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:20992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.388503] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.388519] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:21120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.388532] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.388549] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:21248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.603 [2024-12-06 17:47:33.388563] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.603 [2024-12-06 17:47:33.388579] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:21376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.388593] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.388610] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:21504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.388624] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.388639] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:21632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.388653] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.388677] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:21760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.388693] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.388713] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:21888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.388729] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.388745] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:22016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.388759] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.388776] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:22144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.388790] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.388806] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:22272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.388820] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.388836] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:22400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.388851] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.388867] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:22528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.388882] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.388897] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:22656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.388912] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.388929] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:22784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.388943] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.388959] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:22912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.388973] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.388989] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.389003] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.389020] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.389034] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.389050] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.389064] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.389081] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:23424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.389099] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.389116] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:23552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.389130] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.389147] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.389161] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.389177] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:23808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.389192] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.389208] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.389222] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.389238] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:24064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.389252] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.389268] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:24192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.389282] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.389299] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:24320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.389312] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.389329] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:24448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.389343] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.389358] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1096430 is same with the state(6) to be set 00:31:51.604 [2024-12-06 17:47:33.390588] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:16384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.390611] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.390633] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:16512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.390649] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.390673] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:16640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.390690] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.390706] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:16768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.390726] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.390744] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.390758] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.390774] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:17024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.390789] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.390804] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:17152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.390824] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.390841] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:17280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.390855] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.390871] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:17408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.390886] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.390901] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.390916] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.390932] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:17664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.390945] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.390961] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:17792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.390975] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.390991] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:17920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.391005] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.391021] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:18048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.391035] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.604 [2024-12-06 17:47:33.391052] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:18176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.604 [2024-12-06 17:47:33.391066] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391082] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391096] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391115] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:18432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391130] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391146] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:18560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391160] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391176] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:18688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391190] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391206] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:18816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391220] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391236] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:18944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391250] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391265] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:19072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391280] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391295] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:19200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391315] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391331] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:19328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391345] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391362] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:19456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391376] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391393] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:19584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391407] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391423] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:19712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391436] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391452] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:19840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391466] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391482] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:19968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391504] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391521] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:20096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391535] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391551] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:20224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391565] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391581] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:20352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391595] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391611] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:20480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391625] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391641] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:20608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391655] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391679] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:20736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391695] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391711] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:20864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391725] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391741] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:20992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391755] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391771] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:21120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391785] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391801] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:21248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391820] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391837] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:21376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391851] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391866] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:21504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391880] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391900] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:21632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391915] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391931] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:21760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391945] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391961] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:21888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.391975] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.391991] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:22016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.392005] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.392021] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:22144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.392034] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.392051] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:22272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.392065] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.392082] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:22400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.392096] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.392111] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:22528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.392125] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.392141] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:22656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.392155] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.392171] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:22784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.392185] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.392201] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:22912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.392215] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.392236] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.392251] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.392267] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.392284] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.605 [2024-12-06 17:47:33.392300] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.605 [2024-12-06 17:47:33.392315] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.392331] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:23424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.392346] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.392362] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:23552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.392376] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.392392] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.392407] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.392423] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:23808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.392437] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.392453] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.392468] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.392483] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:24064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.392497] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.392513] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:24192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.392527] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.392543] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:24320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.392557] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.392573] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:24448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.392587] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.392601] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x137ec30 is same with the state(6) to be set 00:31:51.606 [2024-12-06 17:47:33.393857] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.393880] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.393901] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:17024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.393921] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.393938] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:17152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.393953] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.393969] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:17280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.393985] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.394002] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:17408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.394016] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.394033] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.394048] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.394064] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:17664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.394078] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.394094] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:17792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.394108] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.394124] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:17920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.394139] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.394155] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:18048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.394169] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.394186] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:18176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.394200] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.394216] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.394229] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.394245] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:18432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.394260] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.394276] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:18560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.394290] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.394310] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:18688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.394326] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.394342] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:18816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.394357] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.394374] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:18944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.394388] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.394404] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:19072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.394419] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.394435] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:19200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.394449] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.394466] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:19328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.394481] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.394497] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:19456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.394513] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.394529] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:19584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.394543] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.394560] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:19712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.394574] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.394590] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:19840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.394605] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.394621] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:19968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.394636] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.606 [2024-12-06 17:47:33.394652] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:20096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.606 [2024-12-06 17:47:33.394675] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.394694] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:20224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.394712] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.394730] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:20352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.394744] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.394760] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:20480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.394774] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.394790] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:20608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.394804] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.394820] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:20736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.394834] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.394850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:20864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.394865] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.394880] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:20992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.394894] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.394910] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:21120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.394924] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.394940] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:21248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.394954] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.394970] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:21376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.394984] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.395000] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:21504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.395015] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.395031] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:21632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.395045] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.395061] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:21760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.395075] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.395091] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:21888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.395109] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.395126] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:22016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.395140] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.395156] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:22144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.395171] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.395187] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:22272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.395201] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.395217] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:22400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.395231] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.402771] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:22528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.402829] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.402848] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:22656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.402862] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.402879] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:22784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.402894] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.402910] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:22912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.402926] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.402942] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.402957] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.402973] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.402988] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.403004] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.403017] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.403034] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:23424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.403048] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.403075] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:23552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.403091] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.403108] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.403122] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.403138] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:23808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.403153] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.403169] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.403183] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.403199] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:24064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.403214] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.403231] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:24192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.403245] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.403262] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:24320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.403276] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.403292] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:24448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.403306] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.403322] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x137fef0 is same with the state(6) to be set 00:31:51.607 [2024-12-06 17:47:33.404685] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.404719] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.404743] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:17024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.404759] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.404776] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:17152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.404790] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.404807] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:17280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.404822] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.404844] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:17408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.607 [2024-12-06 17:47:33.404859] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.607 [2024-12-06 17:47:33.404874] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.404889] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.404906] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:17664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.404921] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.404937] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:17792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.404951] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.404968] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:17920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.404982] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.404998] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:18048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405012] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405029] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:18176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405043] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405060] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405074] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405090] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:18432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405105] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405121] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:18560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405135] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405151] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:18688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405166] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405182] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:18816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405196] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405212] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:18944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405230] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405247] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:19072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405262] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405278] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:19200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405293] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405309] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:19328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405323] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405339] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:19456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405353] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405369] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:19584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405383] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405399] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:19712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405414] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405430] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:19840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405444] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405461] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:19968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405475] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405491] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:20096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405505] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405521] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:20224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405535] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405552] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:20352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405566] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405581] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:20480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405595] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405616] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:20608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405631] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405647] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:20736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405661] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405686] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:20864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405701] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405717] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:20992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405731] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405748] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:21120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405762] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405778] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:21248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405792] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405808] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:21376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405822] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405838] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:21504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405852] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405869] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:21632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405883] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405899] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:21760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405914] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405930] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:21888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405945] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405961] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:22016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.405975] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.405991] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:22144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.406009] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.406026] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:22272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.406040] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.406058] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:22400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.406073] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.608 [2024-12-06 17:47:33.406089] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:22528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.608 [2024-12-06 17:47:33.406103] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.609 [2024-12-06 17:47:33.406119] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:22656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.609 [2024-12-06 17:47:33.406133] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.609 [2024-12-06 17:47:33.406149] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:22784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.609 [2024-12-06 17:47:33.406163] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.609 [2024-12-06 17:47:33.406179] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:22912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.609 [2024-12-06 17:47:33.406195] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.609 [2024-12-06 17:47:33.406211] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.609 [2024-12-06 17:47:33.406225] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.609 [2024-12-06 17:47:33.406240] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.609 [2024-12-06 17:47:33.406254] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.609 [2024-12-06 17:47:33.406271] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.609 [2024-12-06 17:47:33.406285] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.609 [2024-12-06 17:47:33.406301] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:23424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.609 [2024-12-06 17:47:33.406315] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.609 [2024-12-06 17:47:33.406331] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:23552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.609 [2024-12-06 17:47:33.406345] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.609 [2024-12-06 17:47:33.406361] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.609 [2024-12-06 17:47:33.406375] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.609 [2024-12-06 17:47:33.406396] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:23808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.609 [2024-12-06 17:47:33.406410] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.609 [2024-12-06 17:47:33.406426] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.609 [2024-12-06 17:47:33.406442] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.609 [2024-12-06 17:47:33.406458] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.609 [2024-12-06 17:47:33.406472] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.609 [2024-12-06 17:47:33.406488] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.609 [2024-12-06 17:47:33.406502] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.609 [2024-12-06 17:47:33.406519] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.609 [2024-12-06 17:47:33.406533] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.609 [2024-12-06 17:47:33.406548] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.609 [2024-12-06 17:47:33.406562] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.609 [2024-12-06 17:47:33.406578] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:24064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.609 [2024-12-06 17:47:33.406592] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.609 [2024-12-06 17:47:33.406608] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:24192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.609 [2024-12-06 17:47:33.406622] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.609 [2024-12-06 17:47:33.406638] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:24320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.609 [2024-12-06 17:47:33.406652] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.609 [2024-12-06 17:47:33.406675] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:24448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.609 [2024-12-06 17:47:33.406691] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.609 [2024-12-06 17:47:33.406706] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x11b3d20 is same with the state(6) to be set 00:31:51.871 [2024-12-06 17:47:33.407913] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 1] resetting controller 00:31:51.871 [2024-12-06 17:47:33.407948] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode4, 1] resetting controller 00:31:51.871 [2024-12-06 17:47:33.407972] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode5, 1] resetting controller 00:31:51.871 [2024-12-06 17:47:33.407994] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode6, 1] resetting controller 00:31:51.871 [2024-12-06 17:47:33.408111] bdev_nvme.c:3172:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: [nqn.2016-06.io.spdk:cnode10, 1] Unable to perform failover, already in progress. 00:31:51.871 [2024-12-06 17:47:33.408154] bdev_nvme.c:3172:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: [nqn.2016-06.io.spdk:cnode7, 1] Unable to perform failover, already in progress. 00:31:51.871 [2024-12-06 17:47:33.408283] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode7, 1] resetting controller 00:31:51.871 [2024-12-06 17:47:33.408312] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode10, 1] resetting controller 00:31:51.871 [2024-12-06 17:47:33.408516] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:31:51.871 [2024-12-06 17:47:33.408546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xc99630 with addr=10.0.0.2, port=4420 00:31:51.871 [2024-12-06 17:47:33.408564] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc99630 is same with the state(6) to be set 00:31:51.871 [2024-12-06 17:47:33.408672] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:31:51.871 [2024-12-06 17:47:33.408698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xc96080 with addr=10.0.0.2, port=4420 00:31:51.871 [2024-12-06 17:47:33.408714] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc96080 is same with the state(6) to be set 00:31:51.871 [2024-12-06 17:47:33.408789] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:31:51.871 [2024-12-06 17:47:33.408814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x10c4680 with addr=10.0.0.2, port=4420 00:31:51.871 [2024-12-06 17:47:33.408830] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x10c4680 is same with the state(6) to be set 00:31:51.871 [2024-12-06 17:47:33.408915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:31:51.871 [2024-12-06 17:47:33.408939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x10eb9e0 with addr=10.0.0.2, port=4420 00:31:51.871 [2024-12-06 17:47:33.408954] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x10eb9e0 is same with the state(6) to be set 00:31:51.871 [2024-12-06 17:47:33.410310] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.871 [2024-12-06 17:47:33.410335] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.871 [2024-12-06 17:47:33.410361] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:17024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.871 [2024-12-06 17:47:33.410377] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.871 [2024-12-06 17:47:33.410394] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:17152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.871 [2024-12-06 17:47:33.410408] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.871 [2024-12-06 17:47:33.410425] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:17280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.871 [2024-12-06 17:47:33.410440] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.871 [2024-12-06 17:47:33.410456] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:17408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.871 [2024-12-06 17:47:33.410470] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.871 [2024-12-06 17:47:33.410487] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.871 [2024-12-06 17:47:33.410502] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.871 [2024-12-06 17:47:33.410524] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:17664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.871 [2024-12-06 17:47:33.410539] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.410555] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:17792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.410569] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.410585] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:17920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.410600] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.410616] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:18048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.410630] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.410646] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:18176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.410660] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.410685] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.410712] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.410728] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:18432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.410743] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.410759] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:18560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.410773] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.410790] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:18688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.410805] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.410821] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:18816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.410835] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.410853] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:18944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.410868] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.410885] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:19072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.410900] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.410916] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:19200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.410937] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.410955] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:19328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.410970] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.410986] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:19456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.411001] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.411017] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:19584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.411032] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.411047] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:19712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.411062] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.411078] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:19840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.411092] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.411108] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:19968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.411122] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.411138] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:20096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.411153] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.411169] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:20224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.411183] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.411199] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:20352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.411213] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.411230] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:20480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.411244] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.411260] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:20608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.411274] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.411290] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:20736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.411304] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.411324] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:20864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.411339] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.411356] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:20992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.411371] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.411387] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:21120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.411402] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.411418] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:21248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.411432] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.411449] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:21376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.411463] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.411480] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:21504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.411494] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.872 [2024-12-06 17:47:33.411510] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:21632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.872 [2024-12-06 17:47:33.411525] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.411541] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:21760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.411555] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.411571] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:21888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.411585] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.411601] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:22016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.411615] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.411631] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:22144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.411645] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.411661] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:22272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.411685] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.411703] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:22400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.411721] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.411738] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:22528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.411753] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.411769] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:22656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.411784] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.411800] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:22784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.411814] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.411830] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:22912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.411845] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.411862] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.411876] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.411892] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.411906] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.411922] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.411937] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.411953] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:23424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.411967] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.411983] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:23552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.411997] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.412014] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.412028] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.412044] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:23808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.412058] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.412074] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.412089] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.412109] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:24064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.412125] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.412142] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.412156] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.412172] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.412187] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.412203] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.412217] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.412233] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.412248] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.412264] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:24192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.412279] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.412295] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:24320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.412309] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.412325] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:24448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.412340] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.412355] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x13803c0 is same with the state(6) to be set 00:31:51.873 [2024-12-06 17:47:33.413615] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:16384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.413639] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.413660] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:16512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.413683] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.413700] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:16640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.413714] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.873 [2024-12-06 17:47:33.413731] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:16768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.873 [2024-12-06 17:47:33.413746] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.413767] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.413783] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.413798] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:17024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.413813] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.413830] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:17152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.413844] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.413860] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:17280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.413874] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.413890] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:17408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.413906] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.413923] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.413937] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.413953] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:17664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.413968] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.413983] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:17792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.414000] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.414016] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:17920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.414030] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.414047] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:18048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.414061] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.414078] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:18176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.414093] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.414110] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.414124] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.414140] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:18432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.414154] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.414174] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:18560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.414189] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.414205] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:18688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.414219] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.414235] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:18816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.414249] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.414266] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:18944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.414279] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.414296] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:19072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.414311] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.414327] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:19200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.414341] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.414358] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:19328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.414371] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.414388] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:19456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.414402] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.414419] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:19584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.414433] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.414449] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:19712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.414463] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.414479] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:19840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.414494] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.414510] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:19968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.414524] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.414540] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:20096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.414559] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.414576] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:20224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.414590] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.414606] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:20352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.414620] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.414636] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:20480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.414650] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.414672] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:20608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.414688] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.414704] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:20736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.414720] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.874 [2024-12-06 17:47:33.414736] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:20864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.874 [2024-12-06 17:47:33.414750] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.414766] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:20992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.875 [2024-12-06 17:47:33.414780] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.414797] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:21120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.875 [2024-12-06 17:47:33.414811] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.414827] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:21248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.875 [2024-12-06 17:47:33.414841] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.414857] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:21376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.875 [2024-12-06 17:47:33.414871] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.414887] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:21504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.875 [2024-12-06 17:47:33.414901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.414918] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:21632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.875 [2024-12-06 17:47:33.414933] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.414954] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:21760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.875 [2024-12-06 17:47:33.414969] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.414985] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:21888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.875 [2024-12-06 17:47:33.414999] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.415015] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:22016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.875 [2024-12-06 17:47:33.415029] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.415046] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:22144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.875 [2024-12-06 17:47:33.415061] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.415077] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:22272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.875 [2024-12-06 17:47:33.415091] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.415108] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:22400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.875 [2024-12-06 17:47:33.415122] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.415138] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:22528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.875 [2024-12-06 17:47:33.415153] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.415169] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:22656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.875 [2024-12-06 17:47:33.415182] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.415198] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:22784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.875 [2024-12-06 17:47:33.415213] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.415228] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:22912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.875 [2024-12-06 17:47:33.415243] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.415259] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.875 [2024-12-06 17:47:33.415273] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.415289] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.875 [2024-12-06 17:47:33.415303] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.415319] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.875 [2024-12-06 17:47:33.415337] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.415354] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:23424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.875 [2024-12-06 17:47:33.415369] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.415385] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:23552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.875 [2024-12-06 17:47:33.415399] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.415414] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.875 [2024-12-06 17:47:33.415428] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.415444] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:23808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.875 [2024-12-06 17:47:33.415458] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.415474] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.875 [2024-12-06 17:47:33.415488] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.415504] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:24064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.875 [2024-12-06 17:47:33.415517] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.415534] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:24192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.875 [2024-12-06 17:47:33.415548] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.415564] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:24320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.875 [2024-12-06 17:47:33.415577] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.415594] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:24448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:31:51.875 [2024-12-06 17:47:33.415609] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:31:51.875 [2024-12-06 17:47:33.415623] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x10a05e0 is same with the state(6) to be set 00:31:51.875 [2024-12-06 17:47:33.417534] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode2, 1] resetting controller 00:31:51.875 [2024-12-06 17:47:33.417573] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode3, 1] resetting controller 00:31:51.875 [2024-12-06 17:47:33.417603] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode8, 1] resetting controller 00:31:51.875 task offset: 26240 on job bdev=Nvme3n1 fails 00:31:51.875 00:31:51.875 Latency(us) 00:31:51.875 [2024-12-06T16:47:33.715Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:31:51.876 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:51.876 Job: Nvme1n1 ended in about 0.94 seconds with error 00:31:51.876 Verification LBA range: start 0x0 length 0x400 00:31:51.876 Nvme1n1 : 0.94 210.36 13.15 68.34 0.00 227091.59 18835.53 239230.67 00:31:51.876 Job: Nvme2n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:51.876 Job: Nvme2n1 ended in about 0.92 seconds with error 00:31:51.876 Verification LBA range: start 0x0 length 0x400 00:31:51.876 Nvme2n1 : 0.92 207.60 12.97 69.20 0.00 224082.11 20583.16 257872.02 00:31:51.876 Job: Nvme3n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:51.876 Job: Nvme3n1 ended in about 0.92 seconds with error 00:31:51.876 Verification LBA range: start 0x0 length 0x400 00:31:51.876 Nvme3n1 : 0.92 208.16 13.01 69.39 0.00 218875.07 17573.36 253211.69 00:31:51.876 Job: Nvme4n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:51.876 Job: Nvme4n1 ended in about 0.94 seconds with error 00:31:51.876 Verification LBA range: start 0x0 length 0x400 00:31:51.876 Nvme4n1 : 0.94 204.31 12.77 68.10 0.00 218588.16 18641.35 253211.69 00:31:51.876 Job: Nvme5n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:51.876 Job: Nvme5n1 ended in about 0.94 seconds with error 00:31:51.876 Verification LBA range: start 0x0 length 0x400 00:31:51.876 Nvme5n1 : 0.94 135.74 8.48 67.87 0.00 286598.89 22039.51 268746.15 00:31:51.876 Job: Nvme6n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:51.876 Job: Nvme6n1 ended in about 0.95 seconds with error 00:31:51.876 Verification LBA range: start 0x0 length 0x400 00:31:51.876 Nvme6n1 : 0.95 135.27 8.45 67.64 0.00 281531.48 22524.97 273406.48 00:31:51.876 Job: Nvme7n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:51.876 Job: Nvme7n1 ended in about 0.96 seconds with error 00:31:51.876 Verification LBA range: start 0x0 length 0x400 00:31:51.876 Nvme7n1 : 0.96 137.94 8.62 62.70 0.00 278399.05 19320.98 254765.13 00:31:51.876 Job: Nvme8n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:51.876 Job: Nvme8n1 ended in about 0.97 seconds with error 00:31:51.876 Verification LBA range: start 0x0 length 0x400 00:31:51.876 Nvme8n1 : 0.97 136.65 8.54 66.26 0.00 270403.37 18738.44 256318.58 00:31:51.876 Job: Nvme9n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:51.876 Job: Nvme9n1 ended in about 0.97 seconds with error 00:31:51.876 Verification LBA range: start 0x0 length 0x400 00:31:51.876 Nvme9n1 : 0.97 132.07 8.25 66.03 0.00 271267.97 38641.97 256318.58 00:31:51.876 Job: Nvme10n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:31:51.876 Job: Nvme10n1 ended in about 0.96 seconds with error 00:31:51.876 Verification LBA range: start 0x0 length 0x400 00:31:51.876 Nvme10n1 : 0.96 137.46 8.59 66.65 0.00 256999.94 20000.62 274959.93 00:31:51.876 [2024-12-06T16:47:33.715Z] =================================================================================================================== 00:31:51.876 [2024-12-06T16:47:33.715Z] Total : 1645.56 102.85 672.17 0.00 249709.76 17573.36 274959.93 00:31:51.876 [2024-12-06 17:47:33.444265] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:31:51.876 [2024-12-06 17:47:33.444364] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode9, 1] resetting controller 00:31:51.876 [2024-12-06 17:47:33.444647] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:31:51.876 [2024-12-06 17:47:33.444692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xba4610 with addr=10.0.0.2, port=4420 00:31:51.876 [2024-12-06 17:47:33.444713] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xba4610 is same with the state(6) to be set 00:31:51.876 [2024-12-06 17:47:33.444802] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:31:51.876 [2024-12-06 17:47:33.444829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x10f3710 with addr=10.0.0.2, port=4420 00:31:51.876 [2024-12-06 17:47:33.444846] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x10f3710 is same with the state(6) to be set 00:31:51.876 [2024-12-06 17:47:33.444885] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xc99630 (9): Bad file descriptor 00:31:51.876 [2024-12-06 17:47:33.444909] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xc96080 (9): Bad file descriptor 00:31:51.876 [2024-12-06 17:47:33.444929] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x10c4680 (9): Bad file descriptor 00:31:51.876 [2024-12-06 17:47:33.444948] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x10eb9e0 (9): Bad file descriptor 00:31:51.876 [2024-12-06 17:47:33.445249] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:31:51.876 [2024-12-06 17:47:33.445279] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xc8d630 with addr=10.0.0.2, port=4420 00:31:51.876 [2024-12-06 17:47:33.445297] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8d630 is same with the state(6) to be set 00:31:51.876 [2024-12-06 17:47:33.445380] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:31:51.876 [2024-12-06 17:47:33.445407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xc95950 with addr=10.0.0.2, port=4420 00:31:51.876 [2024-12-06 17:47:33.445423] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc95950 is same with the state(6) to be set 00:31:51.876 [2024-12-06 17:47:33.445513] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:31:51.876 [2024-12-06 17:47:33.445539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xc8ea40 with addr=10.0.0.2, port=4420 00:31:51.876 [2024-12-06 17:47:33.445556] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8ea40 is same with the state(6) to be set 00:31:51.876 [2024-12-06 17:47:33.445635] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:31:51.876 [2024-12-06 17:47:33.445660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xc8e500 with addr=10.0.0.2, port=4420 00:31:51.876 [2024-12-06 17:47:33.445691] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8e500 is same with the state(6) to be set 00:31:51.876 [2024-12-06 17:47:33.445711] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xba4610 (9): Bad file descriptor 00:31:51.876 [2024-12-06 17:47:33.445731] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x10f3710 (9): Bad file descriptor 00:31:51.876 [2024-12-06 17:47:33.445749] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] Ctrlr is in error state 00:31:51.876 [2024-12-06 17:47:33.445763] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] controller reinitialization failed 00:31:51.876 [2024-12-06 17:47:33.445779] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] in failed state. 00:31:51.876 [2024-12-06 17:47:33.445797] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] Resetting controller failed. 00:31:51.876 [2024-12-06 17:47:33.445813] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] Ctrlr is in error state 00:31:51.876 [2024-12-06 17:47:33.445826] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] controller reinitialization failed 00:31:51.876 [2024-12-06 17:47:33.445839] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] in failed state. 00:31:51.876 [2024-12-06 17:47:33.445853] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] Resetting controller failed. 00:31:51.876 [2024-12-06 17:47:33.445866] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] Ctrlr is in error state 00:31:51.876 [2024-12-06 17:47:33.445879] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] controller reinitialization failed 00:31:51.876 [2024-12-06 17:47:33.445892] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] in failed state. 00:31:51.876 [2024-12-06 17:47:33.445905] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] Resetting controller failed. 00:31:51.876 [2024-12-06 17:47:33.445939] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] Ctrlr is in error state 00:31:51.877 [2024-12-06 17:47:33.445961] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] controller reinitialization failed 00:31:51.877 [2024-12-06 17:47:33.445974] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] in failed state. 00:31:51.877 [2024-12-06 17:47:33.445987] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] Resetting controller failed. 00:31:51.877 [2024-12-06 17:47:33.446053] bdev_nvme.c:3172:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: [nqn.2016-06.io.spdk:cnode10, 1] Unable to perform failover, already in progress. 00:31:51.877 [2024-12-06 17:47:33.446079] bdev_nvme.c:3172:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: [nqn.2016-06.io.spdk:cnode7, 1] Unable to perform failover, already in progress. 00:31:51.877 [2024-12-06 17:47:33.446776] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xc8d630 (9): Bad file descriptor 00:31:51.877 [2024-12-06 17:47:33.446806] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xc95950 (9): Bad file descriptor 00:31:51.877 [2024-12-06 17:47:33.446826] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xc8ea40 (9): Bad file descriptor 00:31:51.877 [2024-12-06 17:47:33.446846] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xc8e500 (9): Bad file descriptor 00:31:51.877 [2024-12-06 17:47:33.446862] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] Ctrlr is in error state 00:31:51.877 [2024-12-06 17:47:33.446876] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] controller reinitialization failed 00:31:51.877 [2024-12-06 17:47:33.446889] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] in failed state. 00:31:51.877 [2024-12-06 17:47:33.446902] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] Resetting controller failed. 00:31:51.877 [2024-12-06 17:47:33.446916] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode10, 1] Ctrlr is in error state 00:31:51.877 [2024-12-06 17:47:33.446929] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode10, 1] controller reinitialization failed 00:31:51.877 [2024-12-06 17:47:33.446942] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode10, 1] in failed state. 00:31:51.877 [2024-12-06 17:47:33.446955] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode10, 1] Resetting controller failed. 00:31:51.877 [2024-12-06 17:47:33.447024] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode6, 1] resetting controller 00:31:51.877 [2024-12-06 17:47:33.447049] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode5, 1] resetting controller 00:31:51.877 [2024-12-06 17:47:33.447068] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode4, 1] resetting controller 00:31:51.877 [2024-12-06 17:47:33.447085] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 1] resetting controller 00:31:51.877 [2024-12-06 17:47:33.447127] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] Ctrlr is in error state 00:31:51.877 [2024-12-06 17:47:33.447145] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] controller reinitialization failed 00:31:51.877 [2024-12-06 17:47:33.447159] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] in failed state. 00:31:51.877 [2024-12-06 17:47:33.447173] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] Resetting controller failed. 00:31:51.877 [2024-12-06 17:47:33.447187] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] Ctrlr is in error state 00:31:51.877 [2024-12-06 17:47:33.447200] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] controller reinitialization failed 00:31:51.877 [2024-12-06 17:47:33.447218] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] in failed state. 00:31:51.877 [2024-12-06 17:47:33.447232] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] Resetting controller failed. 00:31:51.877 [2024-12-06 17:47:33.447245] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode8, 1] Ctrlr is in error state 00:31:51.877 [2024-12-06 17:47:33.447258] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode8, 1] controller reinitialization failed 00:31:51.877 [2024-12-06 17:47:33.447271] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode8, 1] in failed state. 00:31:51.877 [2024-12-06 17:47:33.447283] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode8, 1] Resetting controller failed. 00:31:51.877 [2024-12-06 17:47:33.447297] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode9, 1] Ctrlr is in error state 00:31:51.877 [2024-12-06 17:47:33.447310] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode9, 1] controller reinitialization failed 00:31:51.877 [2024-12-06 17:47:33.447323] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode9, 1] in failed state. 00:31:51.877 [2024-12-06 17:47:33.447336] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode9, 1] Resetting controller failed. 00:31:51.877 [2024-12-06 17:47:33.447463] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:31:51.877 [2024-12-06 17:47:33.447491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x10eb9e0 with addr=10.0.0.2, port=4420 00:31:51.877 [2024-12-06 17:47:33.447508] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x10eb9e0 is same with the state(6) to be set 00:31:51.877 [2024-12-06 17:47:33.447596] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:31:51.877 [2024-12-06 17:47:33.447622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x10c4680 with addr=10.0.0.2, port=4420 00:31:51.877 [2024-12-06 17:47:33.447638] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x10c4680 is same with the state(6) to be set 00:31:51.877 [2024-12-06 17:47:33.447737] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:31:51.877 [2024-12-06 17:47:33.447763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xc96080 with addr=10.0.0.2, port=4420 00:31:51.877 [2024-12-06 17:47:33.447779] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc96080 is same with the state(6) to be set 00:31:51.877 [2024-12-06 17:47:33.447862] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:31:51.877 [2024-12-06 17:47:33.447887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xc99630 with addr=10.0.0.2, port=4420 00:31:51.877 [2024-12-06 17:47:33.447903] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc99630 is same with the state(6) to be set 00:31:51.877 [2024-12-06 17:47:33.447948] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x10eb9e0 (9): Bad file descriptor 00:31:51.877 [2024-12-06 17:47:33.447973] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x10c4680 (9): Bad file descriptor 00:31:51.877 [2024-12-06 17:47:33.447996] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xc96080 (9): Bad file descriptor 00:31:51.877 [2024-12-06 17:47:33.448014] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xc99630 (9): Bad file descriptor 00:31:51.877 [2024-12-06 17:47:33.448064] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] Ctrlr is in error state 00:31:51.877 [2024-12-06 17:47:33.448083] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] controller reinitialization failed 00:31:51.877 [2024-12-06 17:47:33.448098] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] in failed state. 00:31:51.878 [2024-12-06 17:47:33.448116] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] Resetting controller failed. 00:31:51.878 [2024-12-06 17:47:33.448131] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] Ctrlr is in error state 00:31:51.878 [2024-12-06 17:47:33.448145] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] controller reinitialization failed 00:31:51.878 [2024-12-06 17:47:33.448158] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] in failed state. 00:31:51.878 [2024-12-06 17:47:33.448171] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] Resetting controller failed. 00:31:51.878 [2024-12-06 17:47:33.448185] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] Ctrlr is in error state 00:31:51.878 [2024-12-06 17:47:33.448198] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] controller reinitialization failed 00:31:51.878 [2024-12-06 17:47:33.448211] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] in failed state. 00:31:51.878 [2024-12-06 17:47:33.448223] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] Resetting controller failed. 00:31:51.878 [2024-12-06 17:47:33.448237] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] Ctrlr is in error state 00:31:51.878 [2024-12-06 17:47:33.448248] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] controller reinitialization failed 00:31:51.878 [2024-12-06 17:47:33.448261] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] in failed state. 00:31:51.878 [2024-12-06 17:47:33.448273] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] Resetting controller failed. 00:31:52.143 17:47:33 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@137 -- # sleep 1 00:31:53.080 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@138 -- # NOT wait 337355 00:31:53.080 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@652 -- # local es=0 00:31:53.080 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@654 -- # valid_exec_arg wait 337355 00:31:53.080 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@640 -- # local arg=wait 00:31:53.080 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:31:53.080 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@644 -- # type -t wait 00:31:53.080 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:31:53.080 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@655 -- # wait 337355 00:31:53.080 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@655 -- # es=255 00:31:53.080 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:31:53.080 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@664 -- # es=127 00:31:53.080 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@665 -- # case "$es" in 00:31:53.080 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@672 -- # es=1 00:31:53.080 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:31:53.080 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@140 -- # stoptarget 00:31:53.080 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@42 -- # rm -f ./local-job0-0-verify.state 00:31:53.080 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@43 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:31:53.081 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@44 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:31:53.081 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@46 -- # nvmftestfini 00:31:53.081 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@516 -- # nvmfcleanup 00:31:53.081 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@121 -- # sync 00:31:53.081 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:31:53.081 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@124 -- # set +e 00:31:53.081 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@125 -- # for i in {1..20} 00:31:53.081 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:31:53.081 rmmod nvme_tcp 00:31:53.081 rmmod nvme_fabrics 00:31:53.081 rmmod nvme_keyring 00:31:53.341 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:31:53.341 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@128 -- # set -e 00:31:53.341 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@129 -- # return 0 00:31:53.341 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@517 -- # '[' -n 337174 ']' 00:31:53.341 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@518 -- # killprocess 337174 00:31:53.341 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@954 -- # '[' -z 337174 ']' 00:31:53.341 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@958 -- # kill -0 337174 00:31:53.341 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 958: kill: (337174) - No such process 00:31:53.341 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@981 -- # echo 'Process with pid 337174 is not found' 00:31:53.341 Process with pid 337174 is not found 00:31:53.341 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:31:53.341 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:31:53.341 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:31:53.341 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@297 -- # iptr 00:31:53.341 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@791 -- # iptables-save 00:31:53.341 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@791 -- # iptables-restore 00:31:53.341 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:31:53.341 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:31:53.341 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@302 -- # remove_spdk_ns 00:31:53.341 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:31:53.341 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:31:53.341 17:47:34 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:31:55.245 17:47:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:31:55.245 00:31:55.245 real 0m7.484s 00:31:55.245 user 0m18.559s 00:31:55.245 sys 0m1.459s 00:31:55.245 17:47:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@1130 -- # xtrace_disable 00:31:55.245 17:47:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:31:55.245 ************************************ 00:31:55.245 END TEST nvmf_shutdown_tc3 00:31:55.245 ************************************ 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@166 -- # [[ e810 == \e\8\1\0 ]] 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@166 -- # [[ tcp == \r\d\m\a ]] 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@167 -- # run_test nvmf_shutdown_tc4 nvmf_shutdown_tc4 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1111 -- # xtrace_disable 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@10 -- # set +x 00:31:55.245 ************************************ 00:31:55.245 START TEST nvmf_shutdown_tc4 00:31:55.245 ************************************ 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@1129 -- # nvmf_shutdown_tc4 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@145 -- # starttarget 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@16 -- # nvmftestinit 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@476 -- # prepare_net_devs 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@438 -- # local -g is_hw=no 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@440 -- # remove_spdk_ns 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@309 -- # xtrace_disable 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@315 -- # pci_devs=() 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@315 -- # local -a pci_devs 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@316 -- # pci_net_devs=() 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@317 -- # pci_drivers=() 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@317 -- # local -A pci_drivers 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@319 -- # net_devs=() 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@319 -- # local -ga net_devs 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@320 -- # e810=() 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@320 -- # local -ga e810 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@321 -- # x722=() 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@321 -- # local -ga x722 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@322 -- # mlx=() 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@322 -- # local -ga mlx 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:31:55.245 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:31:55.245 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:31:55.245 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@418 -- # [[ up == up ]] 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:31:55.246 Found net devices under 0000:0a:00.0: cvl_0_0 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@418 -- # [[ up == up ]] 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:31:55.246 Found net devices under 0000:0a:00.1: cvl_0_1 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@442 -- # is_hw=yes 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:31:55.246 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:31:55.505 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:31:55.505 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.164 ms 00:31:55.505 00:31:55.505 --- 10.0.0.2 ping statistics --- 00:31:55.505 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:31:55.505 rtt min/avg/max/mdev = 0.164/0.164/0.164/0.000 ms 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:31:55.505 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:31:55.505 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.068 ms 00:31:55.505 00:31:55.505 --- 10.0.0.1 ping statistics --- 00:31:55.505 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:31:55.505 rtt min/avg/max/mdev = 0.068/0.068/0.068/0.000 ms 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@450 -- # return 0 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@19 -- # nvmfappstart -m 0x1E 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@726 -- # xtrace_disable 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@509 -- # nvmfpid=338256 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1E 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@510 -- # waitforlisten 338256 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@835 -- # '[' -z 338256 ']' 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@840 -- # local max_retries=100 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:31:55.505 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@844 -- # xtrace_disable 00:31:55.505 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:31:55.764 [2024-12-06 17:47:37.352556] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:31:55.764 [2024-12-06 17:47:37.352638] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:31:55.764 [2024-12-06 17:47:37.425397] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:31:55.764 [2024-12-06 17:47:37.474057] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:31:55.764 [2024-12-06 17:47:37.474115] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:31:55.764 [2024-12-06 17:47:37.474129] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:31:55.764 [2024-12-06 17:47:37.474140] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:31:55.764 [2024-12-06 17:47:37.474150] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:31:55.764 [2024-12-06 17:47:37.475649] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:31:55.764 [2024-12-06 17:47:37.475712] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:31:55.764 [2024-12-06 17:47:37.475782] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:31:55.764 [2024-12-06 17:47:37.475785] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:31:55.764 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:31:55.764 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@868 -- # return 0 00:31:55.764 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:31:55.764 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@732 -- # xtrace_disable 00:31:55.764 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@21 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:31:56.022 [2024-12-06 17:47:37.627888] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@23 -- # num_subsystems=({1..10}) 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@25 -- # timing_enter create_subsystems 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@726 -- # xtrace_disable 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@27 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@36 -- # rpc_cmd 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@563 -- # xtrace_disable 00:31:56.022 17:47:37 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:31:56.022 Malloc1 00:31:56.022 [2024-12-06 17:47:37.732728] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:31:56.022 Malloc2 00:31:56.022 Malloc3 00:31:56.022 Malloc4 00:31:56.281 Malloc5 00:31:56.281 Malloc6 00:31:56.281 Malloc7 00:31:56.281 Malloc8 00:31:56.281 Malloc9 00:31:56.538 Malloc10 00:31:56.538 17:47:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:31:56.538 17:47:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@37 -- # timing_exit create_subsystems 00:31:56.538 17:47:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@732 -- # xtrace_disable 00:31:56.538 17:47:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:31:56.538 17:47:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@149 -- # perfpid=338320 00:31:56.538 17:47:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@148 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 128 -o 45056 -O 4096 -w randwrite -t 20 -r 'trtype:tcp adrfam:IPV4 traddr:10.0.0.2 trsvcid:4420' -P 4 00:31:56.538 17:47:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@150 -- # sleep 5 00:31:56.538 [2024-12-06 17:47:38.243184] subsystem.c:1641:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:32:01.806 17:47:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@152 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; kill -9 $perfpid || true; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:32:01.806 17:47:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@155 -- # killprocess 338256 00:32:01.806 17:47:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@954 -- # '[' -z 338256 ']' 00:32:01.806 17:47:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@958 -- # kill -0 338256 00:32:01.806 17:47:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@959 -- # uname 00:32:01.806 17:47:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:32:01.806 17:47:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 338256 00:32:01.806 17:47:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:32:01.806 17:47:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:32:01.806 17:47:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@972 -- # echo 'killing process with pid 338256' 00:32:01.806 killing process with pid 338256 00:32:01.806 17:47:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@973 -- # kill 338256 00:32:01.806 17:47:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@978 -- # wait 338256 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 starting I/O failed: -6 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 starting I/O failed: -6 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 starting I/O failed: -6 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 starting I/O failed: -6 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 starting I/O failed: -6 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 starting I/O failed: -6 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 starting I/O failed: -6 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 starting I/O failed: -6 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 starting I/O failed: -6 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 [2024-12-06 17:47:43.245576] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24eef20 is same with the state(6) to be set 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 starting I/O failed: -6 00:32:01.806 [2024-12-06 17:47:43.245716] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24eef20 is same with the state(6) to be set 00:32:01.806 [2024-12-06 17:47:43.245744] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24eef20 is same with the state(6) to be set 00:32:01.806 [2024-12-06 17:47:43.245757] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24eef20 is same with the state(6) to be set 00:32:01.806 [2024-12-06 17:47:43.245769] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24eef20 is same with the state(6) to be set 00:32:01.806 [2024-12-06 17:47:43.245782] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24eef20 is same with the state(6) to be set 00:32:01.806 [2024-12-06 17:47:43.245801] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24eef20 is same with the state(6) to be set 00:32:01.806 [2024-12-06 17:47:43.245745] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode10, 1] CQ[2024-12-06 17:47:43.245813] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24eef20 is same with transport error -6 (No such device or address) on qpair id 3 00:32:01.806 the state(6) to be set 00:32:01.806 [2024-12-06 17:47:43.245827] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24eef20 is same with the state(6) to be set 00:32:01.806 [2024-12-06 17:47:43.245839] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24eef20 is same with the state(6) to be set 00:32:01.806 [2024-12-06 17:47:43.245852] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24eef20 is same with the state(6) to be set 00:32:01.806 [2024-12-06 17:47:43.245864] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24eef20 is same with the state(6) to be set 00:32:01.806 [2024-12-06 17:47:43.245876] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24eef20 is same with the state(6) to be set 00:32:01.806 [2024-12-06 17:47:43.245887] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24eef20 is same with the state(6) to be set 00:32:01.806 [2024-12-06 17:47:43.245900] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24eef20 is same with the state(6) to be set 00:32:01.806 [2024-12-06 17:47:43.245913] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24eef20 is same with the state(6) to be set 00:32:01.806 [2024-12-06 17:47:43.245946] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24eef20 is same with the state(6) to be set 00:32:01.806 [2024-12-06 17:47:43.245964] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24eef20 is same with the state(6) to be set 00:32:01.806 [2024-12-06 17:47:43.245976] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24eef20 is same with the state(6) to be set 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 [2024-12-06 17:47:43.245989] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24eef20 is same with Write completed with error (sct=0, sc=8) 00:32:01.806 the state(6) to be set 00:32:01.806 starting I/O failed: -6 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 starting I/O failed: -6 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 starting I/O failed: -6 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 starting I/O failed: -6 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 starting I/O failed: -6 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 starting I/O failed: -6 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 starting I/O failed: -6 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 starting I/O failed: -6 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 Write completed with error (sct=0, sc=8) 00:32:01.806 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 [2024-12-06 17:47:43.246489] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24ef3f0 is same with the state(6) to be set 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 [2024-12-06 17:47:43.246523] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24ef3f0 is same with the state(6) to be set 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 [2024-12-06 17:47:43.246542] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24ef3f0 is same with the state(6) to be set 00:32:01.807 starting I/O failed: -6 00:32:01.807 [2024-12-06 17:47:43.246555] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24ef3f0 is same with the state(6) to be set 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 [2024-12-06 17:47:43.246661] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24ee580 is same with the state(6) to be set 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 [2024-12-06 17:47:43.246712] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24ee580 is same with Write completed with error (sct=0, sc=8) 00:32:01.807 the state(6) to be set 00:32:01.807 starting I/O failed: -6 00:32:01.807 [2024-12-06 17:47:43.246734] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24ee580 is same with the state(6) to be set 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 [2024-12-06 17:47:43.246747] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24ee580 is same with starting I/O failed: -6 00:32:01.807 the state(6) to be set 00:32:01.807 [2024-12-06 17:47:43.246761] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24ee580 is same with Write completed with error (sct=0, sc=8) 00:32:01.807 the state(6) to be set 00:32:01.807 [2024-12-06 17:47:43.246774] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24ee580 is same with the state(6) to be set 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 [2024-12-06 17:47:43.246787] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24ee580 is same with starting I/O failed: -6 00:32:01.807 the state(6) to be set 00:32:01.807 [2024-12-06 17:47:43.246800] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24ee580 is same with Write completed with error (sct=0, sc=8) 00:32:01.807 the state(6) to be set 00:32:01.807 starting I/O failed: -6 00:32:01.807 [2024-12-06 17:47:43.246826] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24ee580 is same with the state(6) to be set 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 [2024-12-06 17:47:43.246840] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24ee580 is same with starting I/O failed: -6 00:32:01.807 the state(6) to be set 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 [2024-12-06 17:47:43.247714] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode10, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.807 Write completed with error (sct=0, sc=8) 00:32:01.807 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 [2024-12-06 17:47:43.249508] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode10, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:32:01.808 NVMe io qpair process completion error 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 [2024-12-06 17:47:43.250604] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 [2024-12-06 17:47:43.251714] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.808 starting I/O failed: -6 00:32:01.808 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 [2024-12-06 17:47:43.252917] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 [2024-12-06 17:47:43.254939] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:32:01.809 NVMe io qpair process completion error 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 [2024-12-06 17:47:43.256075] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.809 starting I/O failed: -6 00:32:01.809 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 [2024-12-06 17:47:43.257137] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 [2024-12-06 17:47:43.258340] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.810 starting I/O failed: -6 00:32:01.810 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 [2024-12-06 17:47:43.260018] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:32:01.811 NVMe io qpair process completion error 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 [2024-12-06 17:47:43.261248] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 [2024-12-06 17:47:43.262211] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.811 Write completed with error (sct=0, sc=8) 00:32:01.811 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 [2024-12-06 17:47:43.263467] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 [2024-12-06 17:47:43.265616] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:32:01.812 NVMe io qpair process completion error 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 starting I/O failed: -6 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.812 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 [2024-12-06 17:47:43.266935] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 [2024-12-06 17:47:43.268037] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 [2024-12-06 17:47:43.269167] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.813 starting I/O failed: -6 00:32:01.813 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 [2024-12-06 17:47:43.271221] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:32:01.814 NVMe io qpair process completion error 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 [2024-12-06 17:47:43.272388] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 [2024-12-06 17:47:43.273490] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.814 Write completed with error (sct=0, sc=8) 00:32:01.814 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 [2024-12-06 17:47:43.274704] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 [2024-12-06 17:47:43.279098] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:32:01.815 NVMe io qpair process completion error 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 Write completed with error (sct=0, sc=8) 00:32:01.815 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 [2024-12-06 17:47:43.280469] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 [2024-12-06 17:47:43.281525] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 [2024-12-06 17:47:43.282692] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.816 Write completed with error (sct=0, sc=8) 00:32:01.816 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 [2024-12-06 17:47:43.285575] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:32:01.817 NVMe io qpair process completion error 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 [2024-12-06 17:47:43.286907] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode9, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 [2024-12-06 17:47:43.287991] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode9, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.817 starting I/O failed: -6 00:32:01.817 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 [2024-12-06 17:47:43.289105] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode9, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 [2024-12-06 17:47:43.290867] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode9, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:32:01.818 NVMe io qpair process completion error 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.818 starting I/O failed: -6 00:32:01.818 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 [2024-12-06 17:47:43.292164] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode8, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 [2024-12-06 17:47:43.293207] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode8, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 [2024-12-06 17:47:43.294384] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode8, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.819 starting I/O failed: -6 00:32:01.819 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 [2024-12-06 17:47:43.296548] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode8, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:32:01.820 NVMe io qpair process completion error 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.820 Write completed with error (sct=0, sc=8) 00:32:01.820 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 [2024-12-06 17:47:43.299433] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:32:01.821 starting I/O failed: -6 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 Write completed with error (sct=0, sc=8) 00:32:01.821 starting I/O failed: -6 00:32:01.821 [2024-12-06 17:47:43.301900] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:32:01.821 NVMe io qpair process completion error 00:32:01.821 Initializing NVMe Controllers 00:32:01.821 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode7 00:32:01.821 Controller IO queue size 128, less than required. 00:32:01.821 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:32:01.821 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode9 00:32:01.821 Controller IO queue size 128, less than required. 00:32:01.821 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:32:01.821 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode8 00:32:01.821 Controller IO queue size 128, less than required. 00:32:01.821 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:32:01.821 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode2 00:32:01.821 Controller IO queue size 128, less than required. 00:32:01.821 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:32:01.822 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode10 00:32:01.822 Controller IO queue size 128, less than required. 00:32:01.822 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:32:01.822 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:32:01.822 Controller IO queue size 128, less than required. 00:32:01.822 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:32:01.822 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode4 00:32:01.822 Controller IO queue size 128, less than required. 00:32:01.822 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:32:01.822 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode5 00:32:01.822 Controller IO queue size 128, less than required. 00:32:01.822 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:32:01.822 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode6 00:32:01.822 Controller IO queue size 128, less than required. 00:32:01.822 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:32:01.822 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode3 00:32:01.822 Controller IO queue size 128, less than required. 00:32:01.822 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:32:01.822 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode7) NSID 1 with lcore 0 00:32:01.822 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode9) NSID 1 with lcore 0 00:32:01.822 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode8) NSID 1 with lcore 0 00:32:01.822 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode2) NSID 1 with lcore 0 00:32:01.822 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode10) NSID 1 with lcore 0 00:32:01.822 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:32:01.822 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode4) NSID 1 with lcore 0 00:32:01.822 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode5) NSID 1 with lcore 0 00:32:01.822 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode6) NSID 1 with lcore 0 00:32:01.822 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode3) NSID 1 with lcore 0 00:32:01.822 Initialization complete. Launching workers. 00:32:01.822 ======================================================== 00:32:01.822 Latency(us) 00:32:01.822 Device Information : IOPS MiB/s Average min max 00:32:01.822 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode7) NSID 1 from core 0: 1840.69 79.09 69561.42 1050.88 129063.87 00:32:01.822 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode9) NSID 1 from core 0: 1830.95 78.67 69973.90 953.65 132724.37 00:32:01.822 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode8) NSID 1 from core 0: 1834.85 78.84 69851.62 1037.96 117126.33 00:32:01.822 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode2) NSID 1 from core 0: 1857.56 79.82 69021.12 891.48 115480.80 00:32:01.822 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode10) NSID 1 from core 0: 1782.27 76.58 71078.59 894.00 115661.94 00:32:01.822 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 1816.24 78.04 69773.43 1120.55 113503.44 00:32:01.822 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode4) NSID 1 from core 0: 1837.01 78.93 69012.29 1057.52 115903.96 00:32:01.822 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode5) NSID 1 from core 0: 1825.54 78.44 69470.33 889.75 118913.80 00:32:01.822 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode6) NSID 1 from core 0: 1826.41 78.48 69469.49 1123.49 121530.65 00:32:01.822 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode3) NSID 1 from core 0: 1829.87 78.63 69367.75 822.10 123964.46 00:32:01.822 ======================================================== 00:32:01.822 Total : 18281.39 785.53 69653.09 822.10 132724.37 00:32:01.822 00:32:01.822 [2024-12-06 17:47:43.307600] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1dad140 is same with the state(6) to be set 00:32:01.822 [2024-12-06 17:47:43.307704] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1dad7a0 is same with the state(6) to be set 00:32:01.822 [2024-12-06 17:47:43.307765] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1dad470 is same with the state(6) to be set 00:32:01.822 [2024-12-06 17:47:43.307826] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1dab190 is same with the state(6) to be set 00:32:01.822 [2024-12-06 17:47:43.307885] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1daafb0 is same with the state(6) to be set 00:32:01.822 [2024-12-06 17:47:43.307943] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1db0b30 is same with the state(6) to be set 00:32:01.822 [2024-12-06 17:47:43.308003] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1dab6a0 is same with the state(6) to be set 00:32:01.822 [2024-12-06 17:47:43.308059] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1dab9d0 is same with the state(6) to be set 00:32:01.822 [2024-12-06 17:47:43.308118] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1dace10 is same with the state(6) to be set 00:32:01.822 [2024-12-06 17:47:43.308176] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1dab370 is same with the state(6) to be set 00:32:01.822 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf: errors occurred 00:32:02.081 17:47:43 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@156 -- # sleep 1 00:32:03.019 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@158 -- # NOT wait 338320 00:32:03.019 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@652 -- # local es=0 00:32:03.019 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@654 -- # valid_exec_arg wait 338320 00:32:03.019 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@640 -- # local arg=wait 00:32:03.019 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:32:03.019 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@644 -- # type -t wait 00:32:03.019 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:32:03.019 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@655 -- # wait 338320 00:32:03.019 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@655 -- # es=1 00:32:03.019 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:32:03.019 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:32:03.019 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:32:03.019 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@159 -- # stoptarget 00:32:03.019 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@42 -- # rm -f ./local-job0-0-verify.state 00:32:03.019 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@43 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:32:03.019 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@44 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:32:03.019 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@46 -- # nvmftestfini 00:32:03.019 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@516 -- # nvmfcleanup 00:32:03.019 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@121 -- # sync 00:32:03.019 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:32:03.019 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@124 -- # set +e 00:32:03.019 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@125 -- # for i in {1..20} 00:32:03.020 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:32:03.020 rmmod nvme_tcp 00:32:03.020 rmmod nvme_fabrics 00:32:03.020 rmmod nvme_keyring 00:32:03.020 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:32:03.020 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@128 -- # set -e 00:32:03.020 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@129 -- # return 0 00:32:03.020 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@517 -- # '[' -n 338256 ']' 00:32:03.020 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@518 -- # killprocess 338256 00:32:03.020 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@954 -- # '[' -z 338256 ']' 00:32:03.020 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@958 -- # kill -0 338256 00:32:03.020 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 958: kill: (338256) - No such process 00:32:03.020 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@981 -- # echo 'Process with pid 338256 is not found' 00:32:03.020 Process with pid 338256 is not found 00:32:03.020 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:32:03.020 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:32:03.020 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:32:03.020 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@297 -- # iptr 00:32:03.020 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@791 -- # iptables-save 00:32:03.020 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:32:03.020 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@791 -- # iptables-restore 00:32:03.020 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:32:03.020 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@302 -- # remove_spdk_ns 00:32:03.020 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:03.020 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:03.020 17:47:44 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:05.552 17:47:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:32:05.552 00:32:05.552 real 0m9.807s 00:32:05.552 user 0m23.780s 00:32:05.552 sys 0m5.605s 00:32:05.552 17:47:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@1130 -- # xtrace_disable 00:32:05.552 17:47:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:32:05.552 ************************************ 00:32:05.552 END TEST nvmf_shutdown_tc4 00:32:05.552 ************************************ 00:32:05.552 17:47:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@170 -- # trap - SIGINT SIGTERM EXIT 00:32:05.552 00:32:05.552 real 0m37.175s 00:32:05.552 user 1m40.528s 00:32:05.552 sys 0m11.919s 00:32:05.552 17:47:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1130 -- # xtrace_disable 00:32:05.552 17:47:46 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@10 -- # set +x 00:32:05.552 ************************************ 00:32:05.552 END TEST nvmf_shutdown 00:32:05.552 ************************************ 00:32:05.552 17:47:46 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@67 -- # run_test nvmf_nsid /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nsid.sh --transport=tcp 00:32:05.552 17:47:46 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:32:05.552 17:47:46 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1111 -- # xtrace_disable 00:32:05.552 17:47:46 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:32:05.552 ************************************ 00:32:05.552 START TEST nvmf_nsid 00:32:05.552 ************************************ 00:32:05.552 17:47:46 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nsid.sh --transport=tcp 00:32:05.552 * Looking for test storage... 00:32:05.552 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:32:05.552 17:47:46 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:32:05.552 17:47:46 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1711 -- # lcov --version 00:32:05.552 17:47:46 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@336 -- # IFS=.-: 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@336 -- # read -ra ver1 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@337 -- # IFS=.-: 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@337 -- # read -ra ver2 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@338 -- # local 'op=<' 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@340 -- # ver1_l=2 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@341 -- # ver2_l=1 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@344 -- # case "$op" in 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@345 -- # : 1 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@364 -- # (( v = 0 )) 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@365 -- # decimal 1 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@353 -- # local d=1 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@355 -- # echo 1 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@365 -- # ver1[v]=1 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@366 -- # decimal 2 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@353 -- # local d=2 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@355 -- # echo 2 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@366 -- # ver2[v]=2 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@368 -- # return 0 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:32:05.552 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:05.552 --rc genhtml_branch_coverage=1 00:32:05.552 --rc genhtml_function_coverage=1 00:32:05.552 --rc genhtml_legend=1 00:32:05.552 --rc geninfo_all_blocks=1 00:32:05.552 --rc geninfo_unexecuted_blocks=1 00:32:05.552 00:32:05.552 ' 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:32:05.552 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:05.552 --rc genhtml_branch_coverage=1 00:32:05.552 --rc genhtml_function_coverage=1 00:32:05.552 --rc genhtml_legend=1 00:32:05.552 --rc geninfo_all_blocks=1 00:32:05.552 --rc geninfo_unexecuted_blocks=1 00:32:05.552 00:32:05.552 ' 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:32:05.552 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:05.552 --rc genhtml_branch_coverage=1 00:32:05.552 --rc genhtml_function_coverage=1 00:32:05.552 --rc genhtml_legend=1 00:32:05.552 --rc geninfo_all_blocks=1 00:32:05.552 --rc geninfo_unexecuted_blocks=1 00:32:05.552 00:32:05.552 ' 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:32:05.552 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:05.552 --rc genhtml_branch_coverage=1 00:32:05.552 --rc genhtml_function_coverage=1 00:32:05.552 --rc genhtml_legend=1 00:32:05.552 --rc geninfo_all_blocks=1 00:32:05.552 --rc geninfo_unexecuted_blocks=1 00:32:05.552 00:32:05.552 ' 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@7 -- # uname -s 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@15 -- # shopt -s extglob 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:32:05.552 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- paths/export.sh@5 -- # export PATH 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@51 -- # : 0 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:32:05.553 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@55 -- # have_pci_nics=0 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@11 -- # subnqn1=nqn.2024-10.io.spdk:cnode0 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@12 -- # subnqn2=nqn.2024-10.io.spdk:cnode1 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@13 -- # subnqn3=nqn.2024-10.io.spdk:cnode2 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@14 -- # tgt2sock=/var/tmp/tgt2.sock 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@15 -- # tgt2pid= 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@46 -- # nvmftestinit 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@476 -- # prepare_net_devs 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@438 -- # local -g is_hw=no 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@440 -- # remove_spdk_ns 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@309 -- # xtrace_disable 00:32:05.553 17:47:47 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@10 -- # set +x 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@315 -- # pci_devs=() 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@315 -- # local -a pci_devs 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@316 -- # pci_net_devs=() 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@317 -- # pci_drivers=() 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@317 -- # local -A pci_drivers 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@319 -- # net_devs=() 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@319 -- # local -ga net_devs 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@320 -- # e810=() 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@320 -- # local -ga e810 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@321 -- # x722=() 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@321 -- # local -ga x722 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@322 -- # mlx=() 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@322 -- # local -ga mlx 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:32:07.451 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:32:07.452 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:32:07.452 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@418 -- # [[ up == up ]] 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:32:07.452 Found net devices under 0000:0a:00.0: cvl_0_0 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@418 -- # [[ up == up ]] 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:32:07.452 Found net devices under 0000:0a:00.1: cvl_0_1 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@442 -- # is_hw=yes 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:32:07.452 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:32:07.710 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:32:07.710 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:32:07.710 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:32:07.710 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:32:07.710 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:32:07.710 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:32:07.710 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:32:07.710 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:32:07.710 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:32:07.710 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:32:07.710 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:32:07.710 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.228 ms 00:32:07.710 00:32:07.710 --- 10.0.0.2 ping statistics --- 00:32:07.710 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:32:07.710 rtt min/avg/max/mdev = 0.228/0.228/0.228/0.000 ms 00:32:07.710 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:32:07.710 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:32:07.710 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.148 ms 00:32:07.710 00:32:07.710 --- 10.0.0.1 ping statistics --- 00:32:07.710 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:32:07.710 rtt min/avg/max/mdev = 0.148/0.148/0.148/0.000 ms 00:32:07.710 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:32:07.710 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@450 -- # return 0 00:32:07.710 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:32:07.710 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:32:07.710 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:32:07.710 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:32:07.710 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:32:07.710 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:32:07.710 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:32:07.710 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@47 -- # nvmfappstart -m 1 00:32:07.710 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:32:07.710 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@726 -- # xtrace_disable 00:32:07.710 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@10 -- # set +x 00:32:07.710 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@509 -- # nvmfpid=341054 00:32:07.710 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 1 00:32:07.710 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@510 -- # waitforlisten 341054 00:32:07.710 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@835 -- # '[' -z 341054 ']' 00:32:07.711 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:32:07.711 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@840 -- # local max_retries=100 00:32:07.711 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:32:07.711 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:32:07.711 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@844 -- # xtrace_disable 00:32:07.711 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@10 -- # set +x 00:32:07.711 [2024-12-06 17:47:49.481300] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:32:07.711 [2024-12-06 17:47:49.481381] [ DPDK EAL parameters: nvmf -c 1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:32:07.968 [2024-12-06 17:47:49.553536] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:07.968 [2024-12-06 17:47:49.601616] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:32:07.968 [2024-12-06 17:47:49.601702] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:32:07.968 [2024-12-06 17:47:49.601725] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:32:07.968 [2024-12-06 17:47:49.601736] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:32:07.968 [2024-12-06 17:47:49.601746] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:32:07.968 [2024-12-06 17:47:49.602357] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:32:07.968 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:32:07.968 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@868 -- # return 0 00:32:07.968 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:32:07.969 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@732 -- # xtrace_disable 00:32:07.969 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@10 -- # set +x 00:32:07.969 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:32:07.969 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@49 -- # trap cleanup SIGINT SIGTERM EXIT 00:32:07.969 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@52 -- # tgt2pid=341148 00:32:07.969 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@51 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 2 -r /var/tmp/tgt2.sock 00:32:07.969 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@54 -- # tgt1addr=10.0.0.2 00:32:07.969 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@55 -- # get_main_ns_ip 00:32:07.969 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@769 -- # local ip 00:32:07.969 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@770 -- # ip_candidates=() 00:32:07.969 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@770 -- # local -A ip_candidates 00:32:07.969 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:32:07.969 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:32:07.969 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:32:07.969 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:32:07.969 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:32:07.969 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:32:07.969 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:32:07.969 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@55 -- # tgt2addr=10.0.0.1 00:32:07.969 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@56 -- # uuidgen 00:32:07.969 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@56 -- # ns1uuid=7c8a8d4e-9eb3-4811-be2c-cf3366110b1c 00:32:07.969 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@57 -- # uuidgen 00:32:07.969 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@57 -- # ns2uuid=278fba43-f3bf-436b-8d13-d4edd3b125ee 00:32:07.969 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@58 -- # uuidgen 00:32:07.969 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@58 -- # ns3uuid=b80a4fce-2865-4272-9fac-bd6754dce778 00:32:07.969 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@63 -- # rpc_cmd 00:32:07.969 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:07.969 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@10 -- # set +x 00:32:07.969 null0 00:32:07.969 null1 00:32:07.969 null2 00:32:07.969 [2024-12-06 17:47:49.788912] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:32:07.969 [2024-12-06 17:47:49.804161] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:32:07.969 [2024-12-06 17:47:49.804251] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid341148 ] 00:32:08.227 [2024-12-06 17:47:49.813122] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:32:08.227 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:08.227 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@79 -- # waitforlisten 341148 /var/tmp/tgt2.sock 00:32:08.227 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@835 -- # '[' -z 341148 ']' 00:32:08.227 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/tgt2.sock 00:32:08.227 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@840 -- # local max_retries=100 00:32:08.227 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/tgt2.sock...' 00:32:08.227 Waiting for process to start up and listen on UNIX domain socket /var/tmp/tgt2.sock... 00:32:08.227 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@844 -- # xtrace_disable 00:32:08.227 17:47:49 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@10 -- # set +x 00:32:08.227 [2024-12-06 17:47:49.872472] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:08.227 [2024-12-06 17:47:49.919878] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:32:08.484 17:47:50 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:32:08.484 17:47:50 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@868 -- # return 0 00:32:08.484 17:47:50 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@80 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/tgt2.sock 00:32:08.742 [2024-12-06 17:47:50.567465] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:32:08.999 [2024-12-06 17:47:50.583676] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.1 port 4421 *** 00:32:08.999 nvme0n1 nvme0n2 00:32:08.999 nvme1n1 00:32:08.999 17:47:50 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@94 -- # nvme_connect 00:32:08.999 17:47:50 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@23 -- # local ctrlr 00:32:08.999 17:47:50 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@25 -- # nvme connect -t tcp -a 10.0.0.1 -s 4421 -n nqn.2024-10.io.spdk:cnode2 --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 00:32:09.566 17:47:51 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@28 -- # for ctrlr in /sys/class/nvme/nvme* 00:32:09.566 17:47:51 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@30 -- # [[ -e /sys/class/nvme/nvme0/subsysnqn ]] 00:32:09.566 17:47:51 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@30 -- # [[ nqn.2024-10.io.spdk:cnode2 == \n\q\n\.\2\0\2\4\-\1\0\.\i\o\.\s\p\d\k\:\c\n\o\d\e\2 ]] 00:32:09.566 17:47:51 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@31 -- # echo nvme0 00:32:09.566 17:47:51 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@32 -- # return 0 00:32:09.566 17:47:51 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@94 -- # ctrlr=nvme0 00:32:09.566 17:47:51 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@95 -- # waitforblk nvme0n1 00:32:09.566 17:47:51 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1239 -- # local i=0 00:32:09.566 17:47:51 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1240 -- # lsblk -l -o NAME 00:32:09.566 17:47:51 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1240 -- # grep -q -w nvme0n1 00:32:09.566 17:47:51 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1241 -- # '[' 0 -lt 15 ']' 00:32:09.566 17:47:51 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1242 -- # i=1 00:32:09.566 17:47:51 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1243 -- # sleep 1 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1240 -- # lsblk -l -o NAME 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1240 -- # grep -q -w nvme0n1 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1246 -- # lsblk -l -o NAME 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1246 -- # grep -q -w nvme0n1 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1250 -- # return 0 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@96 -- # uuid2nguid 7c8a8d4e-9eb3-4811-be2c-cf3366110b1c 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@787 -- # tr -d - 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@96 -- # nvme_get_nguid nvme0 1 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@40 -- # local ctrlr=nvme0 nsid=1 nguid 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@42 -- # nvme id-ns /dev/nvme0n1 -o json 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@42 -- # jq -r .nguid 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@42 -- # nguid=7c8a8d4e9eb34811be2ccf3366110b1c 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@43 -- # echo 7C8A8D4E9EB34811BE2CCF3366110B1C 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@96 -- # [[ 7C8A8D4E9EB34811BE2CCF3366110B1C == \7\C\8\A\8\D\4\E\9\E\B\3\4\8\1\1\B\E\2\C\C\F\3\3\6\6\1\1\0\B\1\C ]] 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@97 -- # waitforblk nvme0n2 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1239 -- # local i=0 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1240 -- # lsblk -l -o NAME 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1240 -- # grep -q -w nvme0n2 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1246 -- # lsblk -l -o NAME 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1246 -- # grep -q -w nvme0n2 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1250 -- # return 0 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@98 -- # uuid2nguid 278fba43-f3bf-436b-8d13-d4edd3b125ee 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@787 -- # tr -d - 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@98 -- # nvme_get_nguid nvme0 2 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@40 -- # local ctrlr=nvme0 nsid=2 nguid 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@42 -- # nvme id-ns /dev/nvme0n2 -o json 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@42 -- # jq -r .nguid 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@42 -- # nguid=278fba43f3bf436b8d13d4edd3b125ee 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@43 -- # echo 278FBA43F3BF436B8D13D4EDD3B125EE 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@98 -- # [[ 278FBA43F3BF436B8D13D4EDD3B125EE == \2\7\8\F\B\A\4\3\F\3\B\F\4\3\6\B\8\D\1\3\D\4\E\D\D\3\B\1\2\5\E\E ]] 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@99 -- # waitforblk nvme0n3 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1239 -- # local i=0 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1240 -- # lsblk -l -o NAME 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1240 -- # grep -q -w nvme0n3 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1246 -- # lsblk -l -o NAME 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1246 -- # grep -q -w nvme0n3 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1250 -- # return 0 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@100 -- # uuid2nguid b80a4fce-2865-4272-9fac-bd6754dce778 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@787 -- # tr -d - 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@100 -- # nvme_get_nguid nvme0 3 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@40 -- # local ctrlr=nvme0 nsid=3 nguid 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@42 -- # nvme id-ns /dev/nvme0n3 -o json 00:32:10.523 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@42 -- # jq -r .nguid 00:32:10.781 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@42 -- # nguid=b80a4fce286542729facbd6754dce778 00:32:10.781 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@43 -- # echo B80A4FCE286542729FACBD6754DCE778 00:32:10.781 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@100 -- # [[ B80A4FCE286542729FACBD6754DCE778 == \B\8\0\A\4\F\C\E\2\8\6\5\4\2\7\2\9\F\A\C\B\D\6\7\5\4\D\C\E\7\7\8 ]] 00:32:10.781 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@101 -- # nvme disconnect -d /dev/nvme0 00:32:10.781 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@103 -- # trap - SIGINT SIGTERM EXIT 00:32:10.781 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@104 -- # cleanup 00:32:10.781 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@18 -- # killprocess 341148 00:32:10.781 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@954 -- # '[' -z 341148 ']' 00:32:10.781 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@958 -- # kill -0 341148 00:32:10.781 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@959 -- # uname 00:32:10.781 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:32:10.781 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 341148 00:32:10.781 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:32:10.781 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:32:10.781 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@972 -- # echo 'killing process with pid 341148' 00:32:10.781 killing process with pid 341148 00:32:10.781 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@973 -- # kill 341148 00:32:10.781 17:47:52 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@978 -- # wait 341148 00:32:11.346 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- target/nsid.sh@19 -- # nvmftestfini 00:32:11.346 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@516 -- # nvmfcleanup 00:32:11.346 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@121 -- # sync 00:32:11.346 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:32:11.346 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@124 -- # set +e 00:32:11.346 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@125 -- # for i in {1..20} 00:32:11.346 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:32:11.346 rmmod nvme_tcp 00:32:11.346 rmmod nvme_fabrics 00:32:11.346 rmmod nvme_keyring 00:32:11.346 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:32:11.346 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@128 -- # set -e 00:32:11.346 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@129 -- # return 0 00:32:11.346 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@517 -- # '[' -n 341054 ']' 00:32:11.346 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@518 -- # killprocess 341054 00:32:11.346 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@954 -- # '[' -z 341054 ']' 00:32:11.346 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@958 -- # kill -0 341054 00:32:11.346 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@959 -- # uname 00:32:11.346 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:32:11.346 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 341054 00:32:11.346 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:32:11.346 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:32:11.346 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@972 -- # echo 'killing process with pid 341054' 00:32:11.346 killing process with pid 341054 00:32:11.346 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@973 -- # kill 341054 00:32:11.346 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@978 -- # wait 341054 00:32:11.605 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:32:11.605 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:32:11.605 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:32:11.605 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@297 -- # iptr 00:32:11.605 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@791 -- # iptables-save 00:32:11.605 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:32:11.605 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@791 -- # iptables-restore 00:32:11.605 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:32:11.605 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@302 -- # remove_spdk_ns 00:32:11.605 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:11.605 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:11.605 17:47:53 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:13.514 17:47:55 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:32:13.514 00:32:13.514 real 0m8.396s 00:32:13.514 user 0m8.215s 00:32:13.514 sys 0m2.704s 00:32:13.514 17:47:55 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@1130 -- # xtrace_disable 00:32:13.514 17:47:55 nvmf_tcp.nvmf_target_extra.nvmf_nsid -- common/autotest_common.sh@10 -- # set +x 00:32:13.515 ************************************ 00:32:13.515 END TEST nvmf_nsid 00:32:13.515 ************************************ 00:32:13.777 17:47:55 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:32:13.777 00:32:13.777 real 18m8.291s 00:32:13.777 user 50m21.463s 00:32:13.777 sys 3m55.553s 00:32:13.777 17:47:55 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1130 -- # xtrace_disable 00:32:13.777 17:47:55 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:32:13.777 ************************************ 00:32:13.777 END TEST nvmf_target_extra 00:32:13.777 ************************************ 00:32:13.777 17:47:55 nvmf_tcp -- nvmf/nvmf.sh@16 -- # run_test nvmf_host /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_host.sh --transport=tcp 00:32:13.777 17:47:55 nvmf_tcp -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:32:13.777 17:47:55 nvmf_tcp -- common/autotest_common.sh@1111 -- # xtrace_disable 00:32:13.777 17:47:55 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:32:13.777 ************************************ 00:32:13.777 START TEST nvmf_host 00:32:13.777 ************************************ 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_host.sh --transport=tcp 00:32:13.777 * Looking for test storage... 00:32:13.777 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1711 -- # lcov --version 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@333 -- # local ver1 ver1_l 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@334 -- # local ver2 ver2_l 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@336 -- # IFS=.-: 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@336 -- # read -ra ver1 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@337 -- # IFS=.-: 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@337 -- # read -ra ver2 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@338 -- # local 'op=<' 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@340 -- # ver1_l=2 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@341 -- # ver2_l=1 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@344 -- # case "$op" in 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@345 -- # : 1 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@364 -- # (( v = 0 )) 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@365 -- # decimal 1 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@353 -- # local d=1 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@355 -- # echo 1 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@365 -- # ver1[v]=1 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@366 -- # decimal 2 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@353 -- # local d=2 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@355 -- # echo 2 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@366 -- # ver2[v]=2 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@368 -- # return 0 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:32:13.777 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:13.777 --rc genhtml_branch_coverage=1 00:32:13.777 --rc genhtml_function_coverage=1 00:32:13.777 --rc genhtml_legend=1 00:32:13.777 --rc geninfo_all_blocks=1 00:32:13.777 --rc geninfo_unexecuted_blocks=1 00:32:13.777 00:32:13.777 ' 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:32:13.777 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:13.777 --rc genhtml_branch_coverage=1 00:32:13.777 --rc genhtml_function_coverage=1 00:32:13.777 --rc genhtml_legend=1 00:32:13.777 --rc geninfo_all_blocks=1 00:32:13.777 --rc geninfo_unexecuted_blocks=1 00:32:13.777 00:32:13.777 ' 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:32:13.777 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:13.777 --rc genhtml_branch_coverage=1 00:32:13.777 --rc genhtml_function_coverage=1 00:32:13.777 --rc genhtml_legend=1 00:32:13.777 --rc geninfo_all_blocks=1 00:32:13.777 --rc geninfo_unexecuted_blocks=1 00:32:13.777 00:32:13.777 ' 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:32:13.777 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:13.777 --rc genhtml_branch_coverage=1 00:32:13.777 --rc genhtml_function_coverage=1 00:32:13.777 --rc genhtml_legend=1 00:32:13.777 --rc geninfo_all_blocks=1 00:32:13.777 --rc geninfo_unexecuted_blocks=1 00:32:13.777 00:32:13.777 ' 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- nvmf/common.sh@7 -- # uname -s 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@15 -- # shopt -s extglob 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- paths/export.sh@5 -- # export PATH 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- nvmf/common.sh@51 -- # : 0 00:32:13.777 17:47:55 nvmf_tcp.nvmf_host -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:32:13.778 17:47:55 nvmf_tcp.nvmf_host -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:32:13.778 17:47:55 nvmf_tcp.nvmf_host -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:32:13.778 17:47:55 nvmf_tcp.nvmf_host -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:32:13.778 17:47:55 nvmf_tcp.nvmf_host -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:32:13.778 17:47:55 nvmf_tcp.nvmf_host -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:32:13.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:32:13.778 17:47:55 nvmf_tcp.nvmf_host -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:32:13.778 17:47:55 nvmf_tcp.nvmf_host -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:32:13.778 17:47:55 nvmf_tcp.nvmf_host -- nvmf/common.sh@55 -- # have_pci_nics=0 00:32:13.778 17:47:55 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@11 -- # trap 'exit 1' SIGINT SIGTERM EXIT 00:32:13.778 17:47:55 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@13 -- # TEST_ARGS=("$@") 00:32:13.778 17:47:55 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@15 -- # [[ 0 -eq 0 ]] 00:32:13.778 17:47:55 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@16 -- # run_test nvmf_multicontroller /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/multicontroller.sh --transport=tcp 00:32:13.778 17:47:55 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:32:13.778 17:47:55 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:32:13.778 17:47:55 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:32:13.778 ************************************ 00:32:13.778 START TEST nvmf_multicontroller 00:32:13.778 ************************************ 00:32:13.778 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/multicontroller.sh --transport=tcp 00:32:14.036 * Looking for test storage... 00:32:14.036 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1711 -- # lcov --version 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@333 -- # local ver1 ver1_l 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@334 -- # local ver2 ver2_l 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@336 -- # IFS=.-: 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@336 -- # read -ra ver1 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@337 -- # IFS=.-: 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@337 -- # read -ra ver2 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@338 -- # local 'op=<' 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@340 -- # ver1_l=2 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@341 -- # ver2_l=1 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@344 -- # case "$op" in 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@345 -- # : 1 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@364 -- # (( v = 0 )) 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@365 -- # decimal 1 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@353 -- # local d=1 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@355 -- # echo 1 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@365 -- # ver1[v]=1 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@366 -- # decimal 2 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@353 -- # local d=2 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@355 -- # echo 2 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@366 -- # ver2[v]=2 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@368 -- # return 0 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:32:14.036 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:14.036 --rc genhtml_branch_coverage=1 00:32:14.036 --rc genhtml_function_coverage=1 00:32:14.036 --rc genhtml_legend=1 00:32:14.036 --rc geninfo_all_blocks=1 00:32:14.036 --rc geninfo_unexecuted_blocks=1 00:32:14.036 00:32:14.036 ' 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:32:14.036 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:14.036 --rc genhtml_branch_coverage=1 00:32:14.036 --rc genhtml_function_coverage=1 00:32:14.036 --rc genhtml_legend=1 00:32:14.036 --rc geninfo_all_blocks=1 00:32:14.036 --rc geninfo_unexecuted_blocks=1 00:32:14.036 00:32:14.036 ' 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:32:14.036 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:14.036 --rc genhtml_branch_coverage=1 00:32:14.036 --rc genhtml_function_coverage=1 00:32:14.036 --rc genhtml_legend=1 00:32:14.036 --rc geninfo_all_blocks=1 00:32:14.036 --rc geninfo_unexecuted_blocks=1 00:32:14.036 00:32:14.036 ' 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:32:14.036 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:14.036 --rc genhtml_branch_coverage=1 00:32:14.036 --rc genhtml_function_coverage=1 00:32:14.036 --rc genhtml_legend=1 00:32:14.036 --rc geninfo_all_blocks=1 00:32:14.036 --rc geninfo_unexecuted_blocks=1 00:32:14.036 00:32:14.036 ' 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@7 -- # uname -s 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@15 -- # shopt -s extglob 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- paths/export.sh@5 -- # export PATH 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@51 -- # : 0 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:32:14.036 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:32:14.036 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:32:14.037 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:32:14.037 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@55 -- # have_pci_nics=0 00:32:14.037 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@11 -- # MALLOC_BDEV_SIZE=64 00:32:14.037 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:32:14.037 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@13 -- # NVMF_HOST_FIRST_PORT=60000 00:32:14.037 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@14 -- # NVMF_HOST_SECOND_PORT=60001 00:32:14.037 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@16 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:32:14.037 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@18 -- # '[' tcp == rdma ']' 00:32:14.037 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@23 -- # nvmftestinit 00:32:14.037 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:32:14.037 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:32:14.037 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@476 -- # prepare_net_devs 00:32:14.037 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@438 -- # local -g is_hw=no 00:32:14.037 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@440 -- # remove_spdk_ns 00:32:14.037 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:14.037 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:14.037 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:14.037 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:32:14.037 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:32:14.037 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@309 -- # xtrace_disable 00:32:14.037 17:47:55 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@315 -- # pci_devs=() 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@315 -- # local -a pci_devs 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@316 -- # pci_net_devs=() 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@317 -- # pci_drivers=() 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@317 -- # local -A pci_drivers 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@319 -- # net_devs=() 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@319 -- # local -ga net_devs 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@320 -- # e810=() 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@320 -- # local -ga e810 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@321 -- # x722=() 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@321 -- # local -ga x722 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@322 -- # mlx=() 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@322 -- # local -ga mlx 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:32:16.566 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:32:16.566 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@418 -- # [[ up == up ]] 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:32:16.566 Found net devices under 0000:0a:00.0: cvl_0_0 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@418 -- # [[ up == up ]] 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:32:16.566 Found net devices under 0000:0a:00.1: cvl_0_1 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@442 -- # is_hw=yes 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:32:16.566 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:32:16.567 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:32:16.567 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:32:16.567 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:32:16.567 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:32:16.567 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:32:16.567 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:32:16.567 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:32:16.567 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:32:16.567 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:32:16.567 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:32:16.567 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:32:16.567 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:32:16.567 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:32:16.567 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:32:16.567 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:32:16.567 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:32:16.567 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:32:16.567 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:32:16.567 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:32:16.567 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:32:16.567 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:32:16.567 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:32:16.567 17:47:57 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:32:16.567 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:32:16.567 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.192 ms 00:32:16.567 00:32:16.567 --- 10.0.0.2 ping statistics --- 00:32:16.567 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:32:16.567 rtt min/avg/max/mdev = 0.192/0.192/0.192/0.000 ms 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:32:16.567 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:32:16.567 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.118 ms 00:32:16.567 00:32:16.567 --- 10.0.0.1 ping statistics --- 00:32:16.567 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:32:16.567 rtt min/avg/max/mdev = 0.118/0.118/0.118/0.000 ms 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@450 -- # return 0 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@25 -- # nvmfappstart -m 0xE 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@726 -- # xtrace_disable 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@509 -- # nvmfpid=343638 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@510 -- # waitforlisten 343638 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@835 -- # '[' -z 343638 ']' 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@840 -- # local max_retries=100 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:32:16.567 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@844 -- # xtrace_disable 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:16.567 [2024-12-06 17:47:58.085287] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:32:16.567 [2024-12-06 17:47:58.085381] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:32:16.567 [2024-12-06 17:47:58.158356] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:32:16.567 [2024-12-06 17:47:58.204377] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:32:16.567 [2024-12-06 17:47:58.204427] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:32:16.567 [2024-12-06 17:47:58.204450] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:32:16.567 [2024-12-06 17:47:58.204460] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:32:16.567 [2024-12-06 17:47:58.204470] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:32:16.567 [2024-12-06 17:47:58.205945] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:32:16.567 [2024-12-06 17:47:58.206005] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:32:16.567 [2024-12-06 17:47:58.206009] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@868 -- # return 0 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@732 -- # xtrace_disable 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@27 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:16.567 [2024-12-06 17:47:58.355258] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@29 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:16.567 Malloc0 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@30 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.567 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:16.825 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.825 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@31 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:32:16.825 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.825 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:16.825 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.825 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@33 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:32:16.825 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.825 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:16.825 [2024-12-06 17:47:58.420233] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:32:16.825 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.825 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@34 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:32:16.825 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.825 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:16.825 [2024-12-06 17:47:58.428088] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:32:16.825 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.825 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@36 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:32:16.825 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.825 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:16.825 Malloc1 00:32:16.825 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.825 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@37 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 -a -s SPDK00000000000002 00:32:16.825 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.825 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:16.825 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.825 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@38 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 Malloc1 00:32:16.825 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.825 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:16.826 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.826 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@40 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:32:16.826 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.826 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:16.826 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.826 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@41 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4421 00:32:16.826 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:16.826 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:16.826 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:16.826 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@44 -- # bdevperf_pid=343661 00:32:16.826 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@46 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; pap "$testdir/try.txt"; killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:32:16.826 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w write -t 1 -f 00:32:16.826 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@47 -- # waitforlisten 343661 /var/tmp/bdevperf.sock 00:32:16.826 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@835 -- # '[' -z 343661 ']' 00:32:16.826 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:32:16.826 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@840 -- # local max_retries=100 00:32:16.826 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:32:16.826 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:32:16.826 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@844 -- # xtrace_disable 00:32:16.826 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@868 -- # return 0 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@50 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:17.084 NVMe0n1 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@54 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@54 -- # grep -c NVMe 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:17.084 1 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@60 -- # NOT rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -q nqn.2021-09-7.io.spdk:00001 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@652 -- # local es=0 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -q nqn.2021-09-7.io.spdk:00001 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@655 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -q nqn.2021-09-7.io.spdk:00001 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:17.084 request: 00:32:17.084 { 00:32:17.084 "name": "NVMe0", 00:32:17.084 "trtype": "tcp", 00:32:17.084 "traddr": "10.0.0.2", 00:32:17.084 "adrfam": "ipv4", 00:32:17.084 "trsvcid": "4420", 00:32:17.084 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:32:17.084 "hostnqn": "nqn.2021-09-7.io.spdk:00001", 00:32:17.084 "hostaddr": "10.0.0.1", 00:32:17.084 "prchk_reftag": false, 00:32:17.084 "prchk_guard": false, 00:32:17.084 "hdgst": false, 00:32:17.084 "ddgst": false, 00:32:17.084 "allow_unrecognized_csi": false, 00:32:17.084 "method": "bdev_nvme_attach_controller", 00:32:17.084 "req_id": 1 00:32:17.084 } 00:32:17.084 Got JSON-RPC error response 00:32:17.084 response: 00:32:17.084 { 00:32:17.084 "code": -114, 00:32:17.084 "message": "A controller named NVMe0 already exists with the specified network path" 00:32:17.084 } 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@655 -- # es=1 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@65 -- # NOT rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode2 -i 10.0.0.1 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@652 -- # local es=0 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode2 -i 10.0.0.1 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@655 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode2 -i 10.0.0.1 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:17.084 request: 00:32:17.084 { 00:32:17.084 "name": "NVMe0", 00:32:17.084 "trtype": "tcp", 00:32:17.084 "traddr": "10.0.0.2", 00:32:17.084 "adrfam": "ipv4", 00:32:17.084 "trsvcid": "4420", 00:32:17.084 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:32:17.084 "hostaddr": "10.0.0.1", 00:32:17.084 "prchk_reftag": false, 00:32:17.084 "prchk_guard": false, 00:32:17.084 "hdgst": false, 00:32:17.084 "ddgst": false, 00:32:17.084 "allow_unrecognized_csi": false, 00:32:17.084 "method": "bdev_nvme_attach_controller", 00:32:17.084 "req_id": 1 00:32:17.084 } 00:32:17.084 Got JSON-RPC error response 00:32:17.084 response: 00:32:17.084 { 00:32:17.084 "code": -114, 00:32:17.084 "message": "A controller named NVMe0 already exists with the specified network path" 00:32:17.084 } 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@655 -- # es=1 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@69 -- # NOT rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -x disable 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@652 -- # local es=0 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -x disable 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@655 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -x disable 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:17.084 request: 00:32:17.084 { 00:32:17.084 "name": "NVMe0", 00:32:17.084 "trtype": "tcp", 00:32:17.084 "traddr": "10.0.0.2", 00:32:17.084 "adrfam": "ipv4", 00:32:17.084 "trsvcid": "4420", 00:32:17.084 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:32:17.084 "hostaddr": "10.0.0.1", 00:32:17.084 "prchk_reftag": false, 00:32:17.084 "prchk_guard": false, 00:32:17.084 "hdgst": false, 00:32:17.084 "ddgst": false, 00:32:17.084 "multipath": "disable", 00:32:17.084 "allow_unrecognized_csi": false, 00:32:17.084 "method": "bdev_nvme_attach_controller", 00:32:17.084 "req_id": 1 00:32:17.084 } 00:32:17.084 Got JSON-RPC error response 00:32:17.084 response: 00:32:17.084 { 00:32:17.084 "code": -114, 00:32:17.084 "message": "A controller named NVMe0 already exists and multipath is disabled" 00:32:17.084 } 00:32:17.084 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:32:17.085 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@655 -- # es=1 00:32:17.085 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:32:17.085 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:32:17.085 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:32:17.085 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@74 -- # NOT rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -x failover 00:32:17.085 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@652 -- # local es=0 00:32:17.085 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -x failover 00:32:17.085 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:32:17.342 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:32:17.342 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:32:17.342 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:32:17.342 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@655 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -x failover 00:32:17.342 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:17.342 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:17.342 request: 00:32:17.342 { 00:32:17.342 "name": "NVMe0", 00:32:17.342 "trtype": "tcp", 00:32:17.342 "traddr": "10.0.0.2", 00:32:17.342 "adrfam": "ipv4", 00:32:17.342 "trsvcid": "4420", 00:32:17.342 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:32:17.342 "hostaddr": "10.0.0.1", 00:32:17.342 "prchk_reftag": false, 00:32:17.342 "prchk_guard": false, 00:32:17.342 "hdgst": false, 00:32:17.342 "ddgst": false, 00:32:17.342 "multipath": "failover", 00:32:17.342 "allow_unrecognized_csi": false, 00:32:17.342 "method": "bdev_nvme_attach_controller", 00:32:17.342 "req_id": 1 00:32:17.342 } 00:32:17.342 Got JSON-RPC error response 00:32:17.342 response: 00:32:17.342 { 00:32:17.342 "code": -114, 00:32:17.342 "message": "A controller named NVMe0 already exists with the specified network path" 00:32:17.342 } 00:32:17.342 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:32:17.342 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@655 -- # es=1 00:32:17.342 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:32:17.343 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:32:17.343 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:32:17.343 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@79 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:32:17.343 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:17.343 17:47:58 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:17.343 NVMe0n1 00:32:17.343 17:47:59 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:17.343 17:47:59 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@83 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:32:17.343 17:47:59 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:17.343 17:47:59 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:17.343 17:47:59 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:17.343 17:47:59 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@87 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe1 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 00:32:17.343 17:47:59 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:17.343 17:47:59 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:17.343 00:32:17.343 17:47:59 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:17.343 17:47:59 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@90 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:32:17.343 17:47:59 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@90 -- # grep -c NVMe 00:32:17.343 17:47:59 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:17.343 17:47:59 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:17.343 17:47:59 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:17.343 17:47:59 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@90 -- # '[' 2 '!=' 2 ']' 00:32:17.343 17:47:59 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@95 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:32:18.715 { 00:32:18.715 "results": [ 00:32:18.715 { 00:32:18.715 "job": "NVMe0n1", 00:32:18.715 "core_mask": "0x1", 00:32:18.715 "workload": "write", 00:32:18.715 "status": "finished", 00:32:18.715 "queue_depth": 128, 00:32:18.715 "io_size": 4096, 00:32:18.715 "runtime": 1.008873, 00:32:18.715 "iops": 18445.334546568298, 00:32:18.715 "mibps": 72.05208807253241, 00:32:18.715 "io_failed": 0, 00:32:18.715 "io_timeout": 0, 00:32:18.715 "avg_latency_us": 6923.873123916544, 00:32:18.715 "min_latency_us": 2051.034074074074, 00:32:18.715 "max_latency_us": 13398.471111111112 00:32:18.715 } 00:32:18.715 ], 00:32:18.715 "core_count": 1 00:32:18.715 } 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@98 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe1 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@100 -- # [[ -n '' ]] 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@116 -- # killprocess 343661 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@954 -- # '[' -z 343661 ']' 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@958 -- # kill -0 343661 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@959 -- # uname 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 343661 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@972 -- # echo 'killing process with pid 343661' 00:32:18.715 killing process with pid 343661 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@973 -- # kill 343661 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@978 -- # wait 343661 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@118 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@119 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode2 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@121 -- # trap - SIGINT SIGTERM EXIT 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@123 -- # pap /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1599 -- # read -r file 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1598 -- # find /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt -type f 00:32:18.715 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1598 -- # sort -u 00:32:18.716 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1600 -- # cat 00:32:18.716 --- /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt --- 00:32:18.716 [2024-12-06 17:47:58.535017] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:32:18.716 [2024-12-06 17:47:58.535116] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid343661 ] 00:32:18.716 [2024-12-06 17:47:58.603715] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:18.716 [2024-12-06 17:47:58.650747] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:32:18.716 [2024-12-06 17:47:59.079687] bdev.c:4934:bdev_name_add: *ERROR*: Bdev name f0eb616c-59a3-476d-b6ec-cf8d111f69ba already exists 00:32:18.716 [2024-12-06 17:47:59.079729] bdev.c:8154:bdev_register: *ERROR*: Unable to add uuid:f0eb616c-59a3-476d-b6ec-cf8d111f69ba alias for bdev NVMe1n1 00:32:18.716 [2024-12-06 17:47:59.079745] bdev_nvme.c:4665:nvme_bdev_create: *ERROR*: spdk_bdev_register() failed 00:32:18.716 Running I/O for 1 seconds... 00:32:18.716 18388.00 IOPS, 71.83 MiB/s 00:32:18.716 Latency(us) 00:32:18.716 [2024-12-06T16:48:00.555Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:18.716 Job: NVMe0n1 (Core Mask 0x1, workload: write, depth: 128, IO size: 4096) 00:32:18.716 NVMe0n1 : 1.01 18445.33 72.05 0.00 0.00 6923.87 2051.03 13398.47 00:32:18.716 [2024-12-06T16:48:00.555Z] =================================================================================================================== 00:32:18.716 [2024-12-06T16:48:00.555Z] Total : 18445.33 72.05 0.00 0.00 6923.87 2051.03 13398.47 00:32:18.716 Received shutdown signal, test time was about 1.000000 seconds 00:32:18.716 00:32:18.716 Latency(us) 00:32:18.716 [2024-12-06T16:48:00.555Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:18.716 [2024-12-06T16:48:00.555Z] =================================================================================================================== 00:32:18.716 [2024-12-06T16:48:00.555Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:32:18.716 --- /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt --- 00:32:18.716 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1605 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:32:18.716 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1599 -- # read -r file 00:32:18.716 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@124 -- # nvmftestfini 00:32:18.716 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@516 -- # nvmfcleanup 00:32:18.716 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@121 -- # sync 00:32:18.716 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:32:18.716 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@124 -- # set +e 00:32:18.716 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@125 -- # for i in {1..20} 00:32:18.716 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:32:18.716 rmmod nvme_tcp 00:32:18.716 rmmod nvme_fabrics 00:32:18.716 rmmod nvme_keyring 00:32:18.716 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:32:18.973 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@128 -- # set -e 00:32:18.973 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@129 -- # return 0 00:32:18.973 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@517 -- # '[' -n 343638 ']' 00:32:18.973 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@518 -- # killprocess 343638 00:32:18.973 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@954 -- # '[' -z 343638 ']' 00:32:18.973 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@958 -- # kill -0 343638 00:32:18.973 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@959 -- # uname 00:32:18.973 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:32:18.973 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 343638 00:32:18.973 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:32:18.973 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:32:18.974 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@972 -- # echo 'killing process with pid 343638' 00:32:18.974 killing process with pid 343638 00:32:18.974 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@973 -- # kill 343638 00:32:18.974 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@978 -- # wait 343638 00:32:19.233 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:32:19.233 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:32:19.233 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:32:19.233 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@297 -- # iptr 00:32:19.233 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@791 -- # iptables-save 00:32:19.233 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@791 -- # iptables-restore 00:32:19.233 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:32:19.233 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:32:19.233 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@302 -- # remove_spdk_ns 00:32:19.233 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:19.233 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:19.233 17:48:00 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:21.136 17:48:02 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:32:21.136 00:32:21.136 real 0m7.272s 00:32:21.136 user 0m10.626s 00:32:21.136 sys 0m2.439s 00:32:21.136 17:48:02 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1130 -- # xtrace_disable 00:32:21.136 17:48:02 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:32:21.137 ************************************ 00:32:21.137 END TEST nvmf_multicontroller 00:32:21.137 ************************************ 00:32:21.137 17:48:02 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@17 -- # run_test nvmf_aer /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/aer.sh --transport=tcp 00:32:21.137 17:48:02 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:32:21.137 17:48:02 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:32:21.137 17:48:02 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:32:21.137 ************************************ 00:32:21.137 START TEST nvmf_aer 00:32:21.137 ************************************ 00:32:21.137 17:48:02 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/aer.sh --transport=tcp 00:32:21.395 * Looking for test storage... 00:32:21.395 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:32:21.395 17:48:02 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:32:21.395 17:48:02 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1711 -- # lcov --version 00:32:21.395 17:48:02 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@333 -- # local ver1 ver1_l 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@334 -- # local ver2 ver2_l 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@336 -- # IFS=.-: 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@336 -- # read -ra ver1 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@337 -- # IFS=.-: 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@337 -- # read -ra ver2 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@338 -- # local 'op=<' 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@340 -- # ver1_l=2 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@341 -- # ver2_l=1 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@344 -- # case "$op" in 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@345 -- # : 1 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@364 -- # (( v = 0 )) 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@365 -- # decimal 1 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@353 -- # local d=1 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@355 -- # echo 1 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@365 -- # ver1[v]=1 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@366 -- # decimal 2 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@353 -- # local d=2 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@355 -- # echo 2 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@366 -- # ver2[v]=2 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@368 -- # return 0 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:32:21.395 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:21.395 --rc genhtml_branch_coverage=1 00:32:21.395 --rc genhtml_function_coverage=1 00:32:21.395 --rc genhtml_legend=1 00:32:21.395 --rc geninfo_all_blocks=1 00:32:21.395 --rc geninfo_unexecuted_blocks=1 00:32:21.395 00:32:21.395 ' 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:32:21.395 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:21.395 --rc genhtml_branch_coverage=1 00:32:21.395 --rc genhtml_function_coverage=1 00:32:21.395 --rc genhtml_legend=1 00:32:21.395 --rc geninfo_all_blocks=1 00:32:21.395 --rc geninfo_unexecuted_blocks=1 00:32:21.395 00:32:21.395 ' 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:32:21.395 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:21.395 --rc genhtml_branch_coverage=1 00:32:21.395 --rc genhtml_function_coverage=1 00:32:21.395 --rc genhtml_legend=1 00:32:21.395 --rc geninfo_all_blocks=1 00:32:21.395 --rc geninfo_unexecuted_blocks=1 00:32:21.395 00:32:21.395 ' 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:32:21.395 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:21.395 --rc genhtml_branch_coverage=1 00:32:21.395 --rc genhtml_function_coverage=1 00:32:21.395 --rc genhtml_legend=1 00:32:21.395 --rc geninfo_all_blocks=1 00:32:21.395 --rc geninfo_unexecuted_blocks=1 00:32:21.395 00:32:21.395 ' 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@7 -- # uname -s 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:32:21.395 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@15 -- # shopt -s extglob 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- paths/export.sh@5 -- # export PATH 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@51 -- # : 0 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:32:21.396 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@55 -- # have_pci_nics=0 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@11 -- # nvmftestinit 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@476 -- # prepare_net_devs 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@438 -- # local -g is_hw=no 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@440 -- # remove_spdk_ns 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@309 -- # xtrace_disable 00:32:21.396 17:48:03 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@315 -- # pci_devs=() 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@315 -- # local -a pci_devs 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@316 -- # pci_net_devs=() 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@317 -- # pci_drivers=() 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@317 -- # local -A pci_drivers 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@319 -- # net_devs=() 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@319 -- # local -ga net_devs 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@320 -- # e810=() 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@320 -- # local -ga e810 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@321 -- # x722=() 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@321 -- # local -ga x722 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@322 -- # mlx=() 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@322 -- # local -ga mlx 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:32:23.926 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:32:23.926 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@418 -- # [[ up == up ]] 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:32:23.926 Found net devices under 0000:0a:00.0: cvl_0_0 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@418 -- # [[ up == up ]] 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:32:23.926 Found net devices under 0000:0a:00.1: cvl_0_1 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@442 -- # is_hw=yes 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:32:23.926 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:32:23.926 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.163 ms 00:32:23.926 00:32:23.926 --- 10.0.0.2 ping statistics --- 00:32:23.926 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:32:23.926 rtt min/avg/max/mdev = 0.163/0.163/0.163/0.000 ms 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:32:23.926 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:32:23.926 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.073 ms 00:32:23.926 00:32:23.926 --- 10.0.0.1 ping statistics --- 00:32:23.926 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:32:23.926 rtt min/avg/max/mdev = 0.073/0.073/0.073/0.000 ms 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@450 -- # return 0 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:32:23.926 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:32:23.927 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@12 -- # nvmfappstart -m 0xF 00:32:23.927 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:32:23.927 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@726 -- # xtrace_disable 00:32:23.927 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:32:23.927 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@509 -- # nvmfpid=345986 00:32:23.927 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:32:23.927 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@510 -- # waitforlisten 345986 00:32:23.927 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@835 -- # '[' -z 345986 ']' 00:32:23.927 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:32:23.927 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@840 -- # local max_retries=100 00:32:23.927 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:32:23.927 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:32:23.927 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@844 -- # xtrace_disable 00:32:23.927 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:32:23.927 [2024-12-06 17:48:05.498937] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:32:23.927 [2024-12-06 17:48:05.499037] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:32:23.927 [2024-12-06 17:48:05.572384] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:32:23.927 [2024-12-06 17:48:05.617625] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:32:23.927 [2024-12-06 17:48:05.617707] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:32:23.927 [2024-12-06 17:48:05.617732] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:32:23.927 [2024-12-06 17:48:05.617744] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:32:23.927 [2024-12-06 17:48:05.617754] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:32:23.927 [2024-12-06 17:48:05.619262] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:32:23.927 [2024-12-06 17:48:05.619372] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:32:23.927 [2024-12-06 17:48:05.619449] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:32:23.927 [2024-12-06 17:48:05.619451] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:32:23.927 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:32:23.927 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@868 -- # return 0 00:32:23.927 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:32:23.927 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@732 -- # xtrace_disable 00:32:23.927 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:32:23.927 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:32:23.927 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@14 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:32:23.927 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:23.927 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:32:23.927 [2024-12-06 17:48:05.756948] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:32:24.184 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:24.184 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@16 -- # rpc_cmd bdev_malloc_create 64 512 --name Malloc0 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:32:24.185 Malloc0 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@17 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 2 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@18 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@19 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:32:24.185 [2024-12-06 17:48:05.821203] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@21 -- # rpc_cmd nvmf_get_subsystems 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:32:24.185 [ 00:32:24.185 { 00:32:24.185 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:32:24.185 "subtype": "Discovery", 00:32:24.185 "listen_addresses": [], 00:32:24.185 "allow_any_host": true, 00:32:24.185 "hosts": [] 00:32:24.185 }, 00:32:24.185 { 00:32:24.185 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:32:24.185 "subtype": "NVMe", 00:32:24.185 "listen_addresses": [ 00:32:24.185 { 00:32:24.185 "trtype": "TCP", 00:32:24.185 "adrfam": "IPv4", 00:32:24.185 "traddr": "10.0.0.2", 00:32:24.185 "trsvcid": "4420" 00:32:24.185 } 00:32:24.185 ], 00:32:24.185 "allow_any_host": true, 00:32:24.185 "hosts": [], 00:32:24.185 "serial_number": "SPDK00000000000001", 00:32:24.185 "model_number": "SPDK bdev Controller", 00:32:24.185 "max_namespaces": 2, 00:32:24.185 "min_cntlid": 1, 00:32:24.185 "max_cntlid": 65519, 00:32:24.185 "namespaces": [ 00:32:24.185 { 00:32:24.185 "nsid": 1, 00:32:24.185 "bdev_name": "Malloc0", 00:32:24.185 "name": "Malloc0", 00:32:24.185 "nguid": "F4BF8386F8364478A7BEDDC7C5957F84", 00:32:24.185 "uuid": "f4bf8386-f836-4478-a7be-ddc7c5957f84" 00:32:24.185 } 00:32:24.185 ] 00:32:24.185 } 00:32:24.185 ] 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@23 -- # AER_TOUCH_FILE=/tmp/aer_touch_file 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@24 -- # rm -f /tmp/aer_touch_file 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@33 -- # aerpid=346128 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/aer/aer -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' -n 2 -t /tmp/aer_touch_file 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@36 -- # waitforfile /tmp/aer_touch_file 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1269 -- # local i=0 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1270 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1271 -- # '[' 0 -lt 200 ']' 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1272 -- # i=1 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1273 -- # sleep 0.1 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1270 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1271 -- # '[' 1 -lt 200 ']' 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1272 -- # i=2 00:32:24.185 17:48:05 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1273 -- # sleep 0.1 00:32:24.443 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1270 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:32:24.443 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1271 -- # '[' 2 -lt 200 ']' 00:32:24.443 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1272 -- # i=3 00:32:24.443 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1273 -- # sleep 0.1 00:32:24.443 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1270 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:32:24.443 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1276 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:32:24.443 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1280 -- # return 0 00:32:24.443 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@39 -- # rpc_cmd bdev_malloc_create 64 4096 --name Malloc1 00:32:24.443 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:24.443 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:32:24.443 Malloc1 00:32:24.443 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:24.443 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@40 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 2 00:32:24.443 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:24.443 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:32:24.443 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:24.443 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@41 -- # rpc_cmd nvmf_get_subsystems 00:32:24.443 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:24.443 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:32:24.443 Asynchronous Event Request test 00:32:24.443 Attaching to 10.0.0.2 00:32:24.443 Attached to 10.0.0.2 00:32:24.443 Registering asynchronous event callbacks... 00:32:24.443 Starting namespace attribute notice tests for all controllers... 00:32:24.443 10.0.0.2: aer_cb for log page 4, aen_event_type: 0x02, aen_event_info: 0x00 00:32:24.443 aer_cb - Changed Namespace 00:32:24.443 Cleaning up... 00:32:24.443 [ 00:32:24.443 { 00:32:24.443 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:32:24.443 "subtype": "Discovery", 00:32:24.443 "listen_addresses": [], 00:32:24.443 "allow_any_host": true, 00:32:24.443 "hosts": [] 00:32:24.443 }, 00:32:24.443 { 00:32:24.443 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:32:24.443 "subtype": "NVMe", 00:32:24.443 "listen_addresses": [ 00:32:24.443 { 00:32:24.443 "trtype": "TCP", 00:32:24.443 "adrfam": "IPv4", 00:32:24.443 "traddr": "10.0.0.2", 00:32:24.443 "trsvcid": "4420" 00:32:24.443 } 00:32:24.443 ], 00:32:24.443 "allow_any_host": true, 00:32:24.443 "hosts": [], 00:32:24.443 "serial_number": "SPDK00000000000001", 00:32:24.443 "model_number": "SPDK bdev Controller", 00:32:24.443 "max_namespaces": 2, 00:32:24.443 "min_cntlid": 1, 00:32:24.443 "max_cntlid": 65519, 00:32:24.443 "namespaces": [ 00:32:24.443 { 00:32:24.443 "nsid": 1, 00:32:24.443 "bdev_name": "Malloc0", 00:32:24.443 "name": "Malloc0", 00:32:24.443 "nguid": "F4BF8386F8364478A7BEDDC7C5957F84", 00:32:24.443 "uuid": "f4bf8386-f836-4478-a7be-ddc7c5957f84" 00:32:24.443 }, 00:32:24.443 { 00:32:24.443 "nsid": 2, 00:32:24.443 "bdev_name": "Malloc1", 00:32:24.443 "name": "Malloc1", 00:32:24.443 "nguid": "ED1085153D2E446BB66202493B1F5F5D", 00:32:24.443 "uuid": "ed108515-3d2e-446b-b662-02493b1f5f5d" 00:32:24.443 } 00:32:24.443 ] 00:32:24.443 } 00:32:24.443 ] 00:32:24.443 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:24.443 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@43 -- # wait 346128 00:32:24.443 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@45 -- # rpc_cmd bdev_malloc_delete Malloc0 00:32:24.443 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:24.443 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:32:24.443 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:24.444 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@46 -- # rpc_cmd bdev_malloc_delete Malloc1 00:32:24.444 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:24.444 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:32:24.444 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:24.444 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@47 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:32:24.444 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:24.444 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:32:24.701 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:24.701 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@49 -- # trap - SIGINT SIGTERM EXIT 00:32:24.701 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@51 -- # nvmftestfini 00:32:24.701 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@516 -- # nvmfcleanup 00:32:24.701 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@121 -- # sync 00:32:24.701 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:32:24.701 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@124 -- # set +e 00:32:24.701 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@125 -- # for i in {1..20} 00:32:24.701 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:32:24.701 rmmod nvme_tcp 00:32:24.701 rmmod nvme_fabrics 00:32:24.701 rmmod nvme_keyring 00:32:24.701 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:32:24.701 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@128 -- # set -e 00:32:24.701 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@129 -- # return 0 00:32:24.701 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@517 -- # '[' -n 345986 ']' 00:32:24.701 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@518 -- # killprocess 345986 00:32:24.701 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@954 -- # '[' -z 345986 ']' 00:32:24.701 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@958 -- # kill -0 345986 00:32:24.701 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@959 -- # uname 00:32:24.701 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:32:24.701 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 345986 00:32:24.701 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:32:24.701 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:32:24.701 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@972 -- # echo 'killing process with pid 345986' 00:32:24.701 killing process with pid 345986 00:32:24.701 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@973 -- # kill 345986 00:32:24.701 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@978 -- # wait 345986 00:32:24.960 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:32:24.960 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:32:24.960 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:32:24.960 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@297 -- # iptr 00:32:24.960 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@791 -- # iptables-save 00:32:24.960 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:32:24.960 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@791 -- # iptables-restore 00:32:24.960 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:32:24.960 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@302 -- # remove_spdk_ns 00:32:24.960 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:24.960 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:24.960 17:48:06 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:26.862 17:48:08 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:32:26.862 00:32:26.862 real 0m5.670s 00:32:26.862 user 0m4.586s 00:32:26.862 sys 0m2.102s 00:32:26.862 17:48:08 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1130 -- # xtrace_disable 00:32:26.862 17:48:08 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:32:26.862 ************************************ 00:32:26.862 END TEST nvmf_aer 00:32:26.862 ************************************ 00:32:26.862 17:48:08 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@18 -- # run_test nvmf_async_init /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/async_init.sh --transport=tcp 00:32:26.862 17:48:08 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:32:26.862 17:48:08 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:32:26.862 17:48:08 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:32:26.862 ************************************ 00:32:26.862 START TEST nvmf_async_init 00:32:26.862 ************************************ 00:32:26.862 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/async_init.sh --transport=tcp 00:32:26.862 * Looking for test storage... 00:32:26.862 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:32:26.862 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:32:26.862 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1711 -- # lcov --version 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@333 -- # local ver1 ver1_l 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@334 -- # local ver2 ver2_l 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@336 -- # IFS=.-: 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@336 -- # read -ra ver1 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@337 -- # IFS=.-: 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@337 -- # read -ra ver2 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@338 -- # local 'op=<' 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@340 -- # ver1_l=2 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@341 -- # ver2_l=1 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@344 -- # case "$op" in 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@345 -- # : 1 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@364 -- # (( v = 0 )) 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@365 -- # decimal 1 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@353 -- # local d=1 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@355 -- # echo 1 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@365 -- # ver1[v]=1 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@366 -- # decimal 2 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@353 -- # local d=2 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@355 -- # echo 2 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@366 -- # ver2[v]=2 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@368 -- # return 0 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:32:27.122 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:27.122 --rc genhtml_branch_coverage=1 00:32:27.122 --rc genhtml_function_coverage=1 00:32:27.122 --rc genhtml_legend=1 00:32:27.122 --rc geninfo_all_blocks=1 00:32:27.122 --rc geninfo_unexecuted_blocks=1 00:32:27.122 00:32:27.122 ' 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:32:27.122 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:27.122 --rc genhtml_branch_coverage=1 00:32:27.122 --rc genhtml_function_coverage=1 00:32:27.122 --rc genhtml_legend=1 00:32:27.122 --rc geninfo_all_blocks=1 00:32:27.122 --rc geninfo_unexecuted_blocks=1 00:32:27.122 00:32:27.122 ' 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:32:27.122 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:27.122 --rc genhtml_branch_coverage=1 00:32:27.122 --rc genhtml_function_coverage=1 00:32:27.122 --rc genhtml_legend=1 00:32:27.122 --rc geninfo_all_blocks=1 00:32:27.122 --rc geninfo_unexecuted_blocks=1 00:32:27.122 00:32:27.122 ' 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:32:27.122 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:27.122 --rc genhtml_branch_coverage=1 00:32:27.122 --rc genhtml_function_coverage=1 00:32:27.122 --rc genhtml_legend=1 00:32:27.122 --rc geninfo_all_blocks=1 00:32:27.122 --rc geninfo_unexecuted_blocks=1 00:32:27.122 00:32:27.122 ' 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@11 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@7 -- # uname -s 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@15 -- # shopt -s extglob 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- paths/export.sh@5 -- # export PATH 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@51 -- # : 0 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:32:27.122 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:32:27.122 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:32:27.123 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:32:27.123 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@55 -- # have_pci_nics=0 00:32:27.123 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@13 -- # null_bdev_size=1024 00:32:27.123 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@14 -- # null_block_size=512 00:32:27.123 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@15 -- # null_bdev=null0 00:32:27.123 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@16 -- # nvme_bdev=nvme0 00:32:27.123 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@20 -- # uuidgen 00:32:27.123 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@20 -- # tr -d - 00:32:27.123 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@20 -- # nguid=3495752e56284a59bf687e002544e939 00:32:27.123 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@22 -- # nvmftestinit 00:32:27.123 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:32:27.123 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:32:27.123 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@476 -- # prepare_net_devs 00:32:27.123 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@438 -- # local -g is_hw=no 00:32:27.123 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@440 -- # remove_spdk_ns 00:32:27.123 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:27.123 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:27.123 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:27.123 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:32:27.123 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:32:27.123 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@309 -- # xtrace_disable 00:32:27.123 17:48:08 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:32:29.025 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:32:29.025 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@315 -- # pci_devs=() 00:32:29.025 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@315 -- # local -a pci_devs 00:32:29.025 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@316 -- # pci_net_devs=() 00:32:29.025 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:32:29.025 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@317 -- # pci_drivers=() 00:32:29.025 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@317 -- # local -A pci_drivers 00:32:29.025 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@319 -- # net_devs=() 00:32:29.025 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@319 -- # local -ga net_devs 00:32:29.025 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@320 -- # e810=() 00:32:29.025 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@320 -- # local -ga e810 00:32:29.025 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@321 -- # x722=() 00:32:29.025 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@321 -- # local -ga x722 00:32:29.025 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@322 -- # mlx=() 00:32:29.025 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@322 -- # local -ga mlx 00:32:29.025 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:32:29.025 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:32:29.025 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:32:29.025 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:32:29.025 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:32:29.025 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:32:29.025 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:32:29.025 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:32:29.026 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:32:29.026 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@418 -- # [[ up == up ]] 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:32:29.026 Found net devices under 0000:0a:00.0: cvl_0_0 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@418 -- # [[ up == up ]] 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:32:29.026 Found net devices under 0000:0a:00.1: cvl_0_1 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@442 -- # is_hw=yes 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:32:29.026 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:32:29.285 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:32:29.285 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:32:29.285 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:32:29.285 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:32:29.285 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:32:29.285 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:32:29.285 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:32:29.285 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:32:29.285 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:32:29.285 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:32:29.285 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:32:29.285 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:32:29.285 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.138 ms 00:32:29.285 00:32:29.285 --- 10.0.0.2 ping statistics --- 00:32:29.285 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:32:29.285 rtt min/avg/max/mdev = 0.138/0.138/0.138/0.000 ms 00:32:29.285 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:32:29.285 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:32:29.285 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.052 ms 00:32:29.285 00:32:29.285 --- 10.0.0.1 ping statistics --- 00:32:29.285 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:32:29.285 rtt min/avg/max/mdev = 0.052/0.052/0.052/0.000 ms 00:32:29.285 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:32:29.285 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@450 -- # return 0 00:32:29.285 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:32:29.285 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:32:29.285 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:32:29.285 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:32:29.285 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:32:29.285 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:32:29.285 17:48:10 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:32:29.285 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@23 -- # nvmfappstart -m 0x1 00:32:29.285 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:32:29.285 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@726 -- # xtrace_disable 00:32:29.286 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:32:29.286 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@509 -- # nvmfpid=348586 00:32:29.286 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1 00:32:29.286 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@510 -- # waitforlisten 348586 00:32:29.286 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@835 -- # '[' -z 348586 ']' 00:32:29.286 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:32:29.286 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@840 -- # local max_retries=100 00:32:29.286 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:32:29.286 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:32:29.286 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@844 -- # xtrace_disable 00:32:29.286 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:32:29.286 [2024-12-06 17:48:11.077827] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:32:29.286 [2024-12-06 17:48:11.077918] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:32:29.543 [2024-12-06 17:48:11.150466] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:29.543 [2024-12-06 17:48:11.191807] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:32:29.543 [2024-12-06 17:48:11.191867] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:32:29.543 [2024-12-06 17:48:11.191889] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:32:29.543 [2024-12-06 17:48:11.191900] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:32:29.543 [2024-12-06 17:48:11.191910] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:32:29.543 [2024-12-06 17:48:11.192489] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@868 -- # return 0 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@732 -- # xtrace_disable 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@26 -- # rpc_cmd nvmf_create_transport -t tcp -o 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:32:29.543 [2024-12-06 17:48:11.333416] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@27 -- # rpc_cmd bdev_null_create null0 1024 512 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:32:29.543 null0 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@28 -- # rpc_cmd bdev_wait_for_examine 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@29 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@30 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 null0 -g 3495752e56284a59bf687e002544e939 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@31 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:32:29.543 [2024-12-06 17:48:11.373703] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@37 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 10.0.0.2 -f ipv4 -s 4420 -n nqn.2016-06.io.spdk:cnode0 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:29.543 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:32:29.800 nvme0n1 00:32:29.800 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:29.800 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@41 -- # rpc_cmd bdev_get_bdevs -b nvme0n1 00:32:29.800 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:29.800 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:32:29.800 [ 00:32:29.800 { 00:32:29.800 "name": "nvme0n1", 00:32:29.800 "aliases": [ 00:32:29.800 "3495752e-5628-4a59-bf68-7e002544e939" 00:32:29.800 ], 00:32:29.800 "product_name": "NVMe disk", 00:32:29.800 "block_size": 512, 00:32:29.800 "num_blocks": 2097152, 00:32:29.800 "uuid": "3495752e-5628-4a59-bf68-7e002544e939", 00:32:29.800 "numa_id": 0, 00:32:29.800 "assigned_rate_limits": { 00:32:29.800 "rw_ios_per_sec": 0, 00:32:29.800 "rw_mbytes_per_sec": 0, 00:32:29.800 "r_mbytes_per_sec": 0, 00:32:29.800 "w_mbytes_per_sec": 0 00:32:29.800 }, 00:32:29.800 "claimed": false, 00:32:29.800 "zoned": false, 00:32:29.800 "supported_io_types": { 00:32:29.800 "read": true, 00:32:29.800 "write": true, 00:32:29.800 "unmap": false, 00:32:29.800 "flush": true, 00:32:29.800 "reset": true, 00:32:29.800 "nvme_admin": true, 00:32:29.800 "nvme_io": true, 00:32:29.800 "nvme_io_md": false, 00:32:29.800 "write_zeroes": true, 00:32:29.800 "zcopy": false, 00:32:29.800 "get_zone_info": false, 00:32:29.800 "zone_management": false, 00:32:29.800 "zone_append": false, 00:32:29.800 "compare": true, 00:32:29.800 "compare_and_write": true, 00:32:29.800 "abort": true, 00:32:29.800 "seek_hole": false, 00:32:29.800 "seek_data": false, 00:32:29.800 "copy": true, 00:32:29.800 "nvme_iov_md": false 00:32:29.800 }, 00:32:29.800 "memory_domains": [ 00:32:29.800 { 00:32:29.800 "dma_device_id": "system", 00:32:29.800 "dma_device_type": 1 00:32:29.800 } 00:32:29.800 ], 00:32:29.800 "driver_specific": { 00:32:29.800 "nvme": [ 00:32:29.800 { 00:32:29.800 "trid": { 00:32:29.800 "trtype": "TCP", 00:32:29.800 "adrfam": "IPv4", 00:32:29.800 "traddr": "10.0.0.2", 00:32:29.800 "trsvcid": "4420", 00:32:29.800 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:32:29.800 }, 00:32:29.800 "ctrlr_data": { 00:32:29.800 "cntlid": 1, 00:32:29.800 "vendor_id": "0x8086", 00:32:29.800 "model_number": "SPDK bdev Controller", 00:32:29.800 "serial_number": "00000000000000000000", 00:32:29.800 "firmware_revision": "25.01", 00:32:29.800 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:32:29.800 "oacs": { 00:32:29.800 "security": 0, 00:32:29.801 "format": 0, 00:32:29.801 "firmware": 0, 00:32:29.801 "ns_manage": 0 00:32:29.801 }, 00:32:29.801 "multi_ctrlr": true, 00:32:29.801 "ana_reporting": false 00:32:29.801 }, 00:32:29.801 "vs": { 00:32:29.801 "nvme_version": "1.3" 00:32:29.801 }, 00:32:29.801 "ns_data": { 00:32:29.801 "id": 1, 00:32:29.801 "can_share": true 00:32:29.801 } 00:32:29.801 } 00:32:29.801 ], 00:32:29.801 "mp_policy": "active_passive" 00:32:29.801 } 00:32:29.801 } 00:32:29.801 ] 00:32:29.801 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:29.801 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@44 -- # rpc_cmd bdev_nvme_reset_controller nvme0 00:32:29.801 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:29.801 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:32:29.801 [2024-12-06 17:48:11.627298] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:32:29.801 [2024-12-06 17:48:11.627395] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x13ef250 (9): Bad file descriptor 00:32:30.058 [2024-12-06 17:48:11.759788] bdev_nvme.c:2286:bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 2] Resetting controller successful. 00:32:30.058 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:30.058 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@47 -- # rpc_cmd bdev_get_bdevs -b nvme0n1 00:32:30.058 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:30.058 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:32:30.058 [ 00:32:30.058 { 00:32:30.058 "name": "nvme0n1", 00:32:30.058 "aliases": [ 00:32:30.058 "3495752e-5628-4a59-bf68-7e002544e939" 00:32:30.058 ], 00:32:30.058 "product_name": "NVMe disk", 00:32:30.058 "block_size": 512, 00:32:30.058 "num_blocks": 2097152, 00:32:30.058 "uuid": "3495752e-5628-4a59-bf68-7e002544e939", 00:32:30.058 "numa_id": 0, 00:32:30.058 "assigned_rate_limits": { 00:32:30.058 "rw_ios_per_sec": 0, 00:32:30.058 "rw_mbytes_per_sec": 0, 00:32:30.058 "r_mbytes_per_sec": 0, 00:32:30.058 "w_mbytes_per_sec": 0 00:32:30.058 }, 00:32:30.058 "claimed": false, 00:32:30.058 "zoned": false, 00:32:30.058 "supported_io_types": { 00:32:30.058 "read": true, 00:32:30.058 "write": true, 00:32:30.058 "unmap": false, 00:32:30.058 "flush": true, 00:32:30.058 "reset": true, 00:32:30.058 "nvme_admin": true, 00:32:30.058 "nvme_io": true, 00:32:30.058 "nvme_io_md": false, 00:32:30.058 "write_zeroes": true, 00:32:30.058 "zcopy": false, 00:32:30.058 "get_zone_info": false, 00:32:30.058 "zone_management": false, 00:32:30.058 "zone_append": false, 00:32:30.058 "compare": true, 00:32:30.058 "compare_and_write": true, 00:32:30.058 "abort": true, 00:32:30.058 "seek_hole": false, 00:32:30.058 "seek_data": false, 00:32:30.058 "copy": true, 00:32:30.058 "nvme_iov_md": false 00:32:30.058 }, 00:32:30.058 "memory_domains": [ 00:32:30.058 { 00:32:30.058 "dma_device_id": "system", 00:32:30.058 "dma_device_type": 1 00:32:30.058 } 00:32:30.058 ], 00:32:30.058 "driver_specific": { 00:32:30.058 "nvme": [ 00:32:30.058 { 00:32:30.058 "trid": { 00:32:30.058 "trtype": "TCP", 00:32:30.058 "adrfam": "IPv4", 00:32:30.058 "traddr": "10.0.0.2", 00:32:30.058 "trsvcid": "4420", 00:32:30.058 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:32:30.058 }, 00:32:30.058 "ctrlr_data": { 00:32:30.058 "cntlid": 2, 00:32:30.058 "vendor_id": "0x8086", 00:32:30.058 "model_number": "SPDK bdev Controller", 00:32:30.058 "serial_number": "00000000000000000000", 00:32:30.058 "firmware_revision": "25.01", 00:32:30.058 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:32:30.058 "oacs": { 00:32:30.058 "security": 0, 00:32:30.058 "format": 0, 00:32:30.058 "firmware": 0, 00:32:30.058 "ns_manage": 0 00:32:30.058 }, 00:32:30.058 "multi_ctrlr": true, 00:32:30.058 "ana_reporting": false 00:32:30.058 }, 00:32:30.058 "vs": { 00:32:30.058 "nvme_version": "1.3" 00:32:30.058 }, 00:32:30.058 "ns_data": { 00:32:30.058 "id": 1, 00:32:30.058 "can_share": true 00:32:30.058 } 00:32:30.058 } 00:32:30.058 ], 00:32:30.058 "mp_policy": "active_passive" 00:32:30.058 } 00:32:30.058 } 00:32:30.058 ] 00:32:30.058 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:30.058 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@50 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:32:30.058 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:30.058 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:32:30.058 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:30.058 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@53 -- # mktemp 00:32:30.058 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@53 -- # key_path=/tmp/tmp.dcRmJfd3cz 00:32:30.058 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@54 -- # echo -n NVMeTLSkey-1:01:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmZwJEiQ: 00:32:30.058 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@55 -- # chmod 0600 /tmp/tmp.dcRmJfd3cz 00:32:30.058 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@56 -- # rpc_cmd keyring_file_add_key key0 /tmp/tmp.dcRmJfd3cz 00:32:30.058 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:30.058 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:32:30.058 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:30.058 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@57 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode0 --disable 00:32:30.058 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:30.058 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:32:30.058 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:30.058 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@58 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4421 --secure-channel 00:32:30.058 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:30.058 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:32:30.058 [2024-12-06 17:48:11.815905] tcp.c:1049:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:32:30.058 [2024-12-06 17:48:11.816043] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:32:30.059 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:30.059 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@60 -- # rpc_cmd nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode0 nqn.2016-06.io.spdk:host1 --psk key0 00:32:30.059 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:30.059 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:32:30.059 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:30.059 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@66 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 10.0.0.2 -f ipv4 -s 4421 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host1 --psk key0 00:32:30.059 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:30.059 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:32:30.059 [2024-12-06 17:48:11.831943] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:32:30.316 nvme0n1 00:32:30.316 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:30.316 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@70 -- # rpc_cmd bdev_get_bdevs -b nvme0n1 00:32:30.316 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:30.316 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:32:30.316 [ 00:32:30.316 { 00:32:30.316 "name": "nvme0n1", 00:32:30.316 "aliases": [ 00:32:30.316 "3495752e-5628-4a59-bf68-7e002544e939" 00:32:30.316 ], 00:32:30.316 "product_name": "NVMe disk", 00:32:30.316 "block_size": 512, 00:32:30.316 "num_blocks": 2097152, 00:32:30.316 "uuid": "3495752e-5628-4a59-bf68-7e002544e939", 00:32:30.316 "numa_id": 0, 00:32:30.316 "assigned_rate_limits": { 00:32:30.316 "rw_ios_per_sec": 0, 00:32:30.316 "rw_mbytes_per_sec": 0, 00:32:30.316 "r_mbytes_per_sec": 0, 00:32:30.316 "w_mbytes_per_sec": 0 00:32:30.317 }, 00:32:30.317 "claimed": false, 00:32:30.317 "zoned": false, 00:32:30.317 "supported_io_types": { 00:32:30.317 "read": true, 00:32:30.317 "write": true, 00:32:30.317 "unmap": false, 00:32:30.317 "flush": true, 00:32:30.317 "reset": true, 00:32:30.317 "nvme_admin": true, 00:32:30.317 "nvme_io": true, 00:32:30.317 "nvme_io_md": false, 00:32:30.317 "write_zeroes": true, 00:32:30.317 "zcopy": false, 00:32:30.317 "get_zone_info": false, 00:32:30.317 "zone_management": false, 00:32:30.317 "zone_append": false, 00:32:30.317 "compare": true, 00:32:30.317 "compare_and_write": true, 00:32:30.317 "abort": true, 00:32:30.317 "seek_hole": false, 00:32:30.317 "seek_data": false, 00:32:30.317 "copy": true, 00:32:30.317 "nvme_iov_md": false 00:32:30.317 }, 00:32:30.317 "memory_domains": [ 00:32:30.317 { 00:32:30.317 "dma_device_id": "system", 00:32:30.317 "dma_device_type": 1 00:32:30.317 } 00:32:30.317 ], 00:32:30.317 "driver_specific": { 00:32:30.317 "nvme": [ 00:32:30.317 { 00:32:30.317 "trid": { 00:32:30.317 "trtype": "TCP", 00:32:30.317 "adrfam": "IPv4", 00:32:30.317 "traddr": "10.0.0.2", 00:32:30.317 "trsvcid": "4421", 00:32:30.317 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:32:30.317 }, 00:32:30.317 "ctrlr_data": { 00:32:30.317 "cntlid": 3, 00:32:30.317 "vendor_id": "0x8086", 00:32:30.317 "model_number": "SPDK bdev Controller", 00:32:30.317 "serial_number": "00000000000000000000", 00:32:30.317 "firmware_revision": "25.01", 00:32:30.317 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:32:30.317 "oacs": { 00:32:30.317 "security": 0, 00:32:30.317 "format": 0, 00:32:30.317 "firmware": 0, 00:32:30.317 "ns_manage": 0 00:32:30.317 }, 00:32:30.317 "multi_ctrlr": true, 00:32:30.317 "ana_reporting": false 00:32:30.317 }, 00:32:30.317 "vs": { 00:32:30.317 "nvme_version": "1.3" 00:32:30.317 }, 00:32:30.317 "ns_data": { 00:32:30.317 "id": 1, 00:32:30.317 "can_share": true 00:32:30.317 } 00:32:30.317 } 00:32:30.317 ], 00:32:30.317 "mp_policy": "active_passive" 00:32:30.317 } 00:32:30.317 } 00:32:30.317 ] 00:32:30.317 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:30.317 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@73 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:32:30.317 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:30.317 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:32:30.317 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:30.317 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@76 -- # rm -f /tmp/tmp.dcRmJfd3cz 00:32:30.317 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@78 -- # trap - SIGINT SIGTERM EXIT 00:32:30.317 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@79 -- # nvmftestfini 00:32:30.317 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@516 -- # nvmfcleanup 00:32:30.317 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@121 -- # sync 00:32:30.317 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:32:30.317 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@124 -- # set +e 00:32:30.317 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@125 -- # for i in {1..20} 00:32:30.317 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:32:30.317 rmmod nvme_tcp 00:32:30.317 rmmod nvme_fabrics 00:32:30.317 rmmod nvme_keyring 00:32:30.317 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:32:30.317 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@128 -- # set -e 00:32:30.317 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@129 -- # return 0 00:32:30.317 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@517 -- # '[' -n 348586 ']' 00:32:30.317 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@518 -- # killprocess 348586 00:32:30.317 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@954 -- # '[' -z 348586 ']' 00:32:30.317 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@958 -- # kill -0 348586 00:32:30.317 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@959 -- # uname 00:32:30.317 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:32:30.317 17:48:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 348586 00:32:30.317 17:48:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:32:30.317 17:48:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:32:30.317 17:48:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@972 -- # echo 'killing process with pid 348586' 00:32:30.317 killing process with pid 348586 00:32:30.317 17:48:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@973 -- # kill 348586 00:32:30.317 17:48:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@978 -- # wait 348586 00:32:30.576 17:48:12 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:32:30.576 17:48:12 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:32:30.576 17:48:12 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:32:30.576 17:48:12 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@297 -- # iptr 00:32:30.576 17:48:12 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@791 -- # iptables-save 00:32:30.576 17:48:12 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:32:30.576 17:48:12 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@791 -- # iptables-restore 00:32:30.576 17:48:12 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:32:30.576 17:48:12 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@302 -- # remove_spdk_ns 00:32:30.576 17:48:12 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:30.576 17:48:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:30.576 17:48:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:32.480 17:48:14 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:32:32.480 00:32:32.480 real 0m5.573s 00:32:32.480 user 0m2.088s 00:32:32.480 sys 0m1.898s 00:32:32.480 17:48:14 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1130 -- # xtrace_disable 00:32:32.480 17:48:14 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:32:32.480 ************************************ 00:32:32.480 END TEST nvmf_async_init 00:32:32.480 ************************************ 00:32:32.480 17:48:14 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@19 -- # run_test dma /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/dma.sh --transport=tcp 00:32:32.480 17:48:14 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:32:32.480 17:48:14 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:32:32.480 17:48:14 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:32:32.480 ************************************ 00:32:32.480 START TEST dma 00:32:32.480 ************************************ 00:32:32.480 17:48:14 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/dma.sh --transport=tcp 00:32:32.480 * Looking for test storage... 00:32:32.739 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1711 -- # lcov --version 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@333 -- # local ver1 ver1_l 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@334 -- # local ver2 ver2_l 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@336 -- # IFS=.-: 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@336 -- # read -ra ver1 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@337 -- # IFS=.-: 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@337 -- # read -ra ver2 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@338 -- # local 'op=<' 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@340 -- # ver1_l=2 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@341 -- # ver2_l=1 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@344 -- # case "$op" in 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@345 -- # : 1 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@364 -- # (( v = 0 )) 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@365 -- # decimal 1 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@353 -- # local d=1 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@355 -- # echo 1 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@365 -- # ver1[v]=1 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@366 -- # decimal 2 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@353 -- # local d=2 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@355 -- # echo 2 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@366 -- # ver2[v]=2 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@368 -- # return 0 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:32:32.739 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:32.739 --rc genhtml_branch_coverage=1 00:32:32.739 --rc genhtml_function_coverage=1 00:32:32.739 --rc genhtml_legend=1 00:32:32.739 --rc geninfo_all_blocks=1 00:32:32.739 --rc geninfo_unexecuted_blocks=1 00:32:32.739 00:32:32.739 ' 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:32:32.739 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:32.739 --rc genhtml_branch_coverage=1 00:32:32.739 --rc genhtml_function_coverage=1 00:32:32.739 --rc genhtml_legend=1 00:32:32.739 --rc geninfo_all_blocks=1 00:32:32.739 --rc geninfo_unexecuted_blocks=1 00:32:32.739 00:32:32.739 ' 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:32:32.739 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:32.739 --rc genhtml_branch_coverage=1 00:32:32.739 --rc genhtml_function_coverage=1 00:32:32.739 --rc genhtml_legend=1 00:32:32.739 --rc geninfo_all_blocks=1 00:32:32.739 --rc geninfo_unexecuted_blocks=1 00:32:32.739 00:32:32.739 ' 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:32:32.739 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:32.739 --rc genhtml_branch_coverage=1 00:32:32.739 --rc genhtml_function_coverage=1 00:32:32.739 --rc genhtml_legend=1 00:32:32.739 --rc geninfo_all_blocks=1 00:32:32.739 --rc geninfo_unexecuted_blocks=1 00:32:32.739 00:32:32.739 ' 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- host/dma.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@7 -- # uname -s 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@15 -- # shopt -s extglob 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- paths/export.sh@5 -- # export PATH 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@51 -- # : 0 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:32:32.739 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@55 -- # have_pci_nics=0 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- host/dma.sh@12 -- # '[' tcp '!=' rdma ']' 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- host/dma.sh@13 -- # exit 0 00:32:32.739 00:32:32.739 real 0m0.159s 00:32:32.739 user 0m0.104s 00:32:32.739 sys 0m0.063s 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1130 -- # xtrace_disable 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@10 -- # set +x 00:32:32.739 ************************************ 00:32:32.739 END TEST dma 00:32:32.739 ************************************ 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@22 -- # run_test nvmf_identify /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/identify.sh --transport=tcp 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:32:32.739 ************************************ 00:32:32.739 START TEST nvmf_identify 00:32:32.739 ************************************ 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/identify.sh --transport=tcp 00:32:32.739 * Looking for test storage... 00:32:32.739 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1711 -- # lcov --version 00:32:32.739 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@333 -- # local ver1 ver1_l 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@334 -- # local ver2 ver2_l 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@336 -- # IFS=.-: 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@336 -- # read -ra ver1 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@337 -- # IFS=.-: 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@337 -- # read -ra ver2 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@338 -- # local 'op=<' 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@340 -- # ver1_l=2 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@341 -- # ver2_l=1 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@344 -- # case "$op" in 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@345 -- # : 1 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@364 -- # (( v = 0 )) 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@365 -- # decimal 1 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@353 -- # local d=1 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@355 -- # echo 1 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@365 -- # ver1[v]=1 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@366 -- # decimal 2 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@353 -- # local d=2 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@355 -- # echo 2 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@366 -- # ver2[v]=2 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@368 -- # return 0 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:32:32.999 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:32.999 --rc genhtml_branch_coverage=1 00:32:32.999 --rc genhtml_function_coverage=1 00:32:32.999 --rc genhtml_legend=1 00:32:32.999 --rc geninfo_all_blocks=1 00:32:32.999 --rc geninfo_unexecuted_blocks=1 00:32:32.999 00:32:32.999 ' 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:32:32.999 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:32.999 --rc genhtml_branch_coverage=1 00:32:32.999 --rc genhtml_function_coverage=1 00:32:32.999 --rc genhtml_legend=1 00:32:32.999 --rc geninfo_all_blocks=1 00:32:32.999 --rc geninfo_unexecuted_blocks=1 00:32:32.999 00:32:32.999 ' 00:32:32.999 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:32:32.999 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:32.999 --rc genhtml_branch_coverage=1 00:32:32.999 --rc genhtml_function_coverage=1 00:32:32.999 --rc genhtml_legend=1 00:32:32.999 --rc geninfo_all_blocks=1 00:32:32.999 --rc geninfo_unexecuted_blocks=1 00:32:32.999 00:32:33.000 ' 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:32:33.000 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:33.000 --rc genhtml_branch_coverage=1 00:32:33.000 --rc genhtml_function_coverage=1 00:32:33.000 --rc genhtml_legend=1 00:32:33.000 --rc geninfo_all_blocks=1 00:32:33.000 --rc geninfo_unexecuted_blocks=1 00:32:33.000 00:32:33.000 ' 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@7 -- # uname -s 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@15 -- # shopt -s extglob 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- paths/export.sh@5 -- # export PATH 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@51 -- # : 0 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:32:33.000 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@55 -- # have_pci_nics=0 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@11 -- # MALLOC_BDEV_SIZE=64 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@14 -- # nvmftestinit 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@476 -- # prepare_net_devs 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@438 -- # local -g is_hw=no 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@440 -- # remove_spdk_ns 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@309 -- # xtrace_disable 00:32:33.000 17:48:14 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@315 -- # pci_devs=() 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@315 -- # local -a pci_devs 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@316 -- # pci_net_devs=() 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@317 -- # pci_drivers=() 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@317 -- # local -A pci_drivers 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@319 -- # net_devs=() 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@319 -- # local -ga net_devs 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@320 -- # e810=() 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@320 -- # local -ga e810 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@321 -- # x722=() 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@321 -- # local -ga x722 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@322 -- # mlx=() 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@322 -- # local -ga mlx 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:32:34.984 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:32:34.984 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:32:34.984 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@418 -- # [[ up == up ]] 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:32:34.985 Found net devices under 0000:0a:00.0: cvl_0_0 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@418 -- # [[ up == up ]] 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:32:34.985 Found net devices under 0000:0a:00.1: cvl_0_1 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@442 -- # is_hw=yes 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:32:34.985 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:32:34.985 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.234 ms 00:32:34.985 00:32:34.985 --- 10.0.0.2 ping statistics --- 00:32:34.985 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:32:34.985 rtt min/avg/max/mdev = 0.234/0.234/0.234/0.000 ms 00:32:34.985 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:32:34.985 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:32:34.985 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.112 ms 00:32:34.985 00:32:34.985 --- 10.0.0.1 ping statistics --- 00:32:34.985 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:32:34.985 rtt min/avg/max/mdev = 0.112/0.112/0.112/0.000 ms 00:32:35.268 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:32:35.268 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@450 -- # return 0 00:32:35.268 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:32:35.268 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:32:35.268 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:32:35.269 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:32:35.269 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:32:35.269 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:32:35.269 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:32:35.269 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@16 -- # timing_enter start_nvmf_tgt 00:32:35.269 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@726 -- # xtrace_disable 00:32:35.269 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:32:35.269 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@19 -- # nvmfpid=350729 00:32:35.269 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@21 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:32:35.269 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@23 -- # waitforlisten 350729 00:32:35.269 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@18 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:32:35.269 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@835 -- # '[' -z 350729 ']' 00:32:35.269 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:32:35.269 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@840 -- # local max_retries=100 00:32:35.269 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:32:35.269 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:32:35.269 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@844 -- # xtrace_disable 00:32:35.269 17:48:16 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:32:35.269 [2024-12-06 17:48:16.898937] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:32:35.269 [2024-12-06 17:48:16.899040] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:32:35.269 [2024-12-06 17:48:16.975434] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:32:35.269 [2024-12-06 17:48:17.025397] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:32:35.269 [2024-12-06 17:48:17.025473] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:32:35.269 [2024-12-06 17:48:17.025486] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:32:35.269 [2024-12-06 17:48:17.025497] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:32:35.269 [2024-12-06 17:48:17.025505] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:32:35.269 [2024-12-06 17:48:17.028688] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:32:35.269 [2024-12-06 17:48:17.028711] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:32:35.269 [2024-12-06 17:48:17.028771] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:32:35.269 [2024-12-06 17:48:17.028775] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@868 -- # return 0 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@24 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:32:35.544 [2024-12-06 17:48:17.156587] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@25 -- # timing_exit start_nvmf_tgt 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@732 -- # xtrace_disable 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@27 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:32:35.544 Malloc0 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@31 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 --nguid ABCDEF0123456789ABCDEF0123456789 --eui64 ABCDEF0123456789 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@34 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:32:35.544 [2024-12-06 17:48:17.239414] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@35 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@37 -- # rpc_cmd nvmf_get_subsystems 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:32:35.544 [ 00:32:35.544 { 00:32:35.544 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:32:35.544 "subtype": "Discovery", 00:32:35.544 "listen_addresses": [ 00:32:35.544 { 00:32:35.544 "trtype": "TCP", 00:32:35.544 "adrfam": "IPv4", 00:32:35.544 "traddr": "10.0.0.2", 00:32:35.544 "trsvcid": "4420" 00:32:35.544 } 00:32:35.544 ], 00:32:35.544 "allow_any_host": true, 00:32:35.544 "hosts": [] 00:32:35.544 }, 00:32:35.544 { 00:32:35.544 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:32:35.544 "subtype": "NVMe", 00:32:35.544 "listen_addresses": [ 00:32:35.544 { 00:32:35.544 "trtype": "TCP", 00:32:35.544 "adrfam": "IPv4", 00:32:35.544 "traddr": "10.0.0.2", 00:32:35.544 "trsvcid": "4420" 00:32:35.544 } 00:32:35.544 ], 00:32:35.544 "allow_any_host": true, 00:32:35.544 "hosts": [], 00:32:35.544 "serial_number": "SPDK00000000000001", 00:32:35.544 "model_number": "SPDK bdev Controller", 00:32:35.544 "max_namespaces": 32, 00:32:35.544 "min_cntlid": 1, 00:32:35.544 "max_cntlid": 65519, 00:32:35.544 "namespaces": [ 00:32:35.544 { 00:32:35.544 "nsid": 1, 00:32:35.544 "bdev_name": "Malloc0", 00:32:35.544 "name": "Malloc0", 00:32:35.544 "nguid": "ABCDEF0123456789ABCDEF0123456789", 00:32:35.544 "eui64": "ABCDEF0123456789", 00:32:35.544 "uuid": "2399ee42-e095-4bc5-84c3-ba1500018911" 00:32:35.544 } 00:32:35.544 ] 00:32:35.544 } 00:32:35.544 ] 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:35.544 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@39 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2014-08.org.nvmexpress.discovery' -L all 00:32:35.544 [2024-12-06 17:48:17.283603] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:32:35.544 [2024-12-06 17:48:17.283656] [ DPDK EAL parameters: identify --no-shconf -c 0x1 -n 1 -m 0 --no-pci --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid350838 ] 00:32:35.544 [2024-12-06 17:48:17.338672] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 0] setting state to connect adminq (no timeout) 00:32:35.544 [2024-12-06 17:48:17.338743] nvme_tcp.c:2238:nvme_tcp_qpair_connect_sock: *DEBUG*: adrfam 1 ai_family 2 00:32:35.544 [2024-12-06 17:48:17.338754] nvme_tcp.c:2242:nvme_tcp_qpair_connect_sock: *DEBUG*: trsvcid is 4420 00:32:35.544 [2024-12-06 17:48:17.338773] nvme_tcp.c:2263:nvme_tcp_qpair_connect_sock: *DEBUG*: sock_impl_name is (null) 00:32:35.544 [2024-12-06 17:48:17.338787] sock.c: 373:spdk_sock_connect_ext: *DEBUG*: Creating a client socket using impl posix 00:32:35.544 [2024-12-06 17:48:17.343104] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 0] setting state to wait for connect adminq (no timeout) 00:32:35.544 [2024-12-06 17:48:17.343176] nvme_tcp.c:1455:nvme_tcp_send_icreq_complete: *DEBUG*: Complete the icreq send for tqpair=0x2320d80 0 00:32:35.544 [2024-12-06 17:48:17.343384] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 1 00:32:35.544 [2024-12-06 17:48:17.343401] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =1 00:32:35.544 [2024-12-06 17:48:17.343409] nvme_tcp.c:1501:nvme_tcp_icresp_handle: *DEBUG*: host_hdgst_enable: 0 00:32:35.544 [2024-12-06 17:48:17.343415] nvme_tcp.c:1502:nvme_tcp_icresp_handle: *DEBUG*: host_ddgst_enable: 0 00:32:35.544 [2024-12-06 17:48:17.343455] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.544 [2024-12-06 17:48:17.343467] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.544 [2024-12-06 17:48:17.343478] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x2320d80) 00:32:35.544 [2024-12-06 17:48:17.343497] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:0 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x400 00:32:35.544 [2024-12-06 17:48:17.343522] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c480, cid 0, qid 0 00:32:35.544 [2024-12-06 17:48:17.350678] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.544 [2024-12-06 17:48:17.350697] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.544 [2024-12-06 17:48:17.350705] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.544 [2024-12-06 17:48:17.350713] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c480) on tqpair=0x2320d80 00:32:35.544 [2024-12-06 17:48:17.350728] nvme_fabric.c: 621:nvme_fabric_qpair_connect_poll: *DEBUG*: CNTLID 0x0001 00:32:35.544 [2024-12-06 17:48:17.350740] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to read vs (no timeout) 00:32:35.544 [2024-12-06 17:48:17.350749] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to read vs wait for vs (no timeout) 00:32:35.544 [2024-12-06 17:48:17.350771] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.544 [2024-12-06 17:48:17.350780] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.544 [2024-12-06 17:48:17.350787] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x2320d80) 00:32:35.544 [2024-12-06 17:48:17.350798] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.544 [2024-12-06 17:48:17.350822] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c480, cid 0, qid 0 00:32:35.544 [2024-12-06 17:48:17.350951] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.544 [2024-12-06 17:48:17.350967] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.544 [2024-12-06 17:48:17.350974] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.544 [2024-12-06 17:48:17.350981] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c480) on tqpair=0x2320d80 00:32:35.544 [2024-12-06 17:48:17.350990] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to read cap (no timeout) 00:32:35.544 [2024-12-06 17:48:17.351002] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to read cap wait for cap (no timeout) 00:32:35.544 [2024-12-06 17:48:17.351015] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.544 [2024-12-06 17:48:17.351022] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.544 [2024-12-06 17:48:17.351029] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x2320d80) 00:32:35.544 [2024-12-06 17:48:17.351039] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.544 [2024-12-06 17:48:17.351061] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c480, cid 0, qid 0 00:32:35.544 [2024-12-06 17:48:17.351140] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.544 [2024-12-06 17:48:17.351152] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.544 [2024-12-06 17:48:17.351159] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.544 [2024-12-06 17:48:17.351166] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c480) on tqpair=0x2320d80 00:32:35.544 [2024-12-06 17:48:17.351175] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to check en (no timeout) 00:32:35.544 [2024-12-06 17:48:17.351189] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to check en wait for cc (timeout 15000 ms) 00:32:35.545 [2024-12-06 17:48:17.351201] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.545 [2024-12-06 17:48:17.351208] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.545 [2024-12-06 17:48:17.351221] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x2320d80) 00:32:35.545 [2024-12-06 17:48:17.351232] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.545 [2024-12-06 17:48:17.351253] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c480, cid 0, qid 0 00:32:35.545 [2024-12-06 17:48:17.351323] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.545 [2024-12-06 17:48:17.351335] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.545 [2024-12-06 17:48:17.351341] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.545 [2024-12-06 17:48:17.351348] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c480) on tqpair=0x2320d80 00:32:35.545 [2024-12-06 17:48:17.351356] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to disable and wait for CSTS.RDY = 0 (timeout 15000 ms) 00:32:35.545 [2024-12-06 17:48:17.351372] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.545 [2024-12-06 17:48:17.351381] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.545 [2024-12-06 17:48:17.351388] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x2320d80) 00:32:35.545 [2024-12-06 17:48:17.351398] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.545 [2024-12-06 17:48:17.351418] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c480, cid 0, qid 0 00:32:35.545 [2024-12-06 17:48:17.351495] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.545 [2024-12-06 17:48:17.351509] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.545 [2024-12-06 17:48:17.351516] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.545 [2024-12-06 17:48:17.351522] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c480) on tqpair=0x2320d80 00:32:35.545 [2024-12-06 17:48:17.351531] nvme_ctrlr.c:3906:nvme_ctrlr_process_init_wait_for_ready_0: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] CC.EN = 0 && CSTS.RDY = 0 00:32:35.545 [2024-12-06 17:48:17.351539] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to controller is disabled (timeout 15000 ms) 00:32:35.545 [2024-12-06 17:48:17.351552] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to enable controller by writing CC.EN = 1 (timeout 15000 ms) 00:32:35.545 [2024-12-06 17:48:17.351661] nvme_ctrlr.c:4104:nvme_ctrlr_process_init: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] Setting CC.EN = 1 00:32:35.545 [2024-12-06 17:48:17.351679] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to enable controller by writing CC.EN = 1 reg (timeout 15000 ms) 00:32:35.545 [2024-12-06 17:48:17.351694] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.545 [2024-12-06 17:48:17.351701] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.545 [2024-12-06 17:48:17.351707] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x2320d80) 00:32:35.545 [2024-12-06 17:48:17.351718] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY SET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.545 [2024-12-06 17:48:17.351739] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c480, cid 0, qid 0 00:32:35.545 [2024-12-06 17:48:17.351852] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.545 [2024-12-06 17:48:17.351864] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.545 [2024-12-06 17:48:17.351871] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.545 [2024-12-06 17:48:17.351877] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c480) on tqpair=0x2320d80 00:32:35.545 [2024-12-06 17:48:17.351886] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to wait for CSTS.RDY = 1 (timeout 15000 ms) 00:32:35.545 [2024-12-06 17:48:17.351902] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.545 [2024-12-06 17:48:17.351915] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.545 [2024-12-06 17:48:17.351922] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x2320d80) 00:32:35.545 [2024-12-06 17:48:17.351933] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.545 [2024-12-06 17:48:17.351953] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c480, cid 0, qid 0 00:32:35.545 [2024-12-06 17:48:17.352025] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.545 [2024-12-06 17:48:17.352037] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.545 [2024-12-06 17:48:17.352043] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.545 [2024-12-06 17:48:17.352050] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c480) on tqpair=0x2320d80 00:32:35.545 [2024-12-06 17:48:17.352057] nvme_ctrlr.c:3941:nvme_ctrlr_process_init_enable_wait_for_ready_1: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] CC.EN = 1 && CSTS.RDY = 1 - controller is ready 00:32:35.545 [2024-12-06 17:48:17.352066] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to reset admin queue (timeout 30000 ms) 00:32:35.545 [2024-12-06 17:48:17.352079] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to identify controller (no timeout) 00:32:35.545 [2024-12-06 17:48:17.352098] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to wait for identify controller (timeout 30000 ms) 00:32:35.545 [2024-12-06 17:48:17.352113] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.545 [2024-12-06 17:48:17.352120] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x2320d80) 00:32:35.545 [2024-12-06 17:48:17.352131] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:0 nsid:0 cdw10:00000001 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.545 [2024-12-06 17:48:17.352152] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c480, cid 0, qid 0 00:32:35.545 [2024-12-06 17:48:17.352283] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:32:35.545 [2024-12-06 17:48:17.352296] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:32:35.545 [2024-12-06 17:48:17.352302] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:32:35.545 [2024-12-06 17:48:17.352309] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x2320d80): datao=0, datal=4096, cccid=0 00:32:35.545 [2024-12-06 17:48:17.352316] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x238c480) on tqpair(0x2320d80): expected_datao=0, payload_size=4096 00:32:35.545 [2024-12-06 17:48:17.352324] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.545 [2024-12-06 17:48:17.352340] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:32:35.545 [2024-12-06 17:48:17.352350] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:32:35.830 [2024-12-06 17:48:17.392767] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.830 [2024-12-06 17:48:17.392786] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.830 [2024-12-06 17:48:17.392793] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.830 [2024-12-06 17:48:17.392800] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c480) on tqpair=0x2320d80 00:32:35.830 [2024-12-06 17:48:17.392812] nvme_ctrlr.c:2081:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] transport max_xfer_size 4294967295 00:32:35.830 [2024-12-06 17:48:17.392827] nvme_ctrlr.c:2085:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] MDTS max_xfer_size 131072 00:32:35.830 [2024-12-06 17:48:17.392835] nvme_ctrlr.c:2088:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] CNTLID 0x0001 00:32:35.830 [2024-12-06 17:48:17.392844] nvme_ctrlr.c:2112:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] transport max_sges 16 00:32:35.830 [2024-12-06 17:48:17.392855] nvme_ctrlr.c:2127:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] fuses compare and write: 1 00:32:35.830 [2024-12-06 17:48:17.392864] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to configure AER (timeout 30000 ms) 00:32:35.830 [2024-12-06 17:48:17.392880] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to wait for configure aer (timeout 30000 ms) 00:32:35.830 [2024-12-06 17:48:17.392892] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.830 [2024-12-06 17:48:17.392900] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.830 [2024-12-06 17:48:17.392907] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x2320d80) 00:32:35.830 [2024-12-06 17:48:17.392918] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES ASYNC EVENT CONFIGURATION cid:0 cdw10:0000000b SGL DATA BLOCK OFFSET 0x0 len:0x0 00:32:35.831 [2024-12-06 17:48:17.392941] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c480, cid 0, qid 0 00:32:35.831 [2024-12-06 17:48:17.393028] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.831 [2024-12-06 17:48:17.393043] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.831 [2024-12-06 17:48:17.393049] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.393056] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c480) on tqpair=0x2320d80 00:32:35.831 [2024-12-06 17:48:17.393068] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.393075] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.393081] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x2320d80) 00:32:35.831 [2024-12-06 17:48:17.393091] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:32:35.831 [2024-12-06 17:48:17.393101] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.393108] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.393114] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=1 on tqpair(0x2320d80) 00:32:35.831 [2024-12-06 17:48:17.393122] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:32:35.831 [2024-12-06 17:48:17.393132] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.393138] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.393144] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=2 on tqpair(0x2320d80) 00:32:35.831 [2024-12-06 17:48:17.393153] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:32:35.831 [2024-12-06 17:48:17.393162] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.393169] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.393175] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x2320d80) 00:32:35.831 [2024-12-06 17:48:17.393183] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:32:35.831 [2024-12-06 17:48:17.393192] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to set keep alive timeout (timeout 30000 ms) 00:32:35.831 [2024-12-06 17:48:17.393211] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to wait for set keep alive timeout (timeout 30000 ms) 00:32:35.831 [2024-12-06 17:48:17.393224] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.393232] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x2320d80) 00:32:35.831 [2024-12-06 17:48:17.393242] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES KEEP ALIVE TIMER cid:4 cdw10:0000000f SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.831 [2024-12-06 17:48:17.393268] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c480, cid 0, qid 0 00:32:35.831 [2024-12-06 17:48:17.393280] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c600, cid 1, qid 0 00:32:35.831 [2024-12-06 17:48:17.393288] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c780, cid 2, qid 0 00:32:35.831 [2024-12-06 17:48:17.393295] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c900, cid 3, qid 0 00:32:35.831 [2024-12-06 17:48:17.393302] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238ca80, cid 4, qid 0 00:32:35.831 [2024-12-06 17:48:17.393444] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.831 [2024-12-06 17:48:17.393456] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.831 [2024-12-06 17:48:17.393462] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.393469] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238ca80) on tqpair=0x2320d80 00:32:35.831 [2024-12-06 17:48:17.393477] nvme_ctrlr.c:3059:nvme_ctrlr_set_keep_alive_timeout_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] Sending keep alive every 5000000 us 00:32:35.831 [2024-12-06 17:48:17.393486] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to ready (no timeout) 00:32:35.831 [2024-12-06 17:48:17.393503] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.393512] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x2320d80) 00:32:35.831 [2024-12-06 17:48:17.393523] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:4 nsid:0 cdw10:00000001 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.831 [2024-12-06 17:48:17.393543] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238ca80, cid 4, qid 0 00:32:35.831 [2024-12-06 17:48:17.393630] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:32:35.831 [2024-12-06 17:48:17.393642] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:32:35.831 [2024-12-06 17:48:17.393649] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.393655] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x2320d80): datao=0, datal=4096, cccid=4 00:32:35.831 [2024-12-06 17:48:17.393670] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x238ca80) on tqpair(0x2320d80): expected_datao=0, payload_size=4096 00:32:35.831 [2024-12-06 17:48:17.393679] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.393690] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.393697] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.393708] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.831 [2024-12-06 17:48:17.393718] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.831 [2024-12-06 17:48:17.393724] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.393731] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238ca80) on tqpair=0x2320d80 00:32:35.831 [2024-12-06 17:48:17.393749] nvme_ctrlr.c:4202:nvme_ctrlr_process_init: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] Ctrlr already in ready state 00:32:35.831 [2024-12-06 17:48:17.393785] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.393796] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x2320d80) 00:32:35.831 [2024-12-06 17:48:17.393806] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:4 nsid:0 cdw10:00ff0070 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.831 [2024-12-06 17:48:17.393818] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.393825] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.393831] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x2320d80) 00:32:35.831 [2024-12-06 17:48:17.393844] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: KEEP ALIVE (18) qid:0 cid:5 nsid:0 cdw10:00000000 cdw11:00000000 00:32:35.831 [2024-12-06 17:48:17.393871] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238ca80, cid 4, qid 0 00:32:35.831 [2024-12-06 17:48:17.393883] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238cc00, cid 5, qid 0 00:32:35.831 [2024-12-06 17:48:17.394020] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:32:35.831 [2024-12-06 17:48:17.394031] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:32:35.831 [2024-12-06 17:48:17.394038] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.394044] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x2320d80): datao=0, datal=1024, cccid=4 00:32:35.831 [2024-12-06 17:48:17.394052] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x238ca80) on tqpair(0x2320d80): expected_datao=0, payload_size=1024 00:32:35.831 [2024-12-06 17:48:17.394059] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.394068] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.394075] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.394083] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.831 [2024-12-06 17:48:17.394092] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.831 [2024-12-06 17:48:17.394098] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.394105] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238cc00) on tqpair=0x2320d80 00:32:35.831 [2024-12-06 17:48:17.434776] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.831 [2024-12-06 17:48:17.434794] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.831 [2024-12-06 17:48:17.434801] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.434808] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238ca80) on tqpair=0x2320d80 00:32:35.831 [2024-12-06 17:48:17.434825] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.434834] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x2320d80) 00:32:35.831 [2024-12-06 17:48:17.434845] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:4 nsid:0 cdw10:02ff0070 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.831 [2024-12-06 17:48:17.434875] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238ca80, cid 4, qid 0 00:32:35.831 [2024-12-06 17:48:17.434978] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:32:35.831 [2024-12-06 17:48:17.434993] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:32:35.831 [2024-12-06 17:48:17.435000] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.435006] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x2320d80): datao=0, datal=3072, cccid=4 00:32:35.831 [2024-12-06 17:48:17.435014] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x238ca80) on tqpair(0x2320d80): expected_datao=0, payload_size=3072 00:32:35.831 [2024-12-06 17:48:17.435021] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.435031] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.435039] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.435050] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.831 [2024-12-06 17:48:17.435060] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.831 [2024-12-06 17:48:17.435066] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.435072] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238ca80) on tqpair=0x2320d80 00:32:35.831 [2024-12-06 17:48:17.435086] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.831 [2024-12-06 17:48:17.435095] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x2320d80) 00:32:35.831 [2024-12-06 17:48:17.435110] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:4 nsid:0 cdw10:00010070 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.831 [2024-12-06 17:48:17.435139] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238ca80, cid 4, qid 0 00:32:35.832 [2024-12-06 17:48:17.435234] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:32:35.832 [2024-12-06 17:48:17.435246] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:32:35.832 [2024-12-06 17:48:17.435252] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:32:35.832 [2024-12-06 17:48:17.435258] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x2320d80): datao=0, datal=8, cccid=4 00:32:35.832 [2024-12-06 17:48:17.435266] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x238ca80) on tqpair(0x2320d80): expected_datao=0, payload_size=8 00:32:35.832 [2024-12-06 17:48:17.435273] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.832 [2024-12-06 17:48:17.435282] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:32:35.832 [2024-12-06 17:48:17.435289] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:32:35.832 [2024-12-06 17:48:17.479685] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.832 [2024-12-06 17:48:17.479718] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.832 [2024-12-06 17:48:17.479726] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.832 [2024-12-06 17:48:17.479733] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238ca80) on tqpair=0x2320d80 00:32:35.832 ===================================================== 00:32:35.832 NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2014-08.org.nvmexpress.discovery 00:32:35.832 ===================================================== 00:32:35.832 Controller Capabilities/Features 00:32:35.832 ================================ 00:32:35.832 Vendor ID: 0000 00:32:35.832 Subsystem Vendor ID: 0000 00:32:35.832 Serial Number: .................... 00:32:35.832 Model Number: ........................................ 00:32:35.832 Firmware Version: 25.01 00:32:35.832 Recommended Arb Burst: 0 00:32:35.832 IEEE OUI Identifier: 00 00 00 00:32:35.832 Multi-path I/O 00:32:35.832 May have multiple subsystem ports: No 00:32:35.832 May have multiple controllers: No 00:32:35.832 Associated with SR-IOV VF: No 00:32:35.832 Max Data Transfer Size: 131072 00:32:35.832 Max Number of Namespaces: 0 00:32:35.832 Max Number of I/O Queues: 1024 00:32:35.832 NVMe Specification Version (VS): 1.3 00:32:35.832 NVMe Specification Version (Identify): 1.3 00:32:35.832 Maximum Queue Entries: 128 00:32:35.832 Contiguous Queues Required: Yes 00:32:35.832 Arbitration Mechanisms Supported 00:32:35.832 Weighted Round Robin: Not Supported 00:32:35.832 Vendor Specific: Not Supported 00:32:35.832 Reset Timeout: 15000 ms 00:32:35.832 Doorbell Stride: 4 bytes 00:32:35.832 NVM Subsystem Reset: Not Supported 00:32:35.832 Command Sets Supported 00:32:35.832 NVM Command Set: Supported 00:32:35.832 Boot Partition: Not Supported 00:32:35.832 Memory Page Size Minimum: 4096 bytes 00:32:35.832 Memory Page Size Maximum: 4096 bytes 00:32:35.832 Persistent Memory Region: Not Supported 00:32:35.832 Optional Asynchronous Events Supported 00:32:35.832 Namespace Attribute Notices: Not Supported 00:32:35.832 Firmware Activation Notices: Not Supported 00:32:35.832 ANA Change Notices: Not Supported 00:32:35.832 PLE Aggregate Log Change Notices: Not Supported 00:32:35.832 LBA Status Info Alert Notices: Not Supported 00:32:35.832 EGE Aggregate Log Change Notices: Not Supported 00:32:35.832 Normal NVM Subsystem Shutdown event: Not Supported 00:32:35.832 Zone Descriptor Change Notices: Not Supported 00:32:35.832 Discovery Log Change Notices: Supported 00:32:35.832 Controller Attributes 00:32:35.832 128-bit Host Identifier: Not Supported 00:32:35.832 Non-Operational Permissive Mode: Not Supported 00:32:35.832 NVM Sets: Not Supported 00:32:35.832 Read Recovery Levels: Not Supported 00:32:35.832 Endurance Groups: Not Supported 00:32:35.832 Predictable Latency Mode: Not Supported 00:32:35.832 Traffic Based Keep ALive: Not Supported 00:32:35.832 Namespace Granularity: Not Supported 00:32:35.832 SQ Associations: Not Supported 00:32:35.832 UUID List: Not Supported 00:32:35.832 Multi-Domain Subsystem: Not Supported 00:32:35.832 Fixed Capacity Management: Not Supported 00:32:35.832 Variable Capacity Management: Not Supported 00:32:35.832 Delete Endurance Group: Not Supported 00:32:35.832 Delete NVM Set: Not Supported 00:32:35.832 Extended LBA Formats Supported: Not Supported 00:32:35.832 Flexible Data Placement Supported: Not Supported 00:32:35.832 00:32:35.832 Controller Memory Buffer Support 00:32:35.832 ================================ 00:32:35.832 Supported: No 00:32:35.832 00:32:35.832 Persistent Memory Region Support 00:32:35.832 ================================ 00:32:35.832 Supported: No 00:32:35.832 00:32:35.832 Admin Command Set Attributes 00:32:35.832 ============================ 00:32:35.832 Security Send/Receive: Not Supported 00:32:35.832 Format NVM: Not Supported 00:32:35.832 Firmware Activate/Download: Not Supported 00:32:35.832 Namespace Management: Not Supported 00:32:35.832 Device Self-Test: Not Supported 00:32:35.832 Directives: Not Supported 00:32:35.832 NVMe-MI: Not Supported 00:32:35.832 Virtualization Management: Not Supported 00:32:35.832 Doorbell Buffer Config: Not Supported 00:32:35.832 Get LBA Status Capability: Not Supported 00:32:35.832 Command & Feature Lockdown Capability: Not Supported 00:32:35.832 Abort Command Limit: 1 00:32:35.832 Async Event Request Limit: 4 00:32:35.832 Number of Firmware Slots: N/A 00:32:35.832 Firmware Slot 1 Read-Only: N/A 00:32:35.832 Firmware Activation Without Reset: N/A 00:32:35.832 Multiple Update Detection Support: N/A 00:32:35.832 Firmware Update Granularity: No Information Provided 00:32:35.832 Per-Namespace SMART Log: No 00:32:35.832 Asymmetric Namespace Access Log Page: Not Supported 00:32:35.832 Subsystem NQN: nqn.2014-08.org.nvmexpress.discovery 00:32:35.832 Command Effects Log Page: Not Supported 00:32:35.832 Get Log Page Extended Data: Supported 00:32:35.832 Telemetry Log Pages: Not Supported 00:32:35.832 Persistent Event Log Pages: Not Supported 00:32:35.832 Supported Log Pages Log Page: May Support 00:32:35.832 Commands Supported & Effects Log Page: Not Supported 00:32:35.832 Feature Identifiers & Effects Log Page:May Support 00:32:35.832 NVMe-MI Commands & Effects Log Page: May Support 00:32:35.832 Data Area 4 for Telemetry Log: Not Supported 00:32:35.832 Error Log Page Entries Supported: 128 00:32:35.832 Keep Alive: Not Supported 00:32:35.832 00:32:35.832 NVM Command Set Attributes 00:32:35.832 ========================== 00:32:35.832 Submission Queue Entry Size 00:32:35.832 Max: 1 00:32:35.832 Min: 1 00:32:35.832 Completion Queue Entry Size 00:32:35.832 Max: 1 00:32:35.832 Min: 1 00:32:35.832 Number of Namespaces: 0 00:32:35.832 Compare Command: Not Supported 00:32:35.832 Write Uncorrectable Command: Not Supported 00:32:35.832 Dataset Management Command: Not Supported 00:32:35.832 Write Zeroes Command: Not Supported 00:32:35.832 Set Features Save Field: Not Supported 00:32:35.832 Reservations: Not Supported 00:32:35.832 Timestamp: Not Supported 00:32:35.832 Copy: Not Supported 00:32:35.832 Volatile Write Cache: Not Present 00:32:35.832 Atomic Write Unit (Normal): 1 00:32:35.832 Atomic Write Unit (PFail): 1 00:32:35.832 Atomic Compare & Write Unit: 1 00:32:35.832 Fused Compare & Write: Supported 00:32:35.832 Scatter-Gather List 00:32:35.832 SGL Command Set: Supported 00:32:35.832 SGL Keyed: Supported 00:32:35.832 SGL Bit Bucket Descriptor: Not Supported 00:32:35.832 SGL Metadata Pointer: Not Supported 00:32:35.832 Oversized SGL: Not Supported 00:32:35.832 SGL Metadata Address: Not Supported 00:32:35.832 SGL Offset: Supported 00:32:35.832 Transport SGL Data Block: Not Supported 00:32:35.832 Replay Protected Memory Block: Not Supported 00:32:35.832 00:32:35.832 Firmware Slot Information 00:32:35.832 ========================= 00:32:35.832 Active slot: 0 00:32:35.832 00:32:35.832 00:32:35.832 Error Log 00:32:35.832 ========= 00:32:35.832 00:32:35.832 Active Namespaces 00:32:35.832 ================= 00:32:35.832 Discovery Log Page 00:32:35.832 ================== 00:32:35.832 Generation Counter: 2 00:32:35.832 Number of Records: 2 00:32:35.832 Record Format: 0 00:32:35.832 00:32:35.832 Discovery Log Entry 0 00:32:35.832 ---------------------- 00:32:35.832 Transport Type: 3 (TCP) 00:32:35.832 Address Family: 1 (IPv4) 00:32:35.832 Subsystem Type: 3 (Current Discovery Subsystem) 00:32:35.832 Entry Flags: 00:32:35.832 Duplicate Returned Information: 1 00:32:35.832 Explicit Persistent Connection Support for Discovery: 1 00:32:35.832 Transport Requirements: 00:32:35.832 Secure Channel: Not Required 00:32:35.832 Port ID: 0 (0x0000) 00:32:35.832 Controller ID: 65535 (0xffff) 00:32:35.832 Admin Max SQ Size: 128 00:32:35.832 Transport Service Identifier: 4420 00:32:35.832 NVM Subsystem Qualified Name: nqn.2014-08.org.nvmexpress.discovery 00:32:35.832 Transport Address: 10.0.0.2 00:32:35.833 Discovery Log Entry 1 00:32:35.833 ---------------------- 00:32:35.833 Transport Type: 3 (TCP) 00:32:35.833 Address Family: 1 (IPv4) 00:32:35.833 Subsystem Type: 2 (NVM Subsystem) 00:32:35.833 Entry Flags: 00:32:35.833 Duplicate Returned Information: 0 00:32:35.833 Explicit Persistent Connection Support for Discovery: 0 00:32:35.833 Transport Requirements: 00:32:35.833 Secure Channel: Not Required 00:32:35.833 Port ID: 0 (0x0000) 00:32:35.833 Controller ID: 65535 (0xffff) 00:32:35.833 Admin Max SQ Size: 128 00:32:35.833 Transport Service Identifier: 4420 00:32:35.833 NVM Subsystem Qualified Name: nqn.2016-06.io.spdk:cnode1 00:32:35.833 Transport Address: 10.0.0.2 [2024-12-06 17:48:17.479848] nvme_ctrlr.c:4399:nvme_ctrlr_destruct_async: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] Prepare to destruct SSD 00:32:35.833 [2024-12-06 17:48:17.479870] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c480) on tqpair=0x2320d80 00:32:35.833 [2024-12-06 17:48:17.479883] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:35.833 [2024-12-06 17:48:17.479892] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c600) on tqpair=0x2320d80 00:32:35.833 [2024-12-06 17:48:17.479899] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:35.833 [2024-12-06 17:48:17.479907] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c780) on tqpair=0x2320d80 00:32:35.833 [2024-12-06 17:48:17.479915] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:35.833 [2024-12-06 17:48:17.479923] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c900) on tqpair=0x2320d80 00:32:35.833 [2024-12-06 17:48:17.479930] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:35.833 [2024-12-06 17:48:17.479948] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.479957] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.479963] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x2320d80) 00:32:35.833 [2024-12-06 17:48:17.479975] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.833 [2024-12-06 17:48:17.480000] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c900, cid 3, qid 0 00:32:35.833 [2024-12-06 17:48:17.480121] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.833 [2024-12-06 17:48:17.480134] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.833 [2024-12-06 17:48:17.480140] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.480147] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c900) on tqpair=0x2320d80 00:32:35.833 [2024-12-06 17:48:17.480159] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.480166] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.480173] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x2320d80) 00:32:35.833 [2024-12-06 17:48:17.480187] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY SET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.833 [2024-12-06 17:48:17.480215] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c900, cid 3, qid 0 00:32:35.833 [2024-12-06 17:48:17.480302] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.833 [2024-12-06 17:48:17.480314] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.833 [2024-12-06 17:48:17.480321] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.480328] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c900) on tqpair=0x2320d80 00:32:35.833 [2024-12-06 17:48:17.480336] nvme_ctrlr.c:1151:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] RTD3E = 0 us 00:32:35.833 [2024-12-06 17:48:17.480343] nvme_ctrlr.c:1154:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] shutdown timeout = 10000 ms 00:32:35.833 [2024-12-06 17:48:17.480359] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.480368] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.480374] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x2320d80) 00:32:35.833 [2024-12-06 17:48:17.480384] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.833 [2024-12-06 17:48:17.480405] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c900, cid 3, qid 0 00:32:35.833 [2024-12-06 17:48:17.480479] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.833 [2024-12-06 17:48:17.480490] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.833 [2024-12-06 17:48:17.480497] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.480503] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c900) on tqpair=0x2320d80 00:32:35.833 [2024-12-06 17:48:17.480520] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.480529] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.480535] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x2320d80) 00:32:35.833 [2024-12-06 17:48:17.480545] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.833 [2024-12-06 17:48:17.480565] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c900, cid 3, qid 0 00:32:35.833 [2024-12-06 17:48:17.480640] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.833 [2024-12-06 17:48:17.480653] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.833 [2024-12-06 17:48:17.480660] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.480675] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c900) on tqpair=0x2320d80 00:32:35.833 [2024-12-06 17:48:17.480692] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.480701] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.480708] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x2320d80) 00:32:35.833 [2024-12-06 17:48:17.480718] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.833 [2024-12-06 17:48:17.480740] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c900, cid 3, qid 0 00:32:35.833 [2024-12-06 17:48:17.480824] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.833 [2024-12-06 17:48:17.480839] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.833 [2024-12-06 17:48:17.480845] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.480852] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c900) on tqpair=0x2320d80 00:32:35.833 [2024-12-06 17:48:17.480869] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.480882] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.480889] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x2320d80) 00:32:35.833 [2024-12-06 17:48:17.480899] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.833 [2024-12-06 17:48:17.480921] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c900, cid 3, qid 0 00:32:35.833 [2024-12-06 17:48:17.480995] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.833 [2024-12-06 17:48:17.481008] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.833 [2024-12-06 17:48:17.481015] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.481022] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c900) on tqpair=0x2320d80 00:32:35.833 [2024-12-06 17:48:17.481038] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.481047] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.481053] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x2320d80) 00:32:35.833 [2024-12-06 17:48:17.481063] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.833 [2024-12-06 17:48:17.481084] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c900, cid 3, qid 0 00:32:35.833 [2024-12-06 17:48:17.481155] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.833 [2024-12-06 17:48:17.481169] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.833 [2024-12-06 17:48:17.481175] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.481182] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c900) on tqpair=0x2320d80 00:32:35.833 [2024-12-06 17:48:17.481198] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.481207] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.481213] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x2320d80) 00:32:35.833 [2024-12-06 17:48:17.481224] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.833 [2024-12-06 17:48:17.481244] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c900, cid 3, qid 0 00:32:35.833 [2024-12-06 17:48:17.481315] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.833 [2024-12-06 17:48:17.481328] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.833 [2024-12-06 17:48:17.481335] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.481342] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c900) on tqpair=0x2320d80 00:32:35.833 [2024-12-06 17:48:17.481358] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.481367] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.481373] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x2320d80) 00:32:35.833 [2024-12-06 17:48:17.481384] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.833 [2024-12-06 17:48:17.481404] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c900, cid 3, qid 0 00:32:35.833 [2024-12-06 17:48:17.481480] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.833 [2024-12-06 17:48:17.481494] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.833 [2024-12-06 17:48:17.481500] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.481507] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c900) on tqpair=0x2320d80 00:32:35.833 [2024-12-06 17:48:17.481523] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.481532] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.833 [2024-12-06 17:48:17.481542] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x2320d80) 00:32:35.834 [2024-12-06 17:48:17.481553] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.834 [2024-12-06 17:48:17.481573] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c900, cid 3, qid 0 00:32:35.834 [2024-12-06 17:48:17.481647] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.834 [2024-12-06 17:48:17.481661] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.834 [2024-12-06 17:48:17.481677] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.481684] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c900) on tqpair=0x2320d80 00:32:35.834 [2024-12-06 17:48:17.481701] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.481710] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.481717] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x2320d80) 00:32:35.834 [2024-12-06 17:48:17.481727] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.834 [2024-12-06 17:48:17.481748] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c900, cid 3, qid 0 00:32:35.834 [2024-12-06 17:48:17.481821] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.834 [2024-12-06 17:48:17.481833] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.834 [2024-12-06 17:48:17.481840] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.481846] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c900) on tqpair=0x2320d80 00:32:35.834 [2024-12-06 17:48:17.481862] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.481871] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.481878] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x2320d80) 00:32:35.834 [2024-12-06 17:48:17.481888] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.834 [2024-12-06 17:48:17.481908] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c900, cid 3, qid 0 00:32:35.834 [2024-12-06 17:48:17.481985] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.834 [2024-12-06 17:48:17.481999] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.834 [2024-12-06 17:48:17.482006] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.482012] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c900) on tqpair=0x2320d80 00:32:35.834 [2024-12-06 17:48:17.482028] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.482037] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.482043] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x2320d80) 00:32:35.834 [2024-12-06 17:48:17.482053] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.834 [2024-12-06 17:48:17.482073] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c900, cid 3, qid 0 00:32:35.834 [2024-12-06 17:48:17.482143] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.834 [2024-12-06 17:48:17.482155] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.834 [2024-12-06 17:48:17.482162] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.482168] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c900) on tqpair=0x2320d80 00:32:35.834 [2024-12-06 17:48:17.482184] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.482193] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.482199] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x2320d80) 00:32:35.834 [2024-12-06 17:48:17.482214] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.834 [2024-12-06 17:48:17.482235] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c900, cid 3, qid 0 00:32:35.834 [2024-12-06 17:48:17.482309] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.834 [2024-12-06 17:48:17.482323] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.834 [2024-12-06 17:48:17.482330] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.482336] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c900) on tqpair=0x2320d80 00:32:35.834 [2024-12-06 17:48:17.482352] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.482361] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.482368] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x2320d80) 00:32:35.834 [2024-12-06 17:48:17.482378] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.834 [2024-12-06 17:48:17.482398] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c900, cid 3, qid 0 00:32:35.834 [2024-12-06 17:48:17.482467] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.834 [2024-12-06 17:48:17.482478] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.834 [2024-12-06 17:48:17.482485] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.482491] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c900) on tqpair=0x2320d80 00:32:35.834 [2024-12-06 17:48:17.482507] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.482516] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.482522] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x2320d80) 00:32:35.834 [2024-12-06 17:48:17.482532] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.834 [2024-12-06 17:48:17.482553] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c900, cid 3, qid 0 00:32:35.834 [2024-12-06 17:48:17.482624] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.834 [2024-12-06 17:48:17.482637] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.834 [2024-12-06 17:48:17.482644] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.482651] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c900) on tqpair=0x2320d80 00:32:35.834 [2024-12-06 17:48:17.482673] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.482684] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.482690] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x2320d80) 00:32:35.834 [2024-12-06 17:48:17.482701] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.834 [2024-12-06 17:48:17.482722] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c900, cid 3, qid 0 00:32:35.834 [2024-12-06 17:48:17.482800] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.834 [2024-12-06 17:48:17.482813] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.834 [2024-12-06 17:48:17.482820] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.482826] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c900) on tqpair=0x2320d80 00:32:35.834 [2024-12-06 17:48:17.482842] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.482851] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.482858] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x2320d80) 00:32:35.834 [2024-12-06 17:48:17.482868] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.834 [2024-12-06 17:48:17.482893] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c900, cid 3, qid 0 00:32:35.834 [2024-12-06 17:48:17.482967] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.834 [2024-12-06 17:48:17.482979] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.834 [2024-12-06 17:48:17.482985] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.482992] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c900) on tqpair=0x2320d80 00:32:35.834 [2024-12-06 17:48:17.483008] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.483017] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.483023] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x2320d80) 00:32:35.834 [2024-12-06 17:48:17.483033] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.834 [2024-12-06 17:48:17.483053] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c900, cid 3, qid 0 00:32:35.834 [2024-12-06 17:48:17.483127] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.834 [2024-12-06 17:48:17.483140] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.834 [2024-12-06 17:48:17.483146] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.483153] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c900) on tqpair=0x2320d80 00:32:35.834 [2024-12-06 17:48:17.483169] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.483178] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.483184] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x2320d80) 00:32:35.834 [2024-12-06 17:48:17.483194] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.834 [2024-12-06 17:48:17.483215] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c900, cid 3, qid 0 00:32:35.834 [2024-12-06 17:48:17.483287] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.834 [2024-12-06 17:48:17.483299] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.834 [2024-12-06 17:48:17.483305] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.483312] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c900) on tqpair=0x2320d80 00:32:35.834 [2024-12-06 17:48:17.483327] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.483336] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.483343] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x2320d80) 00:32:35.834 [2024-12-06 17:48:17.483353] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.834 [2024-12-06 17:48:17.483373] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c900, cid 3, qid 0 00:32:35.834 [2024-12-06 17:48:17.483444] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.834 [2024-12-06 17:48:17.483457] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.834 [2024-12-06 17:48:17.483463] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.483470] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c900) on tqpair=0x2320d80 00:32:35.834 [2024-12-06 17:48:17.483486] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.834 [2024-12-06 17:48:17.483495] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.835 [2024-12-06 17:48:17.483501] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x2320d80) 00:32:35.835 [2024-12-06 17:48:17.483512] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.835 [2024-12-06 17:48:17.483537] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c900, cid 3, qid 0 00:32:35.835 [2024-12-06 17:48:17.483608] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.835 [2024-12-06 17:48:17.483621] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.835 [2024-12-06 17:48:17.483628] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.835 [2024-12-06 17:48:17.483634] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c900) on tqpair=0x2320d80 00:32:35.835 [2024-12-06 17:48:17.483650] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.835 [2024-12-06 17:48:17.483659] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.835 [2024-12-06 17:48:17.487677] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x2320d80) 00:32:35.835 [2024-12-06 17:48:17.487691] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.835 [2024-12-06 17:48:17.487714] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x238c900, cid 3, qid 0 00:32:35.835 [2024-12-06 17:48:17.487841] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.835 [2024-12-06 17:48:17.487854] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.835 [2024-12-06 17:48:17.487860] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.835 [2024-12-06 17:48:17.487867] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x238c900) on tqpair=0x2320d80 00:32:35.835 [2024-12-06 17:48:17.487880] nvme_ctrlr.c:1273:nvme_ctrlr_shutdown_poll_async: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] shutdown complete in 7 milliseconds 00:32:35.835 00:32:35.835 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' -L all 00:32:35.835 [2024-12-06 17:48:17.524626] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:32:35.835 [2024-12-06 17:48:17.524686] [ DPDK EAL parameters: identify --no-shconf -c 0x1 -n 1 -m 0 --no-pci --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid350878 ] 00:32:35.835 [2024-12-06 17:48:17.575378] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 0] setting state to connect adminq (no timeout) 00:32:35.835 [2024-12-06 17:48:17.575433] nvme_tcp.c:2238:nvme_tcp_qpair_connect_sock: *DEBUG*: adrfam 1 ai_family 2 00:32:35.835 [2024-12-06 17:48:17.575443] nvme_tcp.c:2242:nvme_tcp_qpair_connect_sock: *DEBUG*: trsvcid is 4420 00:32:35.835 [2024-12-06 17:48:17.575460] nvme_tcp.c:2263:nvme_tcp_qpair_connect_sock: *DEBUG*: sock_impl_name is (null) 00:32:35.835 [2024-12-06 17:48:17.575471] sock.c: 373:spdk_sock_connect_ext: *DEBUG*: Creating a client socket using impl posix 00:32:35.835 [2024-12-06 17:48:17.578949] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 0] setting state to wait for connect adminq (no timeout) 00:32:35.835 [2024-12-06 17:48:17.579002] nvme_tcp.c:1455:nvme_tcp_send_icreq_complete: *DEBUG*: Complete the icreq send for tqpair=0x6d3d80 0 00:32:35.835 [2024-12-06 17:48:17.585682] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 1 00:32:35.835 [2024-12-06 17:48:17.585703] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =1 00:32:35.835 [2024-12-06 17:48:17.585711] nvme_tcp.c:1501:nvme_tcp_icresp_handle: *DEBUG*: host_hdgst_enable: 0 00:32:35.835 [2024-12-06 17:48:17.585718] nvme_tcp.c:1502:nvme_tcp_icresp_handle: *DEBUG*: host_ddgst_enable: 0 00:32:35.835 [2024-12-06 17:48:17.585751] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.835 [2024-12-06 17:48:17.585763] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.835 [2024-12-06 17:48:17.585773] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x6d3d80) 00:32:35.835 [2024-12-06 17:48:17.585790] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:0 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x400 00:32:35.835 [2024-12-06 17:48:17.585817] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73f480, cid 0, qid 0 00:32:35.835 [2024-12-06 17:48:17.593686] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.835 [2024-12-06 17:48:17.593704] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.835 [2024-12-06 17:48:17.593712] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.835 [2024-12-06 17:48:17.593720] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73f480) on tqpair=0x6d3d80 00:32:35.835 [2024-12-06 17:48:17.593738] nvme_fabric.c: 621:nvme_fabric_qpair_connect_poll: *DEBUG*: CNTLID 0x0001 00:32:35.835 [2024-12-06 17:48:17.593749] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to read vs (no timeout) 00:32:35.835 [2024-12-06 17:48:17.593759] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to read vs wait for vs (no timeout) 00:32:35.835 [2024-12-06 17:48:17.593777] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.835 [2024-12-06 17:48:17.593786] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.835 [2024-12-06 17:48:17.593793] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x6d3d80) 00:32:35.835 [2024-12-06 17:48:17.593805] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.835 [2024-12-06 17:48:17.593829] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73f480, cid 0, qid 0 00:32:35.835 [2024-12-06 17:48:17.593941] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.835 [2024-12-06 17:48:17.593956] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.835 [2024-12-06 17:48:17.593963] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.835 [2024-12-06 17:48:17.593970] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73f480) on tqpair=0x6d3d80 00:32:35.835 [2024-12-06 17:48:17.593978] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to read cap (no timeout) 00:32:35.835 [2024-12-06 17:48:17.593992] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to read cap wait for cap (no timeout) 00:32:35.835 [2024-12-06 17:48:17.594005] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.835 [2024-12-06 17:48:17.594013] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.835 [2024-12-06 17:48:17.594019] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x6d3d80) 00:32:35.835 [2024-12-06 17:48:17.594031] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.835 [2024-12-06 17:48:17.594053] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73f480, cid 0, qid 0 00:32:35.835 [2024-12-06 17:48:17.594127] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.835 [2024-12-06 17:48:17.594139] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.835 [2024-12-06 17:48:17.594146] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.835 [2024-12-06 17:48:17.594153] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73f480) on tqpair=0x6d3d80 00:32:35.835 [2024-12-06 17:48:17.594162] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to check en (no timeout) 00:32:35.835 [2024-12-06 17:48:17.594176] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to check en wait for cc (timeout 15000 ms) 00:32:35.835 [2024-12-06 17:48:17.594188] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.835 [2024-12-06 17:48:17.594196] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.835 [2024-12-06 17:48:17.594202] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x6d3d80) 00:32:35.835 [2024-12-06 17:48:17.594217] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.835 [2024-12-06 17:48:17.594251] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73f480, cid 0, qid 0 00:32:35.835 [2024-12-06 17:48:17.594325] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.835 [2024-12-06 17:48:17.594339] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.835 [2024-12-06 17:48:17.594346] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.835 [2024-12-06 17:48:17.594353] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73f480) on tqpair=0x6d3d80 00:32:35.835 [2024-12-06 17:48:17.594362] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to disable and wait for CSTS.RDY = 0 (timeout 15000 ms) 00:32:35.835 [2024-12-06 17:48:17.594379] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.835 [2024-12-06 17:48:17.594388] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.835 [2024-12-06 17:48:17.594395] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x6d3d80) 00:32:35.835 [2024-12-06 17:48:17.594406] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.835 [2024-12-06 17:48:17.594427] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73f480, cid 0, qid 0 00:32:35.835 [2024-12-06 17:48:17.594500] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.835 [2024-12-06 17:48:17.594514] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.835 [2024-12-06 17:48:17.594521] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.835 [2024-12-06 17:48:17.594528] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73f480) on tqpair=0x6d3d80 00:32:35.835 [2024-12-06 17:48:17.594536] nvme_ctrlr.c:3906:nvme_ctrlr_process_init_wait_for_ready_0: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] CC.EN = 0 && CSTS.RDY = 0 00:32:35.835 [2024-12-06 17:48:17.594544] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to controller is disabled (timeout 15000 ms) 00:32:35.836 [2024-12-06 17:48:17.594557] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to enable controller by writing CC.EN = 1 (timeout 15000 ms) 00:32:35.836 [2024-12-06 17:48:17.594673] nvme_ctrlr.c:4104:nvme_ctrlr_process_init: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] Setting CC.EN = 1 00:32:35.836 [2024-12-06 17:48:17.594684] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to enable controller by writing CC.EN = 1 reg (timeout 15000 ms) 00:32:35.836 [2024-12-06 17:48:17.594697] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.836 [2024-12-06 17:48:17.594705] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.836 [2024-12-06 17:48:17.594711] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x6d3d80) 00:32:35.836 [2024-12-06 17:48:17.594722] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY SET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.836 [2024-12-06 17:48:17.594744] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73f480, cid 0, qid 0 00:32:35.836 [2024-12-06 17:48:17.594844] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.836 [2024-12-06 17:48:17.594856] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.836 [2024-12-06 17:48:17.594863] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.836 [2024-12-06 17:48:17.594870] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73f480) on tqpair=0x6d3d80 00:32:35.836 [2024-12-06 17:48:17.594878] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to wait for CSTS.RDY = 1 (timeout 15000 ms) 00:32:35.836 [2024-12-06 17:48:17.594895] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.836 [2024-12-06 17:48:17.594905] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.836 [2024-12-06 17:48:17.594911] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x6d3d80) 00:32:35.836 [2024-12-06 17:48:17.594926] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.836 [2024-12-06 17:48:17.594955] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73f480, cid 0, qid 0 00:32:35.836 [2024-12-06 17:48:17.595026] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.836 [2024-12-06 17:48:17.595041] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.836 [2024-12-06 17:48:17.595048] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.836 [2024-12-06 17:48:17.595055] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73f480) on tqpair=0x6d3d80 00:32:35.836 [2024-12-06 17:48:17.595062] nvme_ctrlr.c:3941:nvme_ctrlr_process_init_enable_wait_for_ready_1: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] CC.EN = 1 && CSTS.RDY = 1 - controller is ready 00:32:35.836 [2024-12-06 17:48:17.595071] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to reset admin queue (timeout 30000 ms) 00:32:35.836 [2024-12-06 17:48:17.595084] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to identify controller (no timeout) 00:32:35.836 [2024-12-06 17:48:17.595101] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to wait for identify controller (timeout 30000 ms) 00:32:35.836 [2024-12-06 17:48:17.595116] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.836 [2024-12-06 17:48:17.595124] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x6d3d80) 00:32:35.836 [2024-12-06 17:48:17.595135] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:0 nsid:0 cdw10:00000001 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.836 [2024-12-06 17:48:17.595157] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73f480, cid 0, qid 0 00:32:35.836 [2024-12-06 17:48:17.595269] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:32:35.836 [2024-12-06 17:48:17.595282] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:32:35.836 [2024-12-06 17:48:17.595289] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:32:35.836 [2024-12-06 17:48:17.595295] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x6d3d80): datao=0, datal=4096, cccid=0 00:32:35.836 [2024-12-06 17:48:17.595303] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x73f480) on tqpair(0x6d3d80): expected_datao=0, payload_size=4096 00:32:35.836 [2024-12-06 17:48:17.595311] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.836 [2024-12-06 17:48:17.595328] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:32:35.836 [2024-12-06 17:48:17.595337] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:32:35.836 [2024-12-06 17:48:17.635763] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.836 [2024-12-06 17:48:17.635781] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.836 [2024-12-06 17:48:17.635789] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.836 [2024-12-06 17:48:17.635796] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73f480) on tqpair=0x6d3d80 00:32:35.836 [2024-12-06 17:48:17.635808] nvme_ctrlr.c:2081:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] transport max_xfer_size 4294967295 00:32:35.836 [2024-12-06 17:48:17.635821] nvme_ctrlr.c:2085:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] MDTS max_xfer_size 131072 00:32:35.836 [2024-12-06 17:48:17.635831] nvme_ctrlr.c:2088:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] CNTLID 0x0001 00:32:35.836 [2024-12-06 17:48:17.635838] nvme_ctrlr.c:2112:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] transport max_sges 16 00:32:35.836 [2024-12-06 17:48:17.635845] nvme_ctrlr.c:2127:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] fuses compare and write: 1 00:32:35.836 [2024-12-06 17:48:17.635854] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to configure AER (timeout 30000 ms) 00:32:35.836 [2024-12-06 17:48:17.635872] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to wait for configure aer (timeout 30000 ms) 00:32:35.836 [2024-12-06 17:48:17.635885] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.836 [2024-12-06 17:48:17.635894] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.836 [2024-12-06 17:48:17.635900] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x6d3d80) 00:32:35.836 [2024-12-06 17:48:17.635912] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES ASYNC EVENT CONFIGURATION cid:0 cdw10:0000000b SGL DATA BLOCK OFFSET 0x0 len:0x0 00:32:35.836 [2024-12-06 17:48:17.635935] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73f480, cid 0, qid 0 00:32:35.836 [2024-12-06 17:48:17.636019] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.836 [2024-12-06 17:48:17.636039] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.836 [2024-12-06 17:48:17.636046] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.836 [2024-12-06 17:48:17.636052] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73f480) on tqpair=0x6d3d80 00:32:35.836 [2024-12-06 17:48:17.636063] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.836 [2024-12-06 17:48:17.636070] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.836 [2024-12-06 17:48:17.636077] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x6d3d80) 00:32:35.836 [2024-12-06 17:48:17.636087] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:32:35.836 [2024-12-06 17:48:17.636096] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.836 [2024-12-06 17:48:17.636103] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.836 [2024-12-06 17:48:17.636110] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=1 on tqpair(0x6d3d80) 00:32:35.836 [2024-12-06 17:48:17.636119] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:32:35.836 [2024-12-06 17:48:17.636128] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.836 [2024-12-06 17:48:17.636135] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.836 [2024-12-06 17:48:17.636141] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=2 on tqpair(0x6d3d80) 00:32:35.836 [2024-12-06 17:48:17.636150] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:32:35.836 [2024-12-06 17:48:17.636159] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.836 [2024-12-06 17:48:17.636166] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.836 [2024-12-06 17:48:17.636172] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x6d3d80) 00:32:35.836 [2024-12-06 17:48:17.636181] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:32:35.836 [2024-12-06 17:48:17.636190] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to set keep alive timeout (timeout 30000 ms) 00:32:35.836 [2024-12-06 17:48:17.636209] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to wait for set keep alive timeout (timeout 30000 ms) 00:32:35.836 [2024-12-06 17:48:17.636237] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.836 [2024-12-06 17:48:17.636245] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x6d3d80) 00:32:35.836 [2024-12-06 17:48:17.636255] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES KEEP ALIVE TIMER cid:4 cdw10:0000000f SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.836 [2024-12-06 17:48:17.636278] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73f480, cid 0, qid 0 00:32:35.836 [2024-12-06 17:48:17.636304] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73f600, cid 1, qid 0 00:32:35.836 [2024-12-06 17:48:17.636316] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73f780, cid 2, qid 0 00:32:35.836 [2024-12-06 17:48:17.636325] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73f900, cid 3, qid 0 00:32:35.836 [2024-12-06 17:48:17.636333] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73fa80, cid 4, qid 0 00:32:35.836 [2024-12-06 17:48:17.636458] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.836 [2024-12-06 17:48:17.636472] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.836 [2024-12-06 17:48:17.636479] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.836 [2024-12-06 17:48:17.636486] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73fa80) on tqpair=0x6d3d80 00:32:35.836 [2024-12-06 17:48:17.636494] nvme_ctrlr.c:3059:nvme_ctrlr_set_keep_alive_timeout_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] Sending keep alive every 5000000 us 00:32:35.836 [2024-12-06 17:48:17.636503] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to identify controller iocs specific (timeout 30000 ms) 00:32:35.836 [2024-12-06 17:48:17.636517] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to set number of queues (timeout 30000 ms) 00:32:35.836 [2024-12-06 17:48:17.636528] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to wait for set number of queues (timeout 30000 ms) 00:32:35.836 [2024-12-06 17:48:17.636538] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.836 [2024-12-06 17:48:17.636546] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.836 [2024-12-06 17:48:17.636553] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x6d3d80) 00:32:35.836 [2024-12-06 17:48:17.636563] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES NUMBER OF QUEUES cid:4 cdw10:00000007 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:32:35.836 [2024-12-06 17:48:17.636585] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73fa80, cid 4, qid 0 00:32:35.836 [2024-12-06 17:48:17.636705] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.837 [2024-12-06 17:48:17.636721] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.837 [2024-12-06 17:48:17.636727] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.837 [2024-12-06 17:48:17.636734] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73fa80) on tqpair=0x6d3d80 00:32:35.837 [2024-12-06 17:48:17.636803] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to identify active ns (timeout 30000 ms) 00:32:35.837 [2024-12-06 17:48:17.636823] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to wait for identify active ns (timeout 30000 ms) 00:32:35.837 [2024-12-06 17:48:17.636838] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.837 [2024-12-06 17:48:17.636846] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x6d3d80) 00:32:35.837 [2024-12-06 17:48:17.636857] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:4 nsid:0 cdw10:00000002 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.837 [2024-12-06 17:48:17.636879] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73fa80, cid 4, qid 0 00:32:35.837 [2024-12-06 17:48:17.637023] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:32:35.837 [2024-12-06 17:48:17.637038] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:32:35.837 [2024-12-06 17:48:17.637045] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:32:35.837 [2024-12-06 17:48:17.637051] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x6d3d80): datao=0, datal=4096, cccid=4 00:32:35.837 [2024-12-06 17:48:17.637059] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x73fa80) on tqpair(0x6d3d80): expected_datao=0, payload_size=4096 00:32:35.837 [2024-12-06 17:48:17.637066] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.837 [2024-12-06 17:48:17.637076] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:32:35.837 [2024-12-06 17:48:17.637088] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:32:35.837 [2024-12-06 17:48:17.637100] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:35.837 [2024-12-06 17:48:17.637110] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:35.837 [2024-12-06 17:48:17.637117] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:35.837 [2024-12-06 17:48:17.637124] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73fa80) on tqpair=0x6d3d80 00:32:35.837 [2024-12-06 17:48:17.637144] nvme_ctrlr.c:4735:spdk_nvme_ctrlr_get_ns: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] Namespace 1 was added 00:32:35.837 [2024-12-06 17:48:17.637163] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to identify ns (timeout 30000 ms) 00:32:35.837 [2024-12-06 17:48:17.637181] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to wait for identify ns (timeout 30000 ms) 00:32:35.837 [2024-12-06 17:48:17.637195] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:35.837 [2024-12-06 17:48:17.637203] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x6d3d80) 00:32:35.837 [2024-12-06 17:48:17.637213] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:4 nsid:1 cdw10:00000000 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:35.837 [2024-12-06 17:48:17.637235] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73fa80, cid 4, qid 0 00:32:35.837 [2024-12-06 17:48:17.637343] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:32:35.837 [2024-12-06 17:48:17.637358] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:32:35.837 [2024-12-06 17:48:17.637365] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:32:35.837 [2024-12-06 17:48:17.637371] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x6d3d80): datao=0, datal=4096, cccid=4 00:32:35.837 [2024-12-06 17:48:17.637379] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x73fa80) on tqpair(0x6d3d80): expected_datao=0, payload_size=4096 00:32:35.837 [2024-12-06 17:48:17.637386] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:35.837 [2024-12-06 17:48:17.637403] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:32:35.837 [2024-12-06 17:48:17.637413] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:32:36.111 [2024-12-06 17:48:17.681677] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:36.111 [2024-12-06 17:48:17.681696] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:36.111 [2024-12-06 17:48:17.681704] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:36.111 [2024-12-06 17:48:17.681711] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73fa80) on tqpair=0x6d3d80 00:32:36.111 [2024-12-06 17:48:17.681733] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to identify namespace id descriptors (timeout 30000 ms) 00:32:36.111 [2024-12-06 17:48:17.681754] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to wait for identify namespace id descriptors (timeout 30000 ms) 00:32:36.111 [2024-12-06 17:48:17.681769] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:36.111 [2024-12-06 17:48:17.681778] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x6d3d80) 00:32:36.111 [2024-12-06 17:48:17.681789] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:4 nsid:1 cdw10:00000003 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:36.111 [2024-12-06 17:48:17.681814] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73fa80, cid 4, qid 0 00:32:36.111 [2024-12-06 17:48:17.681937] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:32:36.111 [2024-12-06 17:48:17.681951] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:32:36.111 [2024-12-06 17:48:17.681958] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:32:36.111 [2024-12-06 17:48:17.681965] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x6d3d80): datao=0, datal=4096, cccid=4 00:32:36.111 [2024-12-06 17:48:17.681972] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x73fa80) on tqpair(0x6d3d80): expected_datao=0, payload_size=4096 00:32:36.111 [2024-12-06 17:48:17.681984] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:36.111 [2024-12-06 17:48:17.682003] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:32:36.111 [2024-12-06 17:48:17.682012] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:32:36.111 [2024-12-06 17:48:17.722778] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:36.112 [2024-12-06 17:48:17.722797] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:36.112 [2024-12-06 17:48:17.722804] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.722812] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73fa80) on tqpair=0x6d3d80 00:32:36.112 [2024-12-06 17:48:17.722825] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to identify ns iocs specific (timeout 30000 ms) 00:32:36.112 [2024-12-06 17:48:17.722840] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to set supported log pages (timeout 30000 ms) 00:32:36.112 [2024-12-06 17:48:17.722856] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to set supported features (timeout 30000 ms) 00:32:36.112 [2024-12-06 17:48:17.722870] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to set host behavior support feature (timeout 30000 ms) 00:32:36.112 [2024-12-06 17:48:17.722880] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to set doorbell buffer config (timeout 30000 ms) 00:32:36.112 [2024-12-06 17:48:17.722889] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to set host ID (timeout 30000 ms) 00:32:36.112 [2024-12-06 17:48:17.722897] nvme_ctrlr.c:3147:nvme_ctrlr_set_host_id: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] NVMe-oF transport - not sending Set Features - Host ID 00:32:36.112 [2024-12-06 17:48:17.722905] nvme_ctrlr.c:1561:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to transport ready (timeout 30000 ms) 00:32:36.112 [2024-12-06 17:48:17.722914] nvme_ctrlr.c:1567:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to ready (no timeout) 00:32:36.112 [2024-12-06 17:48:17.722934] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.722942] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x6d3d80) 00:32:36.112 [2024-12-06 17:48:17.722954] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ARBITRATION cid:4 cdw10:00000001 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:36.112 [2024-12-06 17:48:17.722965] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.722973] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.722979] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x6d3d80) 00:32:36.112 [2024-12-06 17:48:17.722988] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: KEEP ALIVE (18) qid:0 cid:5 nsid:0 cdw10:00000000 cdw11:00000000 00:32:36.112 [2024-12-06 17:48:17.723015] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73fa80, cid 4, qid 0 00:32:36.112 [2024-12-06 17:48:17.723027] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73fc00, cid 5, qid 0 00:32:36.112 [2024-12-06 17:48:17.723120] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:36.112 [2024-12-06 17:48:17.723134] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:36.112 [2024-12-06 17:48:17.723141] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.723148] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73fa80) on tqpair=0x6d3d80 00:32:36.112 [2024-12-06 17:48:17.723158] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:36.112 [2024-12-06 17:48:17.723168] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:36.112 [2024-12-06 17:48:17.723174] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.723185] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73fc00) on tqpair=0x6d3d80 00:32:36.112 [2024-12-06 17:48:17.723202] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.723211] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x6d3d80) 00:32:36.112 [2024-12-06 17:48:17.723222] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES POWER MANAGEMENT cid:5 cdw10:00000002 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:36.112 [2024-12-06 17:48:17.723243] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73fc00, cid 5, qid 0 00:32:36.112 [2024-12-06 17:48:17.723323] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:36.112 [2024-12-06 17:48:17.723337] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:36.112 [2024-12-06 17:48:17.723344] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.723351] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73fc00) on tqpair=0x6d3d80 00:32:36.112 [2024-12-06 17:48:17.723367] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.723376] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x6d3d80) 00:32:36.112 [2024-12-06 17:48:17.723387] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES TEMPERATURE THRESHOLD cid:5 cdw10:00000004 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:36.112 [2024-12-06 17:48:17.723408] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73fc00, cid 5, qid 0 00:32:36.112 [2024-12-06 17:48:17.723501] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:36.112 [2024-12-06 17:48:17.723513] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:36.112 [2024-12-06 17:48:17.723520] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.723527] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73fc00) on tqpair=0x6d3d80 00:32:36.112 [2024-12-06 17:48:17.723542] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.723551] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x6d3d80) 00:32:36.112 [2024-12-06 17:48:17.723562] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES NUMBER OF QUEUES cid:5 cdw10:00000007 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:36.112 [2024-12-06 17:48:17.723582] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73fc00, cid 5, qid 0 00:32:36.112 [2024-12-06 17:48:17.723672] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:36.112 [2024-12-06 17:48:17.723686] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:36.112 [2024-12-06 17:48:17.723693] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.723700] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73fc00) on tqpair=0x6d3d80 00:32:36.112 [2024-12-06 17:48:17.723726] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.723737] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x6d3d80) 00:32:36.112 [2024-12-06 17:48:17.723748] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:5 nsid:ffffffff cdw10:07ff0001 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:36.112 [2024-12-06 17:48:17.723760] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.723768] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x6d3d80) 00:32:36.112 [2024-12-06 17:48:17.723778] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:4 nsid:ffffffff cdw10:007f0002 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:36.112 [2024-12-06 17:48:17.723789] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.723797] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=6 on tqpair(0x6d3d80) 00:32:36.112 [2024-12-06 17:48:17.723806] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:6 nsid:ffffffff cdw10:007f0003 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:36.112 [2024-12-06 17:48:17.723822] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.723831] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=7 on tqpair(0x6d3d80) 00:32:36.112 [2024-12-06 17:48:17.723840] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:7 nsid:ffffffff cdw10:03ff0005 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:36.112 [2024-12-06 17:48:17.723863] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73fc00, cid 5, qid 0 00:32:36.112 [2024-12-06 17:48:17.723874] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73fa80, cid 4, qid 0 00:32:36.112 [2024-12-06 17:48:17.723882] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73fd80, cid 6, qid 0 00:32:36.112 [2024-12-06 17:48:17.723890] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73ff00, cid 7, qid 0 00:32:36.112 [2024-12-06 17:48:17.724057] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:32:36.112 [2024-12-06 17:48:17.724069] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:32:36.112 [2024-12-06 17:48:17.724076] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.724082] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x6d3d80): datao=0, datal=8192, cccid=5 00:32:36.112 [2024-12-06 17:48:17.724090] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x73fc00) on tqpair(0x6d3d80): expected_datao=0, payload_size=8192 00:32:36.112 [2024-12-06 17:48:17.724098] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.724115] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.724125] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.724137] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:32:36.112 [2024-12-06 17:48:17.724147] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:32:36.112 [2024-12-06 17:48:17.724153] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.724160] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x6d3d80): datao=0, datal=512, cccid=4 00:32:36.112 [2024-12-06 17:48:17.724167] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x73fa80) on tqpair(0x6d3d80): expected_datao=0, payload_size=512 00:32:36.112 [2024-12-06 17:48:17.724174] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.724184] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.724191] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.724199] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:32:36.112 [2024-12-06 17:48:17.724208] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:32:36.112 [2024-12-06 17:48:17.724214] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.724220] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x6d3d80): datao=0, datal=512, cccid=6 00:32:36.112 [2024-12-06 17:48:17.724228] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x73fd80) on tqpair(0x6d3d80): expected_datao=0, payload_size=512 00:32:36.112 [2024-12-06 17:48:17.724235] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.724244] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.724250] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.724259] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:32:36.112 [2024-12-06 17:48:17.724267] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:32:36.112 [2024-12-06 17:48:17.724274] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.724280] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x6d3d80): datao=0, datal=4096, cccid=7 00:32:36.112 [2024-12-06 17:48:17.724287] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x73ff00) on tqpair(0x6d3d80): expected_datao=0, payload_size=4096 00:32:36.112 [2024-12-06 17:48:17.724298] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.724308] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:32:36.112 [2024-12-06 17:48:17.724316] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:32:36.113 [2024-12-06 17:48:17.724324] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:36.113 [2024-12-06 17:48:17.724333] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:36.113 [2024-12-06 17:48:17.724339] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:36.113 [2024-12-06 17:48:17.724346] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73fc00) on tqpair=0x6d3d80 00:32:36.113 [2024-12-06 17:48:17.724378] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:36.113 [2024-12-06 17:48:17.724390] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:36.113 [2024-12-06 17:48:17.724397] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:36.113 [2024-12-06 17:48:17.724403] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73fa80) on tqpair=0x6d3d80 00:32:36.113 [2024-12-06 17:48:17.724417] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:36.113 [2024-12-06 17:48:17.724427] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:36.113 [2024-12-06 17:48:17.724434] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:36.113 [2024-12-06 17:48:17.724440] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73fd80) on tqpair=0x6d3d80 00:32:36.113 [2024-12-06 17:48:17.724450] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:36.113 [2024-12-06 17:48:17.724459] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:36.113 [2024-12-06 17:48:17.724466] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:36.113 [2024-12-06 17:48:17.724472] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73ff00) on tqpair=0x6d3d80 00:32:36.113 ===================================================== 00:32:36.113 NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:32:36.113 ===================================================== 00:32:36.113 Controller Capabilities/Features 00:32:36.113 ================================ 00:32:36.113 Vendor ID: 8086 00:32:36.113 Subsystem Vendor ID: 8086 00:32:36.113 Serial Number: SPDK00000000000001 00:32:36.113 Model Number: SPDK bdev Controller 00:32:36.113 Firmware Version: 25.01 00:32:36.113 Recommended Arb Burst: 6 00:32:36.113 IEEE OUI Identifier: e4 d2 5c 00:32:36.113 Multi-path I/O 00:32:36.113 May have multiple subsystem ports: Yes 00:32:36.113 May have multiple controllers: Yes 00:32:36.113 Associated with SR-IOV VF: No 00:32:36.113 Max Data Transfer Size: 131072 00:32:36.113 Max Number of Namespaces: 32 00:32:36.113 Max Number of I/O Queues: 127 00:32:36.113 NVMe Specification Version (VS): 1.3 00:32:36.113 NVMe Specification Version (Identify): 1.3 00:32:36.113 Maximum Queue Entries: 128 00:32:36.113 Contiguous Queues Required: Yes 00:32:36.113 Arbitration Mechanisms Supported 00:32:36.113 Weighted Round Robin: Not Supported 00:32:36.113 Vendor Specific: Not Supported 00:32:36.113 Reset Timeout: 15000 ms 00:32:36.113 Doorbell Stride: 4 bytes 00:32:36.113 NVM Subsystem Reset: Not Supported 00:32:36.113 Command Sets Supported 00:32:36.113 NVM Command Set: Supported 00:32:36.113 Boot Partition: Not Supported 00:32:36.113 Memory Page Size Minimum: 4096 bytes 00:32:36.113 Memory Page Size Maximum: 4096 bytes 00:32:36.113 Persistent Memory Region: Not Supported 00:32:36.113 Optional Asynchronous Events Supported 00:32:36.113 Namespace Attribute Notices: Supported 00:32:36.113 Firmware Activation Notices: Not Supported 00:32:36.113 ANA Change Notices: Not Supported 00:32:36.113 PLE Aggregate Log Change Notices: Not Supported 00:32:36.113 LBA Status Info Alert Notices: Not Supported 00:32:36.113 EGE Aggregate Log Change Notices: Not Supported 00:32:36.113 Normal NVM Subsystem Shutdown event: Not Supported 00:32:36.113 Zone Descriptor Change Notices: Not Supported 00:32:36.113 Discovery Log Change Notices: Not Supported 00:32:36.113 Controller Attributes 00:32:36.113 128-bit Host Identifier: Supported 00:32:36.113 Non-Operational Permissive Mode: Not Supported 00:32:36.113 NVM Sets: Not Supported 00:32:36.113 Read Recovery Levels: Not Supported 00:32:36.113 Endurance Groups: Not Supported 00:32:36.113 Predictable Latency Mode: Not Supported 00:32:36.113 Traffic Based Keep ALive: Not Supported 00:32:36.113 Namespace Granularity: Not Supported 00:32:36.113 SQ Associations: Not Supported 00:32:36.113 UUID List: Not Supported 00:32:36.113 Multi-Domain Subsystem: Not Supported 00:32:36.113 Fixed Capacity Management: Not Supported 00:32:36.113 Variable Capacity Management: Not Supported 00:32:36.113 Delete Endurance Group: Not Supported 00:32:36.113 Delete NVM Set: Not Supported 00:32:36.113 Extended LBA Formats Supported: Not Supported 00:32:36.113 Flexible Data Placement Supported: Not Supported 00:32:36.113 00:32:36.113 Controller Memory Buffer Support 00:32:36.113 ================================ 00:32:36.113 Supported: No 00:32:36.113 00:32:36.113 Persistent Memory Region Support 00:32:36.113 ================================ 00:32:36.113 Supported: No 00:32:36.113 00:32:36.113 Admin Command Set Attributes 00:32:36.113 ============================ 00:32:36.113 Security Send/Receive: Not Supported 00:32:36.113 Format NVM: Not Supported 00:32:36.113 Firmware Activate/Download: Not Supported 00:32:36.113 Namespace Management: Not Supported 00:32:36.113 Device Self-Test: Not Supported 00:32:36.113 Directives: Not Supported 00:32:36.113 NVMe-MI: Not Supported 00:32:36.113 Virtualization Management: Not Supported 00:32:36.113 Doorbell Buffer Config: Not Supported 00:32:36.113 Get LBA Status Capability: Not Supported 00:32:36.113 Command & Feature Lockdown Capability: Not Supported 00:32:36.113 Abort Command Limit: 4 00:32:36.113 Async Event Request Limit: 4 00:32:36.113 Number of Firmware Slots: N/A 00:32:36.113 Firmware Slot 1 Read-Only: N/A 00:32:36.113 Firmware Activation Without Reset: N/A 00:32:36.113 Multiple Update Detection Support: N/A 00:32:36.113 Firmware Update Granularity: No Information Provided 00:32:36.113 Per-Namespace SMART Log: No 00:32:36.113 Asymmetric Namespace Access Log Page: Not Supported 00:32:36.113 Subsystem NQN: nqn.2016-06.io.spdk:cnode1 00:32:36.113 Command Effects Log Page: Supported 00:32:36.113 Get Log Page Extended Data: Supported 00:32:36.113 Telemetry Log Pages: Not Supported 00:32:36.113 Persistent Event Log Pages: Not Supported 00:32:36.113 Supported Log Pages Log Page: May Support 00:32:36.113 Commands Supported & Effects Log Page: Not Supported 00:32:36.113 Feature Identifiers & Effects Log Page:May Support 00:32:36.113 NVMe-MI Commands & Effects Log Page: May Support 00:32:36.113 Data Area 4 for Telemetry Log: Not Supported 00:32:36.113 Error Log Page Entries Supported: 128 00:32:36.113 Keep Alive: Supported 00:32:36.113 Keep Alive Granularity: 10000 ms 00:32:36.113 00:32:36.113 NVM Command Set Attributes 00:32:36.113 ========================== 00:32:36.113 Submission Queue Entry Size 00:32:36.113 Max: 64 00:32:36.113 Min: 64 00:32:36.113 Completion Queue Entry Size 00:32:36.113 Max: 16 00:32:36.113 Min: 16 00:32:36.113 Number of Namespaces: 32 00:32:36.113 Compare Command: Supported 00:32:36.113 Write Uncorrectable Command: Not Supported 00:32:36.113 Dataset Management Command: Supported 00:32:36.113 Write Zeroes Command: Supported 00:32:36.113 Set Features Save Field: Not Supported 00:32:36.113 Reservations: Supported 00:32:36.113 Timestamp: Not Supported 00:32:36.113 Copy: Supported 00:32:36.113 Volatile Write Cache: Present 00:32:36.113 Atomic Write Unit (Normal): 1 00:32:36.113 Atomic Write Unit (PFail): 1 00:32:36.113 Atomic Compare & Write Unit: 1 00:32:36.113 Fused Compare & Write: Supported 00:32:36.113 Scatter-Gather List 00:32:36.113 SGL Command Set: Supported 00:32:36.113 SGL Keyed: Supported 00:32:36.113 SGL Bit Bucket Descriptor: Not Supported 00:32:36.113 SGL Metadata Pointer: Not Supported 00:32:36.113 Oversized SGL: Not Supported 00:32:36.113 SGL Metadata Address: Not Supported 00:32:36.113 SGL Offset: Supported 00:32:36.113 Transport SGL Data Block: Not Supported 00:32:36.113 Replay Protected Memory Block: Not Supported 00:32:36.113 00:32:36.113 Firmware Slot Information 00:32:36.113 ========================= 00:32:36.113 Active slot: 1 00:32:36.113 Slot 1 Firmware Revision: 25.01 00:32:36.113 00:32:36.113 00:32:36.113 Commands Supported and Effects 00:32:36.113 ============================== 00:32:36.113 Admin Commands 00:32:36.113 -------------- 00:32:36.113 Get Log Page (02h): Supported 00:32:36.113 Identify (06h): Supported 00:32:36.113 Abort (08h): Supported 00:32:36.113 Set Features (09h): Supported 00:32:36.113 Get Features (0Ah): Supported 00:32:36.113 Asynchronous Event Request (0Ch): Supported 00:32:36.113 Keep Alive (18h): Supported 00:32:36.113 I/O Commands 00:32:36.113 ------------ 00:32:36.113 Flush (00h): Supported LBA-Change 00:32:36.113 Write (01h): Supported LBA-Change 00:32:36.113 Read (02h): Supported 00:32:36.113 Compare (05h): Supported 00:32:36.113 Write Zeroes (08h): Supported LBA-Change 00:32:36.113 Dataset Management (09h): Supported LBA-Change 00:32:36.114 Copy (19h): Supported LBA-Change 00:32:36.114 00:32:36.114 Error Log 00:32:36.114 ========= 00:32:36.114 00:32:36.114 Arbitration 00:32:36.114 =========== 00:32:36.114 Arbitration Burst: 1 00:32:36.114 00:32:36.114 Power Management 00:32:36.114 ================ 00:32:36.114 Number of Power States: 1 00:32:36.114 Current Power State: Power State #0 00:32:36.114 Power State #0: 00:32:36.114 Max Power: 0.00 W 00:32:36.114 Non-Operational State: Operational 00:32:36.114 Entry Latency: Not Reported 00:32:36.114 Exit Latency: Not Reported 00:32:36.114 Relative Read Throughput: 0 00:32:36.114 Relative Read Latency: 0 00:32:36.114 Relative Write Throughput: 0 00:32:36.114 Relative Write Latency: 0 00:32:36.114 Idle Power: Not Reported 00:32:36.114 Active Power: Not Reported 00:32:36.114 Non-Operational Permissive Mode: Not Supported 00:32:36.114 00:32:36.114 Health Information 00:32:36.114 ================== 00:32:36.114 Critical Warnings: 00:32:36.114 Available Spare Space: OK 00:32:36.114 Temperature: OK 00:32:36.114 Device Reliability: OK 00:32:36.114 Read Only: No 00:32:36.114 Volatile Memory Backup: OK 00:32:36.114 Current Temperature: 0 Kelvin (-273 Celsius) 00:32:36.114 Temperature Threshold: 0 Kelvin (-273 Celsius) 00:32:36.114 Available Spare: 0% 00:32:36.114 Available Spare Threshold: 0% 00:32:36.114 Life Percentage Used:[2024-12-06 17:48:17.724582] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:36.114 [2024-12-06 17:48:17.724594] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=7 on tqpair(0x6d3d80) 00:32:36.114 [2024-12-06 17:48:17.724605] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ERROR_RECOVERY cid:7 cdw10:00000005 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:36.114 [2024-12-06 17:48:17.724627] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73ff00, cid 7, qid 0 00:32:36.114 [2024-12-06 17:48:17.724759] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:36.114 [2024-12-06 17:48:17.724775] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:36.114 [2024-12-06 17:48:17.724782] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:36.114 [2024-12-06 17:48:17.724789] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73ff00) on tqpair=0x6d3d80 00:32:36.114 [2024-12-06 17:48:17.724836] nvme_ctrlr.c:4399:nvme_ctrlr_destruct_async: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] Prepare to destruct SSD 00:32:36.114 [2024-12-06 17:48:17.724856] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73f480) on tqpair=0x6d3d80 00:32:36.114 [2024-12-06 17:48:17.724867] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:36.114 [2024-12-06 17:48:17.724876] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73f600) on tqpair=0x6d3d80 00:32:36.114 [2024-12-06 17:48:17.724884] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:36.114 [2024-12-06 17:48:17.724892] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73f780) on tqpair=0x6d3d80 00:32:36.114 [2024-12-06 17:48:17.724900] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:36.114 [2024-12-06 17:48:17.724908] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73f900) on tqpair=0x6d3d80 00:32:36.114 [2024-12-06 17:48:17.724916] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:36.114 [2024-12-06 17:48:17.724932] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:36.114 [2024-12-06 17:48:17.724941] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:36.114 [2024-12-06 17:48:17.724947] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x6d3d80) 00:32:36.114 [2024-12-06 17:48:17.724958] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:36.114 [2024-12-06 17:48:17.724981] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73f900, cid 3, qid 0 00:32:36.114 [2024-12-06 17:48:17.725086] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:36.114 [2024-12-06 17:48:17.725098] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:36.114 [2024-12-06 17:48:17.725105] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:36.114 [2024-12-06 17:48:17.725111] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73f900) on tqpair=0x6d3d80 00:32:36.114 [2024-12-06 17:48:17.725122] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:36.114 [2024-12-06 17:48:17.725130] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:36.114 [2024-12-06 17:48:17.725136] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x6d3d80) 00:32:36.114 [2024-12-06 17:48:17.725147] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY SET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:36.114 [2024-12-06 17:48:17.725173] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73f900, cid 3, qid 0 00:32:36.114 [2024-12-06 17:48:17.725263] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:36.114 [2024-12-06 17:48:17.725277] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:36.114 [2024-12-06 17:48:17.725284] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:36.114 [2024-12-06 17:48:17.725290] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73f900) on tqpair=0x6d3d80 00:32:36.114 [2024-12-06 17:48:17.725298] nvme_ctrlr.c:1151:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] RTD3E = 0 us 00:32:36.114 [2024-12-06 17:48:17.725306] nvme_ctrlr.c:1154:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] shutdown timeout = 10000 ms 00:32:36.114 [2024-12-06 17:48:17.725321] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:36.114 [2024-12-06 17:48:17.725331] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:36.114 [2024-12-06 17:48:17.725337] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x6d3d80) 00:32:36.114 [2024-12-06 17:48:17.725348] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:36.114 [2024-12-06 17:48:17.725368] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73f900, cid 3, qid 0 00:32:36.114 [2024-12-06 17:48:17.725445] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:36.114 [2024-12-06 17:48:17.725459] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:36.114 [2024-12-06 17:48:17.725466] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:36.114 [2024-12-06 17:48:17.725472] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73f900) on tqpair=0x6d3d80 00:32:36.114 [2024-12-06 17:48:17.725489] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:36.114 [2024-12-06 17:48:17.725498] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:36.114 [2024-12-06 17:48:17.725505] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x6d3d80) 00:32:36.114 [2024-12-06 17:48:17.725515] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:36.114 [2024-12-06 17:48:17.725536] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73f900, cid 3, qid 0 00:32:36.114 [2024-12-06 17:48:17.725611] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:36.114 [2024-12-06 17:48:17.725624] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:36.114 [2024-12-06 17:48:17.725631] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:36.114 [2024-12-06 17:48:17.725642] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73f900) on tqpair=0x6d3d80 00:32:36.114 [2024-12-06 17:48:17.725659] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:32:36.114 [2024-12-06 17:48:17.729699] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:32:36.114 [2024-12-06 17:48:17.729708] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x6d3d80) 00:32:36.114 [2024-12-06 17:48:17.729718] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:36.114 [2024-12-06 17:48:17.729741] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x73f900, cid 3, qid 0 00:32:36.114 [2024-12-06 17:48:17.729865] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:32:36.114 [2024-12-06 17:48:17.729877] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:32:36.114 [2024-12-06 17:48:17.729884] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:32:36.114 [2024-12-06 17:48:17.729891] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x73f900) on tqpair=0x6d3d80 00:32:36.114 [2024-12-06 17:48:17.729904] nvme_ctrlr.c:1273:nvme_ctrlr_shutdown_poll_async: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] shutdown complete in 4 milliseconds 00:32:36.114 0% 00:32:36.114 Data Units Read: 0 00:32:36.114 Data Units Written: 0 00:32:36.114 Host Read Commands: 0 00:32:36.114 Host Write Commands: 0 00:32:36.114 Controller Busy Time: 0 minutes 00:32:36.114 Power Cycles: 0 00:32:36.114 Power On Hours: 0 hours 00:32:36.114 Unsafe Shutdowns: 0 00:32:36.114 Unrecoverable Media Errors: 0 00:32:36.114 Lifetime Error Log Entries: 0 00:32:36.114 Warning Temperature Time: 0 minutes 00:32:36.114 Critical Temperature Time: 0 minutes 00:32:36.114 00:32:36.114 Number of Queues 00:32:36.114 ================ 00:32:36.114 Number of I/O Submission Queues: 127 00:32:36.114 Number of I/O Completion Queues: 127 00:32:36.114 00:32:36.114 Active Namespaces 00:32:36.114 ================= 00:32:36.114 Namespace ID:1 00:32:36.114 Error Recovery Timeout: Unlimited 00:32:36.114 Command Set Identifier: NVM (00h) 00:32:36.114 Deallocate: Supported 00:32:36.114 Deallocated/Unwritten Error: Not Supported 00:32:36.114 Deallocated Read Value: Unknown 00:32:36.114 Deallocate in Write Zeroes: Not Supported 00:32:36.114 Deallocated Guard Field: 0xFFFF 00:32:36.114 Flush: Supported 00:32:36.114 Reservation: Supported 00:32:36.114 Namespace Sharing Capabilities: Multiple Controllers 00:32:36.114 Size (in LBAs): 131072 (0GiB) 00:32:36.114 Capacity (in LBAs): 131072 (0GiB) 00:32:36.114 Utilization (in LBAs): 131072 (0GiB) 00:32:36.114 NGUID: ABCDEF0123456789ABCDEF0123456789 00:32:36.114 EUI64: ABCDEF0123456789 00:32:36.114 UUID: 2399ee42-e095-4bc5-84c3-ba1500018911 00:32:36.114 Thin Provisioning: Not Supported 00:32:36.114 Per-NS Atomic Units: Yes 00:32:36.114 Atomic Boundary Size (Normal): 0 00:32:36.115 Atomic Boundary Size (PFail): 0 00:32:36.115 Atomic Boundary Offset: 0 00:32:36.115 Maximum Single Source Range Length: 65535 00:32:36.115 Maximum Copy Length: 65535 00:32:36.115 Maximum Source Range Count: 1 00:32:36.115 NGUID/EUI64 Never Reused: No 00:32:36.115 Namespace Write Protected: No 00:32:36.115 Number of LBA Formats: 1 00:32:36.115 Current LBA Format: LBA Format #00 00:32:36.115 LBA Format #00: Data Size: 512 Metadata Size: 0 00:32:36.115 00:32:36.115 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@51 -- # sync 00:32:36.115 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@52 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:32:36.115 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@563 -- # xtrace_disable 00:32:36.115 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:32:36.115 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:32:36.115 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@54 -- # trap - SIGINT SIGTERM EXIT 00:32:36.115 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@56 -- # nvmftestfini 00:32:36.115 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@516 -- # nvmfcleanup 00:32:36.115 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@121 -- # sync 00:32:36.115 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:32:36.115 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@124 -- # set +e 00:32:36.115 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@125 -- # for i in {1..20} 00:32:36.115 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:32:36.115 rmmod nvme_tcp 00:32:36.115 rmmod nvme_fabrics 00:32:36.115 rmmod nvme_keyring 00:32:36.115 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:32:36.115 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@128 -- # set -e 00:32:36.115 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@129 -- # return 0 00:32:36.115 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@517 -- # '[' -n 350729 ']' 00:32:36.115 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@518 -- # killprocess 350729 00:32:36.115 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@954 -- # '[' -z 350729 ']' 00:32:36.115 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@958 -- # kill -0 350729 00:32:36.115 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@959 -- # uname 00:32:36.115 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:32:36.115 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 350729 00:32:36.115 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:32:36.115 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:32:36.115 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@972 -- # echo 'killing process with pid 350729' 00:32:36.115 killing process with pid 350729 00:32:36.115 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@973 -- # kill 350729 00:32:36.115 17:48:17 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@978 -- # wait 350729 00:32:36.375 17:48:18 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:32:36.375 17:48:18 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:32:36.375 17:48:18 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:32:36.375 17:48:18 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@297 -- # iptr 00:32:36.375 17:48:18 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@791 -- # iptables-save 00:32:36.375 17:48:18 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:32:36.375 17:48:18 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@791 -- # iptables-restore 00:32:36.375 17:48:18 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:32:36.375 17:48:18 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@302 -- # remove_spdk_ns 00:32:36.375 17:48:18 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:36.375 17:48:18 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:36.375 17:48:18 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:38.281 17:48:20 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:32:38.281 00:32:38.281 real 0m5.621s 00:32:38.281 user 0m4.845s 00:32:38.281 sys 0m2.007s 00:32:38.281 17:48:20 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1130 -- # xtrace_disable 00:32:38.281 17:48:20 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:32:38.281 ************************************ 00:32:38.281 END TEST nvmf_identify 00:32:38.281 ************************************ 00:32:38.281 17:48:20 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@23 -- # run_test nvmf_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/perf.sh --transport=tcp 00:32:38.281 17:48:20 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:32:38.281 17:48:20 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:32:38.281 17:48:20 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:32:38.539 ************************************ 00:32:38.539 START TEST nvmf_perf 00:32:38.539 ************************************ 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/perf.sh --transport=tcp 00:32:38.539 * Looking for test storage... 00:32:38.539 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1711 -- # lcov --version 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@333 -- # local ver1 ver1_l 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@334 -- # local ver2 ver2_l 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@336 -- # IFS=.-: 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@336 -- # read -ra ver1 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@337 -- # IFS=.-: 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@337 -- # read -ra ver2 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@338 -- # local 'op=<' 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@340 -- # ver1_l=2 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@341 -- # ver2_l=1 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@344 -- # case "$op" in 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@345 -- # : 1 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@364 -- # (( v = 0 )) 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@365 -- # decimal 1 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@353 -- # local d=1 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@355 -- # echo 1 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@365 -- # ver1[v]=1 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@366 -- # decimal 2 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@353 -- # local d=2 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@355 -- # echo 2 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@366 -- # ver2[v]=2 00:32:38.539 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@368 -- # return 0 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:32:38.540 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:38.540 --rc genhtml_branch_coverage=1 00:32:38.540 --rc genhtml_function_coverage=1 00:32:38.540 --rc genhtml_legend=1 00:32:38.540 --rc geninfo_all_blocks=1 00:32:38.540 --rc geninfo_unexecuted_blocks=1 00:32:38.540 00:32:38.540 ' 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:32:38.540 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:38.540 --rc genhtml_branch_coverage=1 00:32:38.540 --rc genhtml_function_coverage=1 00:32:38.540 --rc genhtml_legend=1 00:32:38.540 --rc geninfo_all_blocks=1 00:32:38.540 --rc geninfo_unexecuted_blocks=1 00:32:38.540 00:32:38.540 ' 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:32:38.540 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:38.540 --rc genhtml_branch_coverage=1 00:32:38.540 --rc genhtml_function_coverage=1 00:32:38.540 --rc genhtml_legend=1 00:32:38.540 --rc geninfo_all_blocks=1 00:32:38.540 --rc geninfo_unexecuted_blocks=1 00:32:38.540 00:32:38.540 ' 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:32:38.540 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:38.540 --rc genhtml_branch_coverage=1 00:32:38.540 --rc genhtml_function_coverage=1 00:32:38.540 --rc genhtml_legend=1 00:32:38.540 --rc geninfo_all_blocks=1 00:32:38.540 --rc geninfo_unexecuted_blocks=1 00:32:38.540 00:32:38.540 ' 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@7 -- # uname -s 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@15 -- # shopt -s extglob 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- paths/export.sh@5 -- # export PATH 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@51 -- # : 0 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:32:38.540 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@55 -- # have_pci_nics=0 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@12 -- # MALLOC_BDEV_SIZE=64 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@15 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@17 -- # nvmftestinit 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@476 -- # prepare_net_devs 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@438 -- # local -g is_hw=no 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@440 -- # remove_spdk_ns 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:38.540 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:32:38.541 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:32:38.541 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@309 -- # xtrace_disable 00:32:38.541 17:48:20 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@10 -- # set +x 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@315 -- # pci_devs=() 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@315 -- # local -a pci_devs 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@316 -- # pci_net_devs=() 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@317 -- # pci_drivers=() 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@317 -- # local -A pci_drivers 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@319 -- # net_devs=() 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@319 -- # local -ga net_devs 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@320 -- # e810=() 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@320 -- # local -ga e810 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@321 -- # x722=() 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@321 -- # local -ga x722 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@322 -- # mlx=() 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@322 -- # local -ga mlx 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:32:41.073 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:32:41.073 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@418 -- # [[ up == up ]] 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:32:41.073 Found net devices under 0000:0a:00.0: cvl_0_0 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@418 -- # [[ up == up ]] 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:32:41.073 Found net devices under 0000:0a:00.1: cvl_0_1 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@442 -- # is_hw=yes 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:32:41.073 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:32:41.073 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.254 ms 00:32:41.073 00:32:41.073 --- 10.0.0.2 ping statistics --- 00:32:41.073 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:32:41.073 rtt min/avg/max/mdev = 0.254/0.254/0.254/0.000 ms 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:32:41.073 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:32:41.073 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.102 ms 00:32:41.073 00:32:41.073 --- 10.0.0.1 ping statistics --- 00:32:41.073 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:32:41.073 rtt min/avg/max/mdev = 0.102/0.102/0.102/0.000 ms 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@450 -- # return 0 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:32:41.073 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:32:41.074 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:32:41.074 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:32:41.074 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:32:41.074 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@18 -- # nvmfappstart -m 0xF 00:32:41.074 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:32:41.074 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@726 -- # xtrace_disable 00:32:41.074 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@10 -- # set +x 00:32:41.074 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@509 -- # nvmfpid=352827 00:32:41.074 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:32:41.074 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@510 -- # waitforlisten 352827 00:32:41.074 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@835 -- # '[' -z 352827 ']' 00:32:41.074 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:32:41.074 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@840 -- # local max_retries=100 00:32:41.074 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:32:41.074 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:32:41.074 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@844 -- # xtrace_disable 00:32:41.074 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@10 -- # set +x 00:32:41.074 [2024-12-06 17:48:22.689159] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:32:41.074 [2024-12-06 17:48:22.689258] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:32:41.074 [2024-12-06 17:48:22.760561] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:32:41.074 [2024-12-06 17:48:22.805143] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:32:41.074 [2024-12-06 17:48:22.805200] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:32:41.074 [2024-12-06 17:48:22.805224] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:32:41.074 [2024-12-06 17:48:22.805235] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:32:41.074 [2024-12-06 17:48:22.805244] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:32:41.074 [2024-12-06 17:48:22.806787] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:32:41.074 [2024-12-06 17:48:22.806820] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:32:41.074 [2024-12-06 17:48:22.806880] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:32:41.074 [2024-12-06 17:48:22.806883] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:32:41.332 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:32:41.332 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@868 -- # return 0 00:32:41.332 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:32:41.332 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@732 -- # xtrace_disable 00:32:41.332 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@10 -- # set +x 00:32:41.332 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:32:41.332 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh 00:32:41.332 17:48:22 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py load_subsystem_config 00:32:44.612 17:48:26 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py framework_get_config bdev 00:32:44.612 17:48:26 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@30 -- # jq -r '.[].params | select(.name=="Nvme0").traddr' 00:32:44.612 17:48:26 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@30 -- # local_nvme_trid=0000:88:00.0 00:32:44.612 17:48:26 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:32:44.870 17:48:26 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@31 -- # bdevs=' Malloc0' 00:32:44.870 17:48:26 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@33 -- # '[' -n 0000:88:00.0 ']' 00:32:44.870 17:48:26 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@34 -- # bdevs=' Malloc0 Nvme0n1' 00:32:44.870 17:48:26 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@37 -- # '[' tcp == rdma ']' 00:32:44.870 17:48:26 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:32:45.128 [2024-12-06 17:48:26.914114] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:32:45.128 17:48:26 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:32:45.386 17:48:27 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@45 -- # for bdev in $bdevs 00:32:45.386 17:48:27 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:32:45.643 17:48:27 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@45 -- # for bdev in $bdevs 00:32:45.643 17:48:27 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Nvme0n1 00:32:46.208 17:48:27 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:32:46.209 [2024-12-06 17:48:28.010101] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:32:46.209 17:48:28 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@49 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:32:46.466 17:48:28 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@52 -- # '[' -n 0000:88:00.0 ']' 00:32:46.466 17:48:28 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@53 -- # perf_app -i 0 -q 32 -o 4096 -w randrw -M 50 -t 1 -r 'trtype:PCIe traddr:0000:88:00.0' 00:32:46.466 17:48:28 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@21 -- # '[' 0 -eq 1 ']' 00:32:46.466 17:48:28 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -i 0 -q 32 -o 4096 -w randrw -M 50 -t 1 -r 'trtype:PCIe traddr:0000:88:00.0' 00:32:47.838 Initializing NVMe Controllers 00:32:47.838 Attached to NVMe Controller at 0000:88:00.0 [8086:0a54] 00:32:47.838 Associating PCIE (0000:88:00.0) NSID 1 with lcore 0 00:32:47.838 Initialization complete. Launching workers. 00:32:47.838 ======================================================== 00:32:47.838 Latency(us) 00:32:47.838 Device Information : IOPS MiB/s Average min max 00:32:47.838 PCIE (0000:88:00.0) NSID 1 from core 0: 84589.89 330.43 377.60 35.05 8260.12 00:32:47.838 ======================================================== 00:32:47.838 Total : 84589.89 330.43 377.60 35.05 8260.12 00:32:47.838 00:32:47.838 17:48:29 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 1 -o 4096 -w randrw -M 50 -t 1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:32:49.208 Initializing NVMe Controllers 00:32:49.208 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:32:49.208 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:32:49.208 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:32:49.208 Initialization complete. Launching workers. 00:32:49.208 ======================================================== 00:32:49.208 Latency(us) 00:32:49.208 Device Information : IOPS MiB/s Average min max 00:32:49.208 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 110.00 0.43 9158.73 133.91 46001.37 00:32:49.208 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 51.00 0.20 20536.70 6980.47 48900.74 00:32:49.208 ======================================================== 00:32:49.208 Total : 161.00 0.63 12762.94 133.91 48900.74 00:32:49.208 00:32:49.208 17:48:30 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 32 -o 4096 -w randrw -M 50 -t 1 -HI -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:32:50.579 Initializing NVMe Controllers 00:32:50.579 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:32:50.579 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:32:50.579 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:32:50.579 Initialization complete. Launching workers. 00:32:50.579 ======================================================== 00:32:50.579 Latency(us) 00:32:50.579 Device Information : IOPS MiB/s Average min max 00:32:50.579 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 8518.03 33.27 3757.56 592.92 8448.86 00:32:50.579 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 3814.58 14.90 8389.51 5467.65 16968.21 00:32:50.579 ======================================================== 00:32:50.579 Total : 12332.60 48.17 5190.26 592.92 16968.21 00:32:50.579 00:32:50.579 17:48:32 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@59 -- # [[ e810 == \e\8\1\0 ]] 00:32:50.579 17:48:32 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@59 -- # [[ tcp == \r\d\m\a ]] 00:32:50.579 17:48:32 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 128 -o 262144 -O 16384 -w randrw -M 50 -t 2 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:32:53.105 Initializing NVMe Controllers 00:32:53.105 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:32:53.105 Controller IO queue size 128, less than required. 00:32:53.105 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:32:53.105 Controller IO queue size 128, less than required. 00:32:53.105 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:32:53.105 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:32:53.105 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:32:53.105 Initialization complete. Launching workers. 00:32:53.105 ======================================================== 00:32:53.105 Latency(us) 00:32:53.105 Device Information : IOPS MiB/s Average min max 00:32:53.105 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 1710.97 427.74 75693.38 46751.61 122570.48 00:32:53.105 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 571.99 143.00 231541.08 103841.04 352394.47 00:32:53.105 ======================================================== 00:32:53.105 Total : 2282.95 570.74 114740.63 46751.61 352394.47 00:32:53.105 00:32:53.105 17:48:34 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 128 -o 36964 -O 4096 -w randrw -M 50 -t 5 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -c 0xf -P 4 00:32:53.372 No valid NVMe controllers or AIO or URING devices found 00:32:53.372 Initializing NVMe Controllers 00:32:53.372 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:32:53.372 Controller IO queue size 128, less than required. 00:32:53.372 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:32:53.372 WARNING: IO size 36964 (-o) is not a multiple of nsid 1 sector size 512. Removing this ns from test 00:32:53.372 Controller IO queue size 128, less than required. 00:32:53.372 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:32:53.372 WARNING: IO size 36964 (-o) is not a multiple of nsid 2 sector size 512. Removing this ns from test 00:32:53.372 WARNING: Some requested NVMe devices were skipped 00:32:53.372 17:48:35 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@65 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 128 -o 262144 -w randrw -M 50 -t 2 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' --transport-stat 00:32:55.905 Initializing NVMe Controllers 00:32:55.905 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:32:55.905 Controller IO queue size 128, less than required. 00:32:55.905 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:32:55.905 Controller IO queue size 128, less than required. 00:32:55.905 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:32:55.905 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:32:55.905 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:32:55.905 Initialization complete. Launching workers. 00:32:55.905 00:32:55.905 ==================== 00:32:55.905 lcore 0, ns TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 statistics: 00:32:55.905 TCP transport: 00:32:55.905 polls: 9787 00:32:55.905 idle_polls: 6573 00:32:55.905 sock_completions: 3214 00:32:55.905 nvme_completions: 6103 00:32:55.905 submitted_requests: 9310 00:32:55.905 queued_requests: 1 00:32:55.905 00:32:55.905 ==================== 00:32:55.905 lcore 0, ns TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 statistics: 00:32:55.905 TCP transport: 00:32:55.905 polls: 13224 00:32:55.905 idle_polls: 9676 00:32:55.905 sock_completions: 3548 00:32:55.905 nvme_completions: 6191 00:32:55.905 submitted_requests: 9284 00:32:55.905 queued_requests: 1 00:32:55.905 ======================================================== 00:32:55.905 Latency(us) 00:32:55.905 Device Information : IOPS MiB/s Average min max 00:32:55.905 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 1524.63 381.16 85646.65 62134.26 129296.11 00:32:55.905 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 1546.61 386.65 84205.55 41934.95 132141.43 00:32:55.905 ======================================================== 00:32:55.905 Total : 3071.24 767.81 84920.94 41934.95 132141.43 00:32:55.905 00:32:55.906 17:48:37 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@66 -- # sync 00:32:55.906 17:48:37 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@67 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:32:56.163 17:48:37 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@69 -- # '[' 1 -eq 1 ']' 00:32:56.163 17:48:37 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@71 -- # '[' -n 0000:88:00.0 ']' 00:32:56.163 17:48:37 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore Nvme0n1 lvs_0 00:32:59.441 17:48:41 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@72 -- # ls_guid=77281710-5e44-4911-ac44-045e42f0fbad 00:32:59.441 17:48:41 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@73 -- # get_lvs_free_mb 77281710-5e44-4911-ac44-045e42f0fbad 00:32:59.441 17:48:41 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1368 -- # local lvs_uuid=77281710-5e44-4911-ac44-045e42f0fbad 00:32:59.441 17:48:41 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1369 -- # local lvs_info 00:32:59.441 17:48:41 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1370 -- # local fc 00:32:59.441 17:48:41 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1371 -- # local cs 00:32:59.441 17:48:41 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1372 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores 00:32:59.698 17:48:41 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1372 -- # lvs_info='[ 00:32:59.698 { 00:32:59.698 "uuid": "77281710-5e44-4911-ac44-045e42f0fbad", 00:32:59.698 "name": "lvs_0", 00:32:59.698 "base_bdev": "Nvme0n1", 00:32:59.698 "total_data_clusters": 238234, 00:32:59.698 "free_clusters": 238234, 00:32:59.698 "block_size": 512, 00:32:59.698 "cluster_size": 4194304 00:32:59.698 } 00:32:59.698 ]' 00:32:59.698 17:48:41 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1373 -- # jq '.[] | select(.uuid=="77281710-5e44-4911-ac44-045e42f0fbad") .free_clusters' 00:32:59.698 17:48:41 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1373 -- # fc=238234 00:32:59.699 17:48:41 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1374 -- # jq '.[] | select(.uuid=="77281710-5e44-4911-ac44-045e42f0fbad") .cluster_size' 00:32:59.699 17:48:41 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1374 -- # cs=4194304 00:32:59.699 17:48:41 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1377 -- # free_mb=952936 00:32:59.699 17:48:41 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1378 -- # echo 952936 00:32:59.699 952936 00:32:59.699 17:48:41 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@77 -- # '[' 952936 -gt 20480 ']' 00:32:59.699 17:48:41 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@78 -- # free_mb=20480 00:32:59.699 17:48:41 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@80 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u 77281710-5e44-4911-ac44-045e42f0fbad lbd_0 20480 00:33:00.284 17:48:42 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@80 -- # lb_guid=d76ed0e6-618e-4f7d-a39d-647f5ec3d2f4 00:33:00.284 17:48:42 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore d76ed0e6-618e-4f7d-a39d-647f5ec3d2f4 lvs_n_0 00:33:01.218 17:48:42 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@83 -- # ls_nested_guid=7119c0d6-ead7-4b05-ac98-5d6edc3262ae 00:33:01.218 17:48:42 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@84 -- # get_lvs_free_mb 7119c0d6-ead7-4b05-ac98-5d6edc3262ae 00:33:01.218 17:48:42 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1368 -- # local lvs_uuid=7119c0d6-ead7-4b05-ac98-5d6edc3262ae 00:33:01.218 17:48:42 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1369 -- # local lvs_info 00:33:01.218 17:48:42 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1370 -- # local fc 00:33:01.218 17:48:42 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1371 -- # local cs 00:33:01.218 17:48:42 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1372 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores 00:33:01.476 17:48:43 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1372 -- # lvs_info='[ 00:33:01.476 { 00:33:01.476 "uuid": "77281710-5e44-4911-ac44-045e42f0fbad", 00:33:01.476 "name": "lvs_0", 00:33:01.476 "base_bdev": "Nvme0n1", 00:33:01.476 "total_data_clusters": 238234, 00:33:01.476 "free_clusters": 233114, 00:33:01.476 "block_size": 512, 00:33:01.476 "cluster_size": 4194304 00:33:01.476 }, 00:33:01.476 { 00:33:01.476 "uuid": "7119c0d6-ead7-4b05-ac98-5d6edc3262ae", 00:33:01.476 "name": "lvs_n_0", 00:33:01.476 "base_bdev": "d76ed0e6-618e-4f7d-a39d-647f5ec3d2f4", 00:33:01.476 "total_data_clusters": 5114, 00:33:01.476 "free_clusters": 5114, 00:33:01.476 "block_size": 512, 00:33:01.476 "cluster_size": 4194304 00:33:01.476 } 00:33:01.476 ]' 00:33:01.476 17:48:43 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1373 -- # jq '.[] | select(.uuid=="7119c0d6-ead7-4b05-ac98-5d6edc3262ae") .free_clusters' 00:33:01.476 17:48:43 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1373 -- # fc=5114 00:33:01.476 17:48:43 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1374 -- # jq '.[] | select(.uuid=="7119c0d6-ead7-4b05-ac98-5d6edc3262ae") .cluster_size' 00:33:01.476 17:48:43 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1374 -- # cs=4194304 00:33:01.476 17:48:43 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1377 -- # free_mb=20456 00:33:01.476 17:48:43 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1378 -- # echo 20456 00:33:01.476 20456 00:33:01.476 17:48:43 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@85 -- # '[' 20456 -gt 20480 ']' 00:33:01.476 17:48:43 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u 7119c0d6-ead7-4b05-ac98-5d6edc3262ae lbd_nest_0 20456 00:33:01.734 17:48:43 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@88 -- # lb_nested_guid=2be05395-8c2a-4149-86d1-22330a89c2eb 00:33:01.734 17:48:43 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:33:01.991 17:48:43 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@90 -- # for bdev in $lb_nested_guid 00:33:01.991 17:48:43 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@91 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 2be05395-8c2a-4149-86d1-22330a89c2eb 00:33:02.249 17:48:44 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@93 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:33:02.507 17:48:44 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@95 -- # qd_depth=("1" "32" "128") 00:33:02.507 17:48:44 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@96 -- # io_size=("512" "131072") 00:33:02.507 17:48:44 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@97 -- # for qd in "${qd_depth[@]}" 00:33:02.507 17:48:44 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@98 -- # for o in "${io_size[@]}" 00:33:02.507 17:48:44 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@99 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 1 -o 512 -w randrw -M 50 -t 10 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:33:14.691 Initializing NVMe Controllers 00:33:14.691 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:33:14.691 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:33:14.691 Initialization complete. Launching workers. 00:33:14.691 ======================================================== 00:33:14.691 Latency(us) 00:33:14.691 Device Information : IOPS MiB/s Average min max 00:33:14.691 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 45.90 0.02 21797.97 159.99 45827.44 00:33:14.691 ======================================================== 00:33:14.691 Total : 45.90 0.02 21797.97 159.99 45827.44 00:33:14.691 00:33:14.691 17:48:54 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@98 -- # for o in "${io_size[@]}" 00:33:14.691 17:48:54 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@99 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 1 -o 131072 -w randrw -M 50 -t 10 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:33:24.649 Initializing NVMe Controllers 00:33:24.649 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:33:24.649 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:33:24.649 Initialization complete. Launching workers. 00:33:24.649 ======================================================== 00:33:24.649 Latency(us) 00:33:24.649 Device Information : IOPS MiB/s Average min max 00:33:24.649 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 74.90 9.36 13359.30 5991.19 48877.54 00:33:24.649 ======================================================== 00:33:24.649 Total : 74.90 9.36 13359.30 5991.19 48877.54 00:33:24.649 00:33:24.649 17:49:05 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@97 -- # for qd in "${qd_depth[@]}" 00:33:24.649 17:49:05 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@98 -- # for o in "${io_size[@]}" 00:33:24.649 17:49:05 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@99 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 32 -o 512 -w randrw -M 50 -t 10 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:33:34.608 Initializing NVMe Controllers 00:33:34.608 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:33:34.608 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:33:34.608 Initialization complete. Launching workers. 00:33:34.608 ======================================================== 00:33:34.608 Latency(us) 00:33:34.608 Device Information : IOPS MiB/s Average min max 00:33:34.608 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 7702.96 3.76 4163.08 268.59 43107.48 00:33:34.608 ======================================================== 00:33:34.608 Total : 7702.96 3.76 4163.08 268.59 43107.48 00:33:34.608 00:33:34.608 17:49:15 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@98 -- # for o in "${io_size[@]}" 00:33:34.608 17:49:15 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@99 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 32 -o 131072 -w randrw -M 50 -t 10 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:33:44.569 Initializing NVMe Controllers 00:33:44.570 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:33:44.570 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:33:44.570 Initialization complete. Launching workers. 00:33:44.570 ======================================================== 00:33:44.570 Latency(us) 00:33:44.570 Device Information : IOPS MiB/s Average min max 00:33:44.570 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 3562.05 445.26 8987.37 1412.30 18642.04 00:33:44.570 ======================================================== 00:33:44.570 Total : 3562.05 445.26 8987.37 1412.30 18642.04 00:33:44.570 00:33:44.570 17:49:25 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@97 -- # for qd in "${qd_depth[@]}" 00:33:44.570 17:49:25 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@98 -- # for o in "${io_size[@]}" 00:33:44.570 17:49:25 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@99 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 128 -o 512 -w randrw -M 50 -t 10 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:33:54.528 Initializing NVMe Controllers 00:33:54.528 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:33:54.528 Controller IO queue size 128, less than required. 00:33:54.528 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:33:54.528 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:33:54.528 Initialization complete. Launching workers. 00:33:54.528 ======================================================== 00:33:54.528 Latency(us) 00:33:54.528 Device Information : IOPS MiB/s Average min max 00:33:54.528 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 11937.46 5.83 10729.66 1531.66 24821.64 00:33:54.528 ======================================================== 00:33:54.528 Total : 11937.46 5.83 10729.66 1531.66 24821.64 00:33:54.528 00:33:54.528 17:49:36 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@98 -- # for o in "${io_size[@]}" 00:33:54.528 17:49:36 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@99 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 128 -o 131072 -w randrw -M 50 -t 10 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:34:06.716 Initializing NVMe Controllers 00:34:06.716 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:34:06.716 Controller IO queue size 128, less than required. 00:34:06.716 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:34:06.716 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:34:06.716 Initialization complete. Launching workers. 00:34:06.716 ======================================================== 00:34:06.716 Latency(us) 00:34:06.716 Device Information : IOPS MiB/s Average min max 00:34:06.716 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 1203.90 150.49 106680.86 16115.75 215949.43 00:34:06.716 ======================================================== 00:34:06.717 Total : 1203.90 150.49 106680.86 16115.75 215949.43 00:34:06.717 00:34:06.717 17:49:46 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@104 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:34:06.717 17:49:46 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@105 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete 2be05395-8c2a-4149-86d1-22330a89c2eb 00:34:06.717 17:49:47 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@106 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -l lvs_n_0 00:34:06.717 17:49:47 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@107 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete d76ed0e6-618e-4f7d-a39d-647f5ec3d2f4 00:34:06.717 17:49:48 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@108 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -l lvs_0 00:34:06.717 17:49:48 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@112 -- # trap - SIGINT SIGTERM EXIT 00:34:06.717 17:49:48 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@114 -- # nvmftestfini 00:34:06.717 17:49:48 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@516 -- # nvmfcleanup 00:34:06.717 17:49:48 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@121 -- # sync 00:34:06.717 17:49:48 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:34:06.717 17:49:48 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@124 -- # set +e 00:34:06.717 17:49:48 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@125 -- # for i in {1..20} 00:34:06.717 17:49:48 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:34:06.717 rmmod nvme_tcp 00:34:06.717 rmmod nvme_fabrics 00:34:06.975 rmmod nvme_keyring 00:34:06.975 17:49:48 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:34:06.975 17:49:48 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@128 -- # set -e 00:34:06.975 17:49:48 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@129 -- # return 0 00:34:06.975 17:49:48 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@517 -- # '[' -n 352827 ']' 00:34:06.975 17:49:48 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@518 -- # killprocess 352827 00:34:06.975 17:49:48 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@954 -- # '[' -z 352827 ']' 00:34:06.975 17:49:48 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@958 -- # kill -0 352827 00:34:06.975 17:49:48 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@959 -- # uname 00:34:06.975 17:49:48 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:34:06.975 17:49:48 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 352827 00:34:06.975 17:49:48 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:34:06.975 17:49:48 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:34:06.975 17:49:48 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@972 -- # echo 'killing process with pid 352827' 00:34:06.975 killing process with pid 352827 00:34:06.975 17:49:48 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@973 -- # kill 352827 00:34:06.975 17:49:48 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@978 -- # wait 352827 00:34:08.877 17:49:50 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:34:08.877 17:49:50 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:34:08.877 17:49:50 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:34:08.877 17:49:50 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@297 -- # iptr 00:34:08.877 17:49:50 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@791 -- # iptables-save 00:34:08.877 17:49:50 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:34:08.877 17:49:50 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@791 -- # iptables-restore 00:34:08.877 17:49:50 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:34:08.878 17:49:50 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@302 -- # remove_spdk_ns 00:34:08.878 17:49:50 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:34:08.878 17:49:50 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:34:08.878 17:49:50 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:34:10.783 17:49:52 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:34:10.783 00:34:10.783 real 1m32.139s 00:34:10.783 user 5m39.090s 00:34:10.783 sys 0m15.756s 00:34:10.783 17:49:52 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:34:10.783 17:49:52 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@10 -- # set +x 00:34:10.783 ************************************ 00:34:10.783 END TEST nvmf_perf 00:34:10.783 ************************************ 00:34:10.783 17:49:52 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@24 -- # run_test nvmf_fio_host /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/fio.sh --transport=tcp 00:34:10.783 17:49:52 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:34:10.783 17:49:52 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:34:10.783 17:49:52 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:34:10.783 ************************************ 00:34:10.783 START TEST nvmf_fio_host 00:34:10.783 ************************************ 00:34:10.783 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/fio.sh --transport=tcp 00:34:10.783 * Looking for test storage... 00:34:10.784 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1711 -- # lcov --version 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@333 -- # local ver1 ver1_l 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@334 -- # local ver2 ver2_l 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@336 -- # IFS=.-: 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@336 -- # read -ra ver1 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@337 -- # IFS=.-: 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@337 -- # read -ra ver2 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@338 -- # local 'op=<' 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@340 -- # ver1_l=2 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@341 -- # ver2_l=1 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@344 -- # case "$op" in 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@345 -- # : 1 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@364 -- # (( v = 0 )) 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@365 -- # decimal 1 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@353 -- # local d=1 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@355 -- # echo 1 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@365 -- # ver1[v]=1 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@366 -- # decimal 2 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@353 -- # local d=2 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@355 -- # echo 2 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@366 -- # ver2[v]=2 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@368 -- # return 0 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:34:10.784 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:10.784 --rc genhtml_branch_coverage=1 00:34:10.784 --rc genhtml_function_coverage=1 00:34:10.784 --rc genhtml_legend=1 00:34:10.784 --rc geninfo_all_blocks=1 00:34:10.784 --rc geninfo_unexecuted_blocks=1 00:34:10.784 00:34:10.784 ' 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:34:10.784 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:10.784 --rc genhtml_branch_coverage=1 00:34:10.784 --rc genhtml_function_coverage=1 00:34:10.784 --rc genhtml_legend=1 00:34:10.784 --rc geninfo_all_blocks=1 00:34:10.784 --rc geninfo_unexecuted_blocks=1 00:34:10.784 00:34:10.784 ' 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:34:10.784 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:10.784 --rc genhtml_branch_coverage=1 00:34:10.784 --rc genhtml_function_coverage=1 00:34:10.784 --rc genhtml_legend=1 00:34:10.784 --rc geninfo_all_blocks=1 00:34:10.784 --rc geninfo_unexecuted_blocks=1 00:34:10.784 00:34:10.784 ' 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:34:10.784 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:10.784 --rc genhtml_branch_coverage=1 00:34:10.784 --rc genhtml_function_coverage=1 00:34:10.784 --rc genhtml_legend=1 00:34:10.784 --rc geninfo_all_blocks=1 00:34:10.784 --rc geninfo_unexecuted_blocks=1 00:34:10.784 00:34:10.784 ' 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@15 -- # shopt -s extglob 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@5 -- # export PATH 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@7 -- # uname -s 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@15 -- # shopt -s extglob 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:10.784 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:10.785 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:10.785 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@5 -- # export PATH 00:34:10.785 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:10.785 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@51 -- # : 0 00:34:10.785 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:34:10.785 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:34:10.785 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:34:10.785 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:34:10.785 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:34:10.785 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:34:10.785 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:34:10.785 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:34:10.785 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:34:10.785 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@55 -- # have_pci_nics=0 00:34:10.785 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@12 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:34:10.785 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@14 -- # nvmftestinit 00:34:10.785 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:34:10.785 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:34:10.785 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@476 -- # prepare_net_devs 00:34:10.785 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@438 -- # local -g is_hw=no 00:34:10.785 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@440 -- # remove_spdk_ns 00:34:10.785 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:34:10.785 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:34:10.785 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:34:10.785 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:34:10.785 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:34:10.785 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@309 -- # xtrace_disable 00:34:10.785 17:49:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@10 -- # set +x 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@315 -- # pci_devs=() 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@315 -- # local -a pci_devs 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@316 -- # pci_net_devs=() 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@317 -- # pci_drivers=() 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@317 -- # local -A pci_drivers 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@319 -- # net_devs=() 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@319 -- # local -ga net_devs 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@320 -- # e810=() 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@320 -- # local -ga e810 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@321 -- # x722=() 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@321 -- # local -ga x722 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@322 -- # mlx=() 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@322 -- # local -ga mlx 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:34:13.320 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:34:13.320 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@418 -- # [[ up == up ]] 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:34:13.320 Found net devices under 0000:0a:00.0: cvl_0_0 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@418 -- # [[ up == up ]] 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:34:13.320 Found net devices under 0000:0a:00.1: cvl_0_1 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@442 -- # is_hw=yes 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:34:13.320 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:34:13.321 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:34:13.321 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:34:13.321 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:34:13.321 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.266 ms 00:34:13.321 00:34:13.321 --- 10.0.0.2 ping statistics --- 00:34:13.321 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:34:13.321 rtt min/avg/max/mdev = 0.266/0.266/0.266/0.000 ms 00:34:13.321 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:34:13.321 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:34:13.321 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.152 ms 00:34:13.321 00:34:13.321 --- 10.0.0.1 ping statistics --- 00:34:13.321 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:34:13.321 rtt min/avg/max/mdev = 0.152/0.152/0.152/0.000 ms 00:34:13.321 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:34:13.321 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@450 -- # return 0 00:34:13.321 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:34:13.321 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:34:13.321 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:34:13.321 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:34:13.321 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:34:13.321 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:34:13.321 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:34:13.321 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@16 -- # [[ y != y ]] 00:34:13.321 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@21 -- # timing_enter start_nvmf_tgt 00:34:13.321 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@726 -- # xtrace_disable 00:34:13.321 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@10 -- # set +x 00:34:13.321 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@24 -- # nvmfpid=364938 00:34:13.321 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@23 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:34:13.321 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@26 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:34:13.321 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@28 -- # waitforlisten 364938 00:34:13.321 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@835 -- # '[' -z 364938 ']' 00:34:13.321 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:34:13.321 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@840 -- # local max_retries=100 00:34:13.321 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:34:13.321 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:34:13.321 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@844 -- # xtrace_disable 00:34:13.321 17:49:54 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@10 -- # set +x 00:34:13.321 [2024-12-06 17:49:54.784721] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:34:13.321 [2024-12-06 17:49:54.784806] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:34:13.321 [2024-12-06 17:49:54.863030] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:34:13.321 [2024-12-06 17:49:54.910906] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:34:13.321 [2024-12-06 17:49:54.910972] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:34:13.321 [2024-12-06 17:49:54.910990] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:34:13.321 [2024-12-06 17:49:54.911001] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:34:13.321 [2024-12-06 17:49:54.911024] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:34:13.321 [2024-12-06 17:49:54.912569] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:34:13.321 [2024-12-06 17:49:54.912623] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:34:13.321 [2024-12-06 17:49:54.912694] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:34:13.321 [2024-12-06 17:49:54.912697] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:34:13.321 17:49:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:34:13.321 17:49:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@868 -- # return 0 00:34:13.321 17:49:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:34:13.578 [2024-12-06 17:49:55.270930] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:34:13.578 17:49:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@30 -- # timing_exit start_nvmf_tgt 00:34:13.578 17:49:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@732 -- # xtrace_disable 00:34:13.578 17:49:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@10 -- # set +x 00:34:13.578 17:49:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc1 00:34:13.836 Malloc1 00:34:13.836 17:49:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:34:14.094 17:49:55 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:34:14.352 17:49:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:34:14.610 [2024-12-06 17:49:56.398776] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:34:14.610 17:49:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:34:14.879 17:49:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@38 -- # PLUGIN_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme 00:34:14.879 17:49:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@41 -- # fio_nvme /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/example_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' --bs=4096 00:34:14.879 17:49:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1364 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/example_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' --bs=4096 00:34:14.879 17:49:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1341 -- # local fio_dir=/usr/src/fio 00:34:14.879 17:49:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1343 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:34:14.879 17:49:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1343 -- # local sanitizers 00:34:14.879 17:49:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1344 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:34:14.879 17:49:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # shift 00:34:14.879 17:49:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1347 -- # local asan_lib= 00:34:14.879 17:49:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:34:14.879 17:49:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:34:14.879 17:49:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # grep libasan 00:34:14.879 17:49:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:34:14.879 17:49:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # asan_lib= 00:34:14.879 17:49:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:34:14.879 17:49:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:34:14.879 17:49:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:34:14.879 17:49:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # grep libclang_rt.asan 00:34:14.879 17:49:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:34:14.879 17:49:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # asan_lib= 00:34:14.879 17:49:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:34:14.879 17:49:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1356 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme' 00:34:14.879 17:49:56 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1356 -- # /usr/src/fio/fio /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/example_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' --bs=4096 00:34:15.136 test: (g=0): rw=randrw, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk, iodepth=128 00:34:15.136 fio-3.35 00:34:15.136 Starting 1 thread 00:34:17.661 00:34:17.661 test: (groupid=0, jobs=1): err= 0: pid=365410: Fri Dec 6 17:49:59 2024 00:34:17.661 read: IOPS=7823, BW=30.6MiB/s (32.0MB/s)(61.3MiB/2007msec) 00:34:17.661 slat (nsec): min=1947, max=285004, avg=2586.54, stdev=3261.52 00:34:17.661 clat (usec): min=3681, max=14557, avg=8935.00, stdev=736.88 00:34:17.661 lat (usec): min=3732, max=14559, avg=8937.59, stdev=736.78 00:34:17.661 clat percentiles (usec): 00:34:17.661 | 1.00th=[ 7242], 5.00th=[ 7767], 10.00th=[ 8029], 20.00th=[ 8356], 00:34:17.661 | 30.00th=[ 8586], 40.00th=[ 8717], 50.00th=[ 8979], 60.00th=[ 9110], 00:34:17.661 | 70.00th=[ 9241], 80.00th=[ 9503], 90.00th=[ 9896], 95.00th=[10028], 00:34:17.661 | 99.00th=[10683], 99.50th=[10814], 99.90th=[12387], 99.95th=[14222], 00:34:17.661 | 99.99th=[14484] 00:34:17.661 bw ( KiB/s): min=29768, max=31952, per=99.85%, avg=31244.00, stdev=999.68, samples=4 00:34:17.661 iops : min= 7442, max= 7988, avg=7811.00, stdev=249.92, samples=4 00:34:17.661 write: IOPS=7794, BW=30.4MiB/s (31.9MB/s)(61.1MiB/2007msec); 0 zone resets 00:34:17.661 slat (usec): min=2, max=282, avg= 2.70, stdev= 2.70 00:34:17.661 clat (usec): min=2624, max=14320, avg=7392.19, stdev=664.55 00:34:17.661 lat (usec): min=2640, max=14322, avg=7394.90, stdev=664.69 00:34:17.661 clat percentiles (usec): 00:34:17.661 | 1.00th=[ 5997], 5.00th=[ 6456], 10.00th=[ 6652], 20.00th=[ 6849], 00:34:17.661 | 30.00th=[ 7046], 40.00th=[ 7242], 50.00th=[ 7373], 60.00th=[ 7504], 00:34:17.661 | 70.00th=[ 7701], 80.00th=[ 7898], 90.00th=[ 8160], 95.00th=[ 8455], 00:34:17.661 | 99.00th=[ 9372], 99.50th=[ 9896], 99.90th=[10552], 99.95th=[12387], 00:34:17.661 | 99.99th=[14222] 00:34:17.661 bw ( KiB/s): min=30832, max=31432, per=100.00%, avg=31192.00, stdev=255.08, samples=4 00:34:17.661 iops : min= 7708, max= 7858, avg=7798.00, stdev=63.77, samples=4 00:34:17.662 lat (msec) : 4=0.07%, 10=96.19%, 20=3.74% 00:34:17.662 cpu : usr=63.26%, sys=35.29%, ctx=83, majf=0, minf=35 00:34:17.662 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.1%, >=64=99.8% 00:34:17.662 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:34:17.662 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:34:17.662 issued rwts: total=15701,15644,0,0 short=0,0,0,0 dropped=0,0,0,0 00:34:17.662 latency : target=0, window=0, percentile=100.00%, depth=128 00:34:17.662 00:34:17.662 Run status group 0 (all jobs): 00:34:17.662 READ: bw=30.6MiB/s (32.0MB/s), 30.6MiB/s-30.6MiB/s (32.0MB/s-32.0MB/s), io=61.3MiB (64.3MB), run=2007-2007msec 00:34:17.662 WRITE: bw=30.4MiB/s (31.9MB/s), 30.4MiB/s-30.4MiB/s (31.9MB/s-31.9MB/s), io=61.1MiB (64.1MB), run=2007-2007msec 00:34:17.662 17:49:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@45 -- # fio_nvme /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/mock_sgl_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' 00:34:17.662 17:49:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1364 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/mock_sgl_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' 00:34:17.662 17:49:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1341 -- # local fio_dir=/usr/src/fio 00:34:17.662 17:49:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1343 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:34:17.662 17:49:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1343 -- # local sanitizers 00:34:17.662 17:49:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1344 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:34:17.662 17:49:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # shift 00:34:17.662 17:49:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1347 -- # local asan_lib= 00:34:17.662 17:49:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:34:17.662 17:49:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:34:17.662 17:49:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # grep libasan 00:34:17.662 17:49:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:34:17.662 17:49:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # asan_lib= 00:34:17.662 17:49:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:34:17.662 17:49:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:34:17.662 17:49:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:34:17.662 17:49:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # grep libclang_rt.asan 00:34:17.662 17:49:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:34:17.662 17:49:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # asan_lib= 00:34:17.662 17:49:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:34:17.662 17:49:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1356 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme' 00:34:17.662 17:49:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1356 -- # /usr/src/fio/fio /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/mock_sgl_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' 00:34:17.919 test: (g=0): rw=randrw, bs=(R) 16.0KiB-16.0KiB, (W) 16.0KiB-16.0KiB, (T) 16.0KiB-16.0KiB, ioengine=spdk, iodepth=128 00:34:17.919 fio-3.35 00:34:17.919 Starting 1 thread 00:34:20.445 00:34:20.445 test: (groupid=0, jobs=1): err= 0: pid=365745: Fri Dec 6 17:50:01 2024 00:34:20.445 read: IOPS=8227, BW=129MiB/s (135MB/s)(258MiB/2006msec) 00:34:20.445 slat (nsec): min=2798, max=96580, avg=3656.20, stdev=1720.74 00:34:20.445 clat (usec): min=1824, max=15790, avg=8708.63, stdev=1983.38 00:34:20.445 lat (usec): min=1828, max=15794, avg=8712.29, stdev=1983.41 00:34:20.445 clat percentiles (usec): 00:34:20.445 | 1.00th=[ 4555], 5.00th=[ 5538], 10.00th=[ 6194], 20.00th=[ 7046], 00:34:20.445 | 30.00th=[ 7570], 40.00th=[ 8160], 50.00th=[ 8717], 60.00th=[ 9241], 00:34:20.445 | 70.00th=[ 9765], 80.00th=[10290], 90.00th=[11207], 95.00th=[11994], 00:34:20.445 | 99.00th=[14091], 99.50th=[14615], 99.90th=[15270], 99.95th=[15533], 00:34:20.445 | 99.99th=[15664] 00:34:20.445 bw ( KiB/s): min=58080, max=79808, per=52.93%, avg=69680.00, stdev=9627.64, samples=4 00:34:20.445 iops : min= 3630, max= 4988, avg=4355.00, stdev=601.73, samples=4 00:34:20.445 write: IOPS=4895, BW=76.5MiB/s (80.2MB/s)(142MiB/1859msec); 0 zone resets 00:34:20.445 slat (usec): min=30, max=126, avg=33.84, stdev= 5.37 00:34:20.445 clat (usec): min=5340, max=17920, avg=11549.54, stdev=1911.26 00:34:20.445 lat (usec): min=5374, max=17951, avg=11583.38, stdev=1911.45 00:34:20.445 clat percentiles (usec): 00:34:20.445 | 1.00th=[ 7504], 5.00th=[ 8717], 10.00th=[ 9110], 20.00th=[ 9896], 00:34:20.445 | 30.00th=[10421], 40.00th=[10945], 50.00th=[11469], 60.00th=[11994], 00:34:20.445 | 70.00th=[12518], 80.00th=[13304], 90.00th=[14091], 95.00th=[14746], 00:34:20.445 | 99.00th=[16057], 99.50th=[16319], 99.90th=[17695], 99.95th=[17957], 00:34:20.445 | 99.99th=[17957] 00:34:20.445 bw ( KiB/s): min=59712, max=82208, per=92.17%, avg=72192.00, stdev=9963.71, samples=4 00:34:20.445 iops : min= 3732, max= 5138, avg=4512.00, stdev=622.73, samples=4 00:34:20.445 lat (msec) : 2=0.02%, 4=0.21%, 10=55.98%, 20=43.79% 00:34:20.445 cpu : usr=77.76%, sys=21.00%, ctx=53, majf=0, minf=61 00:34:20.445 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.2%, 16=0.3%, 32=0.6%, >=64=98.8% 00:34:20.445 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:34:20.445 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:34:20.445 issued rwts: total=16504,9100,0,0 short=0,0,0,0 dropped=0,0,0,0 00:34:20.445 latency : target=0, window=0, percentile=100.00%, depth=128 00:34:20.445 00:34:20.445 Run status group 0 (all jobs): 00:34:20.445 READ: bw=129MiB/s (135MB/s), 129MiB/s-129MiB/s (135MB/s-135MB/s), io=258MiB (270MB), run=2006-2006msec 00:34:20.445 WRITE: bw=76.5MiB/s (80.2MB/s), 76.5MiB/s-76.5MiB/s (80.2MB/s-80.2MB/s), io=142MiB (149MB), run=1859-1859msec 00:34:20.445 17:50:01 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:34:20.445 17:50:02 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@49 -- # '[' 1 -eq 1 ']' 00:34:20.445 17:50:02 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@51 -- # bdfs=($(get_nvme_bdfs)) 00:34:20.445 17:50:02 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@51 -- # get_nvme_bdfs 00:34:20.445 17:50:02 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1498 -- # bdfs=() 00:34:20.445 17:50:02 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1498 -- # local bdfs 00:34:20.445 17:50:02 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1499 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:34:20.445 17:50:02 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1499 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh 00:34:20.445 17:50:02 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1499 -- # jq -r '.config[].params.traddr' 00:34:20.445 17:50:02 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1500 -- # (( 1 == 0 )) 00:34:20.445 17:50:02 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1504 -- # printf '%s\n' 0000:88:00.0 00:34:20.445 17:50:02 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_nvme_attach_controller -b Nvme0 -t PCIe -a 0000:88:00.0 -i 10.0.0.2 00:34:23.723 Nvme0n1 00:34:23.723 17:50:05 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore -c 1073741824 Nvme0n1 lvs_0 00:34:27.003 17:50:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@53 -- # ls_guid=f9b3f042-d572-4442-b69e-ec94a2e1cb33 00:34:27.003 17:50:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@54 -- # get_lvs_free_mb f9b3f042-d572-4442-b69e-ec94a2e1cb33 00:34:27.003 17:50:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1368 -- # local lvs_uuid=f9b3f042-d572-4442-b69e-ec94a2e1cb33 00:34:27.003 17:50:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1369 -- # local lvs_info 00:34:27.003 17:50:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1370 -- # local fc 00:34:27.003 17:50:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1371 -- # local cs 00:34:27.003 17:50:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1372 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores 00:34:27.003 17:50:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1372 -- # lvs_info='[ 00:34:27.003 { 00:34:27.003 "uuid": "f9b3f042-d572-4442-b69e-ec94a2e1cb33", 00:34:27.003 "name": "lvs_0", 00:34:27.003 "base_bdev": "Nvme0n1", 00:34:27.003 "total_data_clusters": 930, 00:34:27.003 "free_clusters": 930, 00:34:27.003 "block_size": 512, 00:34:27.003 "cluster_size": 1073741824 00:34:27.003 } 00:34:27.003 ]' 00:34:27.003 17:50:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1373 -- # jq '.[] | select(.uuid=="f9b3f042-d572-4442-b69e-ec94a2e1cb33") .free_clusters' 00:34:27.003 17:50:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1373 -- # fc=930 00:34:27.003 17:50:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1374 -- # jq '.[] | select(.uuid=="f9b3f042-d572-4442-b69e-ec94a2e1cb33") .cluster_size' 00:34:27.003 17:50:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1374 -- # cs=1073741824 00:34:27.003 17:50:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1377 -- # free_mb=952320 00:34:27.003 17:50:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1378 -- # echo 952320 00:34:27.003 952320 00:34:27.003 17:50:08 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -l lvs_0 lbd_0 952320 00:34:27.261 e20a5c4a-a245-418f-8354-4e0ad0821940 00:34:27.261 17:50:09 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 -a -s SPDK00000000000001 00:34:27.519 17:50:09 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 lvs_0/lbd_0 00:34:27.777 17:50:09 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:34:28.342 17:50:09 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@59 -- # fio_nvme /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/example_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' --bs=4096 00:34:28.342 17:50:09 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1364 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/example_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' --bs=4096 00:34:28.342 17:50:09 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1341 -- # local fio_dir=/usr/src/fio 00:34:28.342 17:50:09 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1343 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:34:28.342 17:50:09 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1343 -- # local sanitizers 00:34:28.342 17:50:09 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1344 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:34:28.342 17:50:09 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # shift 00:34:28.342 17:50:09 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1347 -- # local asan_lib= 00:34:28.342 17:50:09 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:34:28.342 17:50:09 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:34:28.342 17:50:09 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # grep libasan 00:34:28.342 17:50:09 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:34:28.342 17:50:09 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # asan_lib= 00:34:28.342 17:50:09 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:34:28.342 17:50:09 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:34:28.342 17:50:09 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:34:28.343 17:50:09 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:34:28.343 17:50:09 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # grep libclang_rt.asan 00:34:28.343 17:50:09 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # asan_lib= 00:34:28.343 17:50:09 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:34:28.343 17:50:09 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1356 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme' 00:34:28.343 17:50:09 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1356 -- # /usr/src/fio/fio /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/example_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' --bs=4096 00:34:28.343 test: (g=0): rw=randrw, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk, iodepth=128 00:34:28.343 fio-3.35 00:34:28.343 Starting 1 thread 00:34:30.871 00:34:30.871 test: (groupid=0, jobs=1): err= 0: pid=367026: Fri Dec 6 17:50:12 2024 00:34:30.871 read: IOPS=5716, BW=22.3MiB/s (23.4MB/s)(44.8MiB/2008msec) 00:34:30.871 slat (nsec): min=1914, max=176405, avg=2679.74, stdev=2612.49 00:34:30.871 clat (usec): min=909, max=171583, avg=12193.29, stdev=11909.39 00:34:30.871 lat (usec): min=912, max=171635, avg=12195.97, stdev=11909.78 00:34:30.871 clat percentiles (msec): 00:34:30.871 | 1.00th=[ 9], 5.00th=[ 10], 10.00th=[ 11], 20.00th=[ 11], 00:34:30.871 | 30.00th=[ 11], 40.00th=[ 12], 50.00th=[ 12], 60.00th=[ 12], 00:34:30.871 | 70.00th=[ 12], 80.00th=[ 13], 90.00th=[ 13], 95.00th=[ 13], 00:34:30.871 | 99.00th=[ 14], 99.50th=[ 157], 99.90th=[ 171], 99.95th=[ 171], 00:34:30.871 | 99.99th=[ 171] 00:34:30.871 bw ( KiB/s): min=15896, max=25568, per=99.87%, avg=22836.00, stdev=4636.97, samples=4 00:34:30.871 iops : min= 3974, max= 6392, avg=5709.00, stdev=1159.24, samples=4 00:34:30.871 write: IOPS=5700, BW=22.3MiB/s (23.3MB/s)(44.7MiB/2008msec); 0 zone resets 00:34:30.871 slat (usec): min=2, max=150, avg= 2.83, stdev= 2.01 00:34:30.871 clat (usec): min=310, max=169408, avg=10024.00, stdev=11161.95 00:34:30.871 lat (usec): min=314, max=169415, avg=10026.83, stdev=11162.30 00:34:30.871 clat percentiles (msec): 00:34:30.871 | 1.00th=[ 8], 5.00th=[ 8], 10.00th=[ 9], 20.00th=[ 9], 00:34:30.871 | 30.00th=[ 9], 40.00th=[ 10], 50.00th=[ 10], 60.00th=[ 10], 00:34:30.871 | 70.00th=[ 10], 80.00th=[ 10], 90.00th=[ 11], 95.00th=[ 11], 00:34:30.871 | 99.00th=[ 12], 99.50th=[ 155], 99.90th=[ 169], 99.95th=[ 169], 00:34:30.871 | 99.99th=[ 169] 00:34:30.871 bw ( KiB/s): min=16872, max=24784, per=99.82%, avg=22762.00, stdev=3927.15, samples=4 00:34:30.871 iops : min= 4218, max= 6196, avg=5690.50, stdev=981.79, samples=4 00:34:30.871 lat (usec) : 500=0.01%, 750=0.01%, 1000=0.01% 00:34:30.871 lat (msec) : 2=0.03%, 4=0.11%, 10=46.49%, 20=52.79%, 250=0.56% 00:34:30.871 cpu : usr=61.78%, sys=36.87%, ctx=113, majf=0, minf=35 00:34:30.871 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.1%, >=64=99.7% 00:34:30.871 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:34:30.871 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:34:30.871 issued rwts: total=11479,11447,0,0 short=0,0,0,0 dropped=0,0,0,0 00:34:30.871 latency : target=0, window=0, percentile=100.00%, depth=128 00:34:30.871 00:34:30.871 Run status group 0 (all jobs): 00:34:30.871 READ: bw=22.3MiB/s (23.4MB/s), 22.3MiB/s-22.3MiB/s (23.4MB/s-23.4MB/s), io=44.8MiB (47.0MB), run=2008-2008msec 00:34:30.871 WRITE: bw=22.3MiB/s (23.3MB/s), 22.3MiB/s-22.3MiB/s (23.3MB/s-23.3MB/s), io=44.7MiB (46.9MB), run=2008-2008msec 00:34:30.871 17:50:12 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode2 00:34:31.130 17:50:12 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --clear-method none lvs_0/lbd_0 lvs_n_0 00:34:32.504 17:50:13 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@64 -- # ls_nested_guid=0121bf6f-7c03-46b8-92c3-597e4c5ca6c4 00:34:32.504 17:50:13 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@65 -- # get_lvs_free_mb 0121bf6f-7c03-46b8-92c3-597e4c5ca6c4 00:34:32.504 17:50:13 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1368 -- # local lvs_uuid=0121bf6f-7c03-46b8-92c3-597e4c5ca6c4 00:34:32.504 17:50:13 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1369 -- # local lvs_info 00:34:32.504 17:50:13 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1370 -- # local fc 00:34:32.504 17:50:13 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1371 -- # local cs 00:34:32.504 17:50:13 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1372 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores 00:34:32.504 17:50:14 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1372 -- # lvs_info='[ 00:34:32.504 { 00:34:32.504 "uuid": "f9b3f042-d572-4442-b69e-ec94a2e1cb33", 00:34:32.504 "name": "lvs_0", 00:34:32.504 "base_bdev": "Nvme0n1", 00:34:32.504 "total_data_clusters": 930, 00:34:32.504 "free_clusters": 0, 00:34:32.504 "block_size": 512, 00:34:32.504 "cluster_size": 1073741824 00:34:32.504 }, 00:34:32.504 { 00:34:32.504 "uuid": "0121bf6f-7c03-46b8-92c3-597e4c5ca6c4", 00:34:32.504 "name": "lvs_n_0", 00:34:32.504 "base_bdev": "e20a5c4a-a245-418f-8354-4e0ad0821940", 00:34:32.504 "total_data_clusters": 237847, 00:34:32.504 "free_clusters": 237847, 00:34:32.504 "block_size": 512, 00:34:32.504 "cluster_size": 4194304 00:34:32.504 } 00:34:32.504 ]' 00:34:32.504 17:50:14 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1373 -- # jq '.[] | select(.uuid=="0121bf6f-7c03-46b8-92c3-597e4c5ca6c4") .free_clusters' 00:34:32.504 17:50:14 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1373 -- # fc=237847 00:34:32.504 17:50:14 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1374 -- # jq '.[] | select(.uuid=="0121bf6f-7c03-46b8-92c3-597e4c5ca6c4") .cluster_size' 00:34:32.504 17:50:14 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1374 -- # cs=4194304 00:34:32.504 17:50:14 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1377 -- # free_mb=951388 00:34:32.504 17:50:14 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1378 -- # echo 951388 00:34:32.504 951388 00:34:32.504 17:50:14 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -l lvs_n_0 lbd_nest_0 951388 00:34:33.438 b9c761c5-7c18-447b-bdb8-b29803d1a0b4 00:34:33.438 17:50:14 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@67 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode3 -a -s SPDK00000000000001 00:34:33.438 17:50:15 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@68 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode3 lvs_n_0/lbd_nest_0 00:34:33.696 17:50:15 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@69 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode3 -t tcp -a 10.0.0.2 -s 4420 00:34:33.954 17:50:15 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@70 -- # fio_nvme /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/example_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' --bs=4096 00:34:33.954 17:50:15 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1364 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/example_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' --bs=4096 00:34:33.954 17:50:15 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1341 -- # local fio_dir=/usr/src/fio 00:34:33.954 17:50:15 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1343 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:34:33.954 17:50:15 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1343 -- # local sanitizers 00:34:33.954 17:50:15 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1344 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:34:33.954 17:50:15 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # shift 00:34:33.954 17:50:15 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1347 -- # local asan_lib= 00:34:33.954 17:50:15 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:34:33.954 17:50:15 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:34:33.954 17:50:15 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # grep libasan 00:34:33.954 17:50:15 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:34:33.954 17:50:15 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # asan_lib= 00:34:33.954 17:50:15 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:34:33.954 17:50:15 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:34:33.954 17:50:15 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:34:33.954 17:50:15 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # grep libclang_rt.asan 00:34:33.954 17:50:15 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:34:34.212 17:50:15 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1349 -- # asan_lib= 00:34:34.212 17:50:15 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:34:34.212 17:50:15 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1356 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme' 00:34:34.212 17:50:15 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1356 -- # /usr/src/fio/fio /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/example_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' --bs=4096 00:34:34.212 test: (g=0): rw=randrw, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk, iodepth=128 00:34:34.212 fio-3.35 00:34:34.212 Starting 1 thread 00:34:36.739 00:34:36.739 test: (groupid=0, jobs=1): err= 0: pid=367878: Fri Dec 6 17:50:18 2024 00:34:36.739 read: IOPS=5776, BW=22.6MiB/s (23.7MB/s)(45.3MiB/2009msec) 00:34:36.739 slat (nsec): min=1849, max=147724, avg=2468.25, stdev=2060.14 00:34:36.739 clat (usec): min=4276, max=20819, avg=12128.50, stdev=1157.09 00:34:36.739 lat (usec): min=4281, max=20822, avg=12130.97, stdev=1156.95 00:34:36.739 clat percentiles (usec): 00:34:36.739 | 1.00th=[ 9634], 5.00th=[10421], 10.00th=[10683], 20.00th=[11207], 00:34:36.739 | 30.00th=[11600], 40.00th=[11863], 50.00th=[12125], 60.00th=[12387], 00:34:36.739 | 70.00th=[12649], 80.00th=[13042], 90.00th=[13435], 95.00th=[13829], 00:34:36.739 | 99.00th=[14746], 99.50th=[15008], 99.90th=[19792], 99.95th=[20579], 00:34:36.739 | 99.99th=[20841] 00:34:36.739 bw ( KiB/s): min=21736, max=23648, per=99.77%, avg=23054.00, stdev=885.45, samples=4 00:34:36.739 iops : min= 5434, max= 5912, avg=5763.50, stdev=221.36, samples=4 00:34:36.739 write: IOPS=5759, BW=22.5MiB/s (23.6MB/s)(45.2MiB/2009msec); 0 zone resets 00:34:36.739 slat (nsec): min=1963, max=113470, avg=2535.63, stdev=1496.08 00:34:36.739 clat (usec): min=2024, max=18376, avg=9873.04, stdev=931.33 00:34:36.739 lat (usec): min=2030, max=18378, avg=9875.57, stdev=931.27 00:34:36.739 clat percentiles (usec): 00:34:36.739 | 1.00th=[ 7767], 5.00th=[ 8455], 10.00th=[ 8848], 20.00th=[ 9241], 00:34:36.739 | 30.00th=[ 9372], 40.00th=[ 9634], 50.00th=[ 9896], 60.00th=[10028], 00:34:36.739 | 70.00th=[10290], 80.00th=[10552], 90.00th=[10945], 95.00th=[11207], 00:34:36.739 | 99.00th=[11863], 99.50th=[12125], 99.90th=[16188], 99.95th=[17433], 00:34:36.739 | 99.99th=[18482] 00:34:36.739 bw ( KiB/s): min=22800, max=23232, per=100.00%, avg=23044.00, stdev=221.85, samples=4 00:34:36.739 iops : min= 5700, max= 5808, avg=5761.00, stdev=55.46, samples=4 00:34:36.739 lat (msec) : 4=0.04%, 10=29.08%, 20=70.83%, 50=0.05% 00:34:36.739 cpu : usr=61.21%, sys=37.55%, ctx=124, majf=0, minf=35 00:34:36.739 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.1%, >=64=99.7% 00:34:36.739 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:34:36.739 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:34:36.739 issued rwts: total=11605,11571,0,0 short=0,0,0,0 dropped=0,0,0,0 00:34:36.739 latency : target=0, window=0, percentile=100.00%, depth=128 00:34:36.739 00:34:36.739 Run status group 0 (all jobs): 00:34:36.739 READ: bw=22.6MiB/s (23.7MB/s), 22.6MiB/s-22.6MiB/s (23.7MB/s-23.7MB/s), io=45.3MiB (47.5MB), run=2009-2009msec 00:34:36.739 WRITE: bw=22.5MiB/s (23.6MB/s), 22.5MiB/s-22.5MiB/s (23.6MB/s-23.6MB/s), io=45.2MiB (47.4MB), run=2009-2009msec 00:34:36.739 17:50:18 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode3 00:34:36.996 17:50:18 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@74 -- # sync 00:34:36.996 17:50:18 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@76 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -t 120 bdev_lvol_delete lvs_n_0/lbd_nest_0 00:34:41.176 17:50:22 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@77 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -l lvs_n_0 00:34:41.176 17:50:22 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@78 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete lvs_0/lbd_0 00:34:44.453 17:50:25 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -l lvs_0 00:34:44.453 17:50:26 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@80 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_nvme_detach_controller Nvme0 00:34:46.354 17:50:27 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:34:46.354 17:50:27 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@85 -- # rm -f ./local-test-0-verify.state 00:34:46.354 17:50:27 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@86 -- # nvmftestfini 00:34:46.354 17:50:27 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@516 -- # nvmfcleanup 00:34:46.354 17:50:27 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@121 -- # sync 00:34:46.354 17:50:27 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:34:46.354 17:50:27 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@124 -- # set +e 00:34:46.354 17:50:27 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@125 -- # for i in {1..20} 00:34:46.354 17:50:27 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:34:46.354 rmmod nvme_tcp 00:34:46.354 rmmod nvme_fabrics 00:34:46.354 rmmod nvme_keyring 00:34:46.354 17:50:28 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:34:46.354 17:50:28 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@128 -- # set -e 00:34:46.354 17:50:28 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@129 -- # return 0 00:34:46.354 17:50:28 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@517 -- # '[' -n 364938 ']' 00:34:46.354 17:50:28 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@518 -- # killprocess 364938 00:34:46.354 17:50:28 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@954 -- # '[' -z 364938 ']' 00:34:46.354 17:50:28 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@958 -- # kill -0 364938 00:34:46.354 17:50:28 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@959 -- # uname 00:34:46.354 17:50:28 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:34:46.354 17:50:28 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 364938 00:34:46.354 17:50:28 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:34:46.354 17:50:28 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:34:46.354 17:50:28 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@972 -- # echo 'killing process with pid 364938' 00:34:46.354 killing process with pid 364938 00:34:46.354 17:50:28 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@973 -- # kill 364938 00:34:46.354 17:50:28 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@978 -- # wait 364938 00:34:46.613 17:50:28 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:34:46.613 17:50:28 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:34:46.613 17:50:28 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:34:46.613 17:50:28 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@297 -- # iptr 00:34:46.613 17:50:28 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@791 -- # iptables-save 00:34:46.613 17:50:28 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:34:46.613 17:50:28 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@791 -- # iptables-restore 00:34:46.613 17:50:28 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:34:46.613 17:50:28 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@302 -- # remove_spdk_ns 00:34:46.613 17:50:28 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:34:46.613 17:50:28 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:34:46.613 17:50:28 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:34:48.521 17:50:30 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:34:48.521 00:34:48.521 real 0m38.025s 00:34:48.521 user 2m26.012s 00:34:48.521 sys 0m7.220s 00:34:48.521 17:50:30 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1130 -- # xtrace_disable 00:34:48.521 17:50:30 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@10 -- # set +x 00:34:48.521 ************************************ 00:34:48.521 END TEST nvmf_fio_host 00:34:48.521 ************************************ 00:34:48.780 17:50:30 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@25 -- # run_test nvmf_failover /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/failover.sh --transport=tcp 00:34:48.780 17:50:30 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:34:48.780 17:50:30 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:34:48.780 17:50:30 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:34:48.780 ************************************ 00:34:48.780 START TEST nvmf_failover 00:34:48.780 ************************************ 00:34:48.780 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/failover.sh --transport=tcp 00:34:48.780 * Looking for test storage... 00:34:48.780 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:34:48.780 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:34:48.780 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1711 -- # lcov --version 00:34:48.780 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:34:48.780 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:34:48.780 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:34:48.780 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@333 -- # local ver1 ver1_l 00:34:48.780 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@334 -- # local ver2 ver2_l 00:34:48.780 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@336 -- # IFS=.-: 00:34:48.780 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@336 -- # read -ra ver1 00:34:48.780 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@337 -- # IFS=.-: 00:34:48.780 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@337 -- # read -ra ver2 00:34:48.780 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@338 -- # local 'op=<' 00:34:48.780 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@340 -- # ver1_l=2 00:34:48.780 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@341 -- # ver2_l=1 00:34:48.780 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:34:48.780 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@344 -- # case "$op" in 00:34:48.780 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@345 -- # : 1 00:34:48.780 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@364 -- # (( v = 0 )) 00:34:48.780 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:34:48.780 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@365 -- # decimal 1 00:34:48.780 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@353 -- # local d=1 00:34:48.780 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@355 -- # echo 1 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@365 -- # ver1[v]=1 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@366 -- # decimal 2 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@353 -- # local d=2 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@355 -- # echo 2 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@366 -- # ver2[v]=2 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@368 -- # return 0 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:34:48.781 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:48.781 --rc genhtml_branch_coverage=1 00:34:48.781 --rc genhtml_function_coverage=1 00:34:48.781 --rc genhtml_legend=1 00:34:48.781 --rc geninfo_all_blocks=1 00:34:48.781 --rc geninfo_unexecuted_blocks=1 00:34:48.781 00:34:48.781 ' 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:34:48.781 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:48.781 --rc genhtml_branch_coverage=1 00:34:48.781 --rc genhtml_function_coverage=1 00:34:48.781 --rc genhtml_legend=1 00:34:48.781 --rc geninfo_all_blocks=1 00:34:48.781 --rc geninfo_unexecuted_blocks=1 00:34:48.781 00:34:48.781 ' 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:34:48.781 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:48.781 --rc genhtml_branch_coverage=1 00:34:48.781 --rc genhtml_function_coverage=1 00:34:48.781 --rc genhtml_legend=1 00:34:48.781 --rc geninfo_all_blocks=1 00:34:48.781 --rc geninfo_unexecuted_blocks=1 00:34:48.781 00:34:48.781 ' 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:34:48.781 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:48.781 --rc genhtml_branch_coverage=1 00:34:48.781 --rc genhtml_function_coverage=1 00:34:48.781 --rc genhtml_legend=1 00:34:48.781 --rc geninfo_all_blocks=1 00:34:48.781 --rc geninfo_unexecuted_blocks=1 00:34:48.781 00:34:48.781 ' 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@7 -- # uname -s 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@15 -- # shopt -s extglob 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- paths/export.sh@5 -- # export PATH 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@51 -- # : 0 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:34:48.781 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@55 -- # have_pci_nics=0 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@11 -- # MALLOC_BDEV_SIZE=64 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@14 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@16 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@18 -- # nvmftestinit 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@476 -- # prepare_net_devs 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@438 -- # local -g is_hw=no 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@440 -- # remove_spdk_ns 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@309 -- # xtrace_disable 00:34:48.781 17:50:30 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@315 -- # pci_devs=() 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@315 -- # local -a pci_devs 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@316 -- # pci_net_devs=() 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@317 -- # pci_drivers=() 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@317 -- # local -A pci_drivers 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@319 -- # net_devs=() 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@319 -- # local -ga net_devs 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@320 -- # e810=() 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@320 -- # local -ga e810 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@321 -- # x722=() 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@321 -- # local -ga x722 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@322 -- # mlx=() 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@322 -- # local -ga mlx 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:34:51.313 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:34:51.313 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@418 -- # [[ up == up ]] 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:34:51.313 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:34:51.314 Found net devices under 0000:0a:00.0: cvl_0_0 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@418 -- # [[ up == up ]] 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:34:51.314 Found net devices under 0000:0a:00.1: cvl_0_1 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@442 -- # is_hw=yes 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:34:51.314 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:34:51.314 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.218 ms 00:34:51.314 00:34:51.314 --- 10.0.0.2 ping statistics --- 00:34:51.314 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:34:51.314 rtt min/avg/max/mdev = 0.218/0.218/0.218/0.000 ms 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:34:51.314 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:34:51.314 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.163 ms 00:34:51.314 00:34:51.314 --- 10.0.0.1 ping statistics --- 00:34:51.314 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:34:51.314 rtt min/avg/max/mdev = 0.163/0.163/0.163/0.000 ms 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@450 -- # return 0 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@20 -- # nvmfappstart -m 0xE 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@726 -- # xtrace_disable 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@509 -- # nvmfpid=371131 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@510 -- # waitforlisten 371131 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@835 -- # '[' -z 371131 ']' 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@840 -- # local max_retries=100 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:34:51.314 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@844 -- # xtrace_disable 00:34:51.314 17:50:32 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:34:51.314 [2024-12-06 17:50:32.932985] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:34:51.314 [2024-12-06 17:50:32.933080] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:34:51.314 [2024-12-06 17:50:33.007600] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:34:51.314 [2024-12-06 17:50:33.055094] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:34:51.314 [2024-12-06 17:50:33.055150] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:34:51.314 [2024-12-06 17:50:33.055179] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:34:51.314 [2024-12-06 17:50:33.055191] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:34:51.314 [2024-12-06 17:50:33.055201] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:34:51.314 [2024-12-06 17:50:33.056766] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:34:51.314 [2024-12-06 17:50:33.060685] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:34:51.314 [2024-12-06 17:50:33.060697] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:34:51.572 17:50:33 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:34:51.572 17:50:33 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@868 -- # return 0 00:34:51.572 17:50:33 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:34:51.572 17:50:33 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@732 -- # xtrace_disable 00:34:51.572 17:50:33 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:34:51.572 17:50:33 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:34:51.572 17:50:33 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@22 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:34:51.830 [2024-12-06 17:50:33.438926] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:34:51.830 17:50:33 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@23 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc0 00:34:52.088 Malloc0 00:34:52.088 17:50:33 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:34:52.346 17:50:34 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:34:52.604 17:50:34 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:34:52.862 [2024-12-06 17:50:34.555250] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:34:52.862 17:50:34 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:34:53.120 [2024-12-06 17:50:34.828043] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:34:53.120 17:50:34 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4422 00:34:53.379 [2024-12-06 17:50:35.100916] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4422 *** 00:34:53.379 17:50:35 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@31 -- # bdevperf_pid=371429 00:34:53.379 17:50:35 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 15 -f 00:34:53.379 17:50:35 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@33 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; cat $testdir/try.txt; rm -f $testdir/try.txt; killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:34:53.379 17:50:35 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@34 -- # waitforlisten 371429 /var/tmp/bdevperf.sock 00:34:53.379 17:50:35 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@835 -- # '[' -z 371429 ']' 00:34:53.379 17:50:35 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:34:53.379 17:50:35 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@840 -- # local max_retries=100 00:34:53.379 17:50:35 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:34:53.379 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:34:53.379 17:50:35 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@844 -- # xtrace_disable 00:34:53.379 17:50:35 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:34:53.637 17:50:35 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:34:53.637 17:50:35 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@868 -- # return 0 00:34:53.637 17:50:35 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x failover 00:34:54.201 NVMe0n1 00:34:54.201 17:50:35 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x failover 00:34:54.458 00:34:54.458 17:50:36 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@39 -- # run_test_pid=371561 00:34:54.458 17:50:36 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:34:54.458 17:50:36 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@41 -- # sleep 1 00:34:55.390 17:50:37 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:34:55.648 [2024-12-06 17:50:37.411831] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.411916] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.411933] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.411945] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.411957] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.411969] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.411981] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.411993] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412005] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412017] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412029] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412040] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412053] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412065] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412076] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412088] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412100] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412126] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412138] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412149] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412160] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412171] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412182] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412194] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412205] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412217] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412228] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412239] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412254] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412265] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412276] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412287] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412299] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412311] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412323] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412334] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412345] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412356] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412367] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412379] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412389] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412400] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412412] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412422] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412433] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412460] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412472] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412484] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412495] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412506] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412518] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412529] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412540] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412552] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412565] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412580] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412592] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412603] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412615] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412627] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412639] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 [2024-12-06 17:50:37.412650] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xb3e190 is same with the state(6) to be set 00:34:55.649 17:50:37 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@45 -- # sleep 3 00:34:58.929 17:50:40 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4422 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x failover 00:34:59.188 00:34:59.189 17:50:41 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:34:59.447 17:50:41 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@50 -- # sleep 3 00:35:02.731 17:50:44 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:35:02.989 [2024-12-06 17:50:44.595594] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:35:02.989 17:50:44 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@55 -- # sleep 1 00:35:03.923 17:50:45 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4422 00:35:04.181 [2024-12-06 17:50:45.869308] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.181 [2024-12-06 17:50:45.869376] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.181 [2024-12-06 17:50:45.869393] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.181 [2024-12-06 17:50:45.869406] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.181 [2024-12-06 17:50:45.869418] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.181 [2024-12-06 17:50:45.869431] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.181 [2024-12-06 17:50:45.869443] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.181 [2024-12-06 17:50:45.869455] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.181 [2024-12-06 17:50:45.869467] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.181 [2024-12-06 17:50:45.869494] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.181 [2024-12-06 17:50:45.869506] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.181 [2024-12-06 17:50:45.869517] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.181 [2024-12-06 17:50:45.869544] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.181 [2024-12-06 17:50:45.869557] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.181 [2024-12-06 17:50:45.869569] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.181 [2024-12-06 17:50:45.869580] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.182 [2024-12-06 17:50:45.869592] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.182 [2024-12-06 17:50:45.869603] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.182 [2024-12-06 17:50:45.869614] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.182 [2024-12-06 17:50:45.869625] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.182 [2024-12-06 17:50:45.869637] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.182 [2024-12-06 17:50:45.869655] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.182 [2024-12-06 17:50:45.869691] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.182 [2024-12-06 17:50:45.869705] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.182 [2024-12-06 17:50:45.869717] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.182 [2024-12-06 17:50:45.869729] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.182 [2024-12-06 17:50:45.869740] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.182 [2024-12-06 17:50:45.869751] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.182 [2024-12-06 17:50:45.869763] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.182 [2024-12-06 17:50:45.869775] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.182 [2024-12-06 17:50:45.869786] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.182 [2024-12-06 17:50:45.869798] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.182 [2024-12-06 17:50:45.869809] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.182 [2024-12-06 17:50:45.869821] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.182 [2024-12-06 17:50:45.869834] tcp.c:1790:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xc8b300 is same with the state(6) to be set 00:35:04.182 17:50:45 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@59 -- # wait 371561 00:35:10.750 { 00:35:10.750 "results": [ 00:35:10.750 { 00:35:10.751 "job": "NVMe0n1", 00:35:10.751 "core_mask": "0x1", 00:35:10.751 "workload": "verify", 00:35:10.751 "status": "finished", 00:35:10.751 "verify_range": { 00:35:10.751 "start": 0, 00:35:10.751 "length": 16384 00:35:10.751 }, 00:35:10.751 "queue_depth": 128, 00:35:10.751 "io_size": 4096, 00:35:10.751 "runtime": 15.013193, 00:35:10.751 "iops": 8523.56990281814, 00:35:10.751 "mibps": 33.29519493288336, 00:35:10.751 "io_failed": 9180, 00:35:10.751 "io_timeout": 0, 00:35:10.751 "avg_latency_us": 13983.817738004, 00:35:10.751 "min_latency_us": 552.2014814814814, 00:35:10.751 "max_latency_us": 16311.182222222222 00:35:10.751 } 00:35:10.751 ], 00:35:10.751 "core_count": 1 00:35:10.751 } 00:35:10.751 17:50:51 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@61 -- # killprocess 371429 00:35:10.751 17:50:51 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@954 -- # '[' -z 371429 ']' 00:35:10.751 17:50:51 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@958 -- # kill -0 371429 00:35:10.751 17:50:51 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@959 -- # uname 00:35:10.751 17:50:51 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:35:10.751 17:50:51 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 371429 00:35:10.751 17:50:51 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:35:10.751 17:50:51 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:35:10.751 17:50:51 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@972 -- # echo 'killing process with pid 371429' 00:35:10.751 killing process with pid 371429 00:35:10.751 17:50:51 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@973 -- # kill 371429 00:35:10.751 17:50:51 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@978 -- # wait 371429 00:35:10.751 17:50:51 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@63 -- # cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:35:10.751 [2024-12-06 17:50:35.163398] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:35:10.751 [2024-12-06 17:50:35.163486] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid371429 ] 00:35:10.751 [2024-12-06 17:50:35.232272] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:35:10.751 [2024-12-06 17:50:35.279558] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:35:10.751 Running I/O for 15 seconds... 00:35:10.751 8634.00 IOPS, 33.73 MiB/s [2024-12-06T16:50:52.590Z] [2024-12-06 17:50:37.414157] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:79088 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.751 [2024-12-06 17:50:37.414197] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.751 [2024-12-06 17:50:37.414225] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:79096 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.751 [2024-12-06 17:50:37.414241] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.751 [2024-12-06 17:50:37.414258] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:79104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.751 [2024-12-06 17:50:37.414273] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.751 [2024-12-06 17:50:37.414288] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:79112 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.751 [2024-12-06 17:50:37.414303] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.751 [2024-12-06 17:50:37.414318] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:79120 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.751 [2024-12-06 17:50:37.414333] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.751 [2024-12-06 17:50:37.414349] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:85 nsid:1 lba:79128 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.751 [2024-12-06 17:50:37.414363] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.751 [2024-12-06 17:50:37.414379] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:79136 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.751 [2024-12-06 17:50:37.414393] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.751 [2024-12-06 17:50:37.414409] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:79144 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.751 [2024-12-06 17:50:37.414423] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.751 [2024-12-06 17:50:37.414439] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:64 nsid:1 lba:79152 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.751 [2024-12-06 17:50:37.414453] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.751 [2024-12-06 17:50:37.414468] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:116 nsid:1 lba:79160 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.751 [2024-12-06 17:50:37.414483] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.751 [2024-12-06 17:50:37.414498] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:79168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.751 [2024-12-06 17:50:37.414513] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.751 [2024-12-06 17:50:37.414535] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:79176 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.751 [2024-12-06 17:50:37.414551] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.751 [2024-12-06 17:50:37.414566] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:79184 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.751 [2024-12-06 17:50:37.414581] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.751 [2024-12-06 17:50:37.414596] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:79192 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.751 [2024-12-06 17:50:37.414610] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.751 [2024-12-06 17:50:37.414625] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:123 nsid:1 lba:79200 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.751 [2024-12-06 17:50:37.414639] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.751 [2024-12-06 17:50:37.414655] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:82 nsid:1 lba:79208 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.751 [2024-12-06 17:50:37.414691] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.751 [2024-12-06 17:50:37.414709] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:79216 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.751 [2024-12-06 17:50:37.414724] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.751 [2024-12-06 17:50:37.414740] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:79224 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.751 [2024-12-06 17:50:37.414754] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.751 [2024-12-06 17:50:37.414771] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:79232 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.751 [2024-12-06 17:50:37.414785] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.751 [2024-12-06 17:50:37.414800] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:107 nsid:1 lba:79240 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.751 [2024-12-06 17:50:37.414814] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.751 [2024-12-06 17:50:37.414830] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:79248 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.751 [2024-12-06 17:50:37.414844] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.751 [2024-12-06 17:50:37.414860] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:79256 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.751 [2024-12-06 17:50:37.414874] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.751 [2024-12-06 17:50:37.414890] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:79264 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.751 [2024-12-06 17:50:37.414904] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.751 [2024-12-06 17:50:37.414920] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:83 nsid:1 lba:79272 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.751 [2024-12-06 17:50:37.414938] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.751 [2024-12-06 17:50:37.414954] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:117 nsid:1 lba:79280 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.751 [2024-12-06 17:50:37.414984] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.751 [2024-12-06 17:50:37.415000] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:79288 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.751 [2024-12-06 17:50:37.415015] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.751 [2024-12-06 17:50:37.415030] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:79296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.751 [2024-12-06 17:50:37.415044] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.751 [2024-12-06 17:50:37.415059] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:79304 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.751 [2024-12-06 17:50:37.415073] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415089] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:79312 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.752 [2024-12-06 17:50:37.415103] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415118] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:79320 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.752 [2024-12-06 17:50:37.415132] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415147] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:79328 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.752 [2024-12-06 17:50:37.415161] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415176] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:79336 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.752 [2024-12-06 17:50:37.415190] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415205] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:75 nsid:1 lba:79344 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.752 [2024-12-06 17:50:37.415219] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415234] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:79352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.752 [2024-12-06 17:50:37.415248] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415263] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:79360 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.752 [2024-12-06 17:50:37.415276] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415292] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:79368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.752 [2024-12-06 17:50:37.415306] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415325] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:79376 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.752 [2024-12-06 17:50:37.415339] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415355] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:79392 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.415368] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415384] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:79400 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.415398] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415413] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:79408 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.415427] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415441] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:79416 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.415455] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415471] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:79424 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.415484] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415499] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:79432 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.415513] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415528] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:79440 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.415542] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415557] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:79448 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.415571] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415586] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:79456 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.415600] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415615] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:79464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.415629] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415644] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:113 nsid:1 lba:79472 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.415658] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415697] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:69 nsid:1 lba:79480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.415717] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415733] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:79488 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.415748] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415763] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:79496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.415777] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415793] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:79504 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.415807] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415823] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:79512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.415837] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415853] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:79520 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.415866] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415882] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:100 nsid:1 lba:79528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.415897] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415912] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:79536 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.415926] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415948] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:79544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.415963] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.415994] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:79552 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.416007] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.416023] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:79560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.416036] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.416052] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:79568 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.416066] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.416081] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:79576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.416094] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.416110] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:79584 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.416127] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.416143] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:99 nsid:1 lba:79592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.416157] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.416172] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:79600 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.416186] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.416201] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:104 nsid:1 lba:79384 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.752 [2024-12-06 17:50:37.416215] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.416230] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:79608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.416244] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.416259] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:79616 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.416272] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.752 [2024-12-06 17:50:37.416288] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:79624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.752 [2024-12-06 17:50:37.416301] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.416316] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:79632 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.416330] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.416345] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:79640 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.416359] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.416374] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:79648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.416387] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.416402] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:79656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.416431] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.416453] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:79664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.416468] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.416483] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:79672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.416497] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.416516] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:79680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.416531] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.416547] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:79688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.416561] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.416577] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:79696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.416591] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.416606] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:79704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.416621] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.416636] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:79712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.416650] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.416676] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:79720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.416693] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.416709] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:70 nsid:1 lba:79728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.416724] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.416739] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:105 nsid:1 lba:79736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.416754] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.416769] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:93 nsid:1 lba:79744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.416784] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.416800] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:80 nsid:1 lba:79752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.416814] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.416829] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:79760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.416843] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.416858] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:79768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.416873] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.416889] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:88 nsid:1 lba:79776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.416907] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.416924] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:98 nsid:1 lba:79784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.416938] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.416958] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:79792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.416973] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.416989] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:79800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.417003] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.417019] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:79808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.417032] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.417048] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:79816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.417062] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.417078] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:79824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.417091] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.417107] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:74 nsid:1 lba:79832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.417121] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.417136] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:79840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.417150] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.417165] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:79848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.417179] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.417195] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:79856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.417210] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.417225] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:79864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.417239] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.417255] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:79872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.417269] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.417284] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:114 nsid:1 lba:79880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.417302] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.417318] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:79888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.417333] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.417348] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:120 nsid:1 lba:79896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.417363] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.417379] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:97 nsid:1 lba:79904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.417393] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.417408] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:79912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.417422] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.417443] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:79920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.417458] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.417473] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:79928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.417487] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.417502] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:79936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.753 [2024-12-06 17:50:37.417517] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.753 [2024-12-06 17:50:37.417532] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:79944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.754 [2024-12-06 17:50:37.417546] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.754 [2024-12-06 17:50:37.417567] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:79952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.754 [2024-12-06 17:50:37.417582] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.754 [2024-12-06 17:50:37.417599] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:79960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.754 [2024-12-06 17:50:37.417613] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.754 [2024-12-06 17:50:37.417629] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:77 nsid:1 lba:79968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.754 [2024-12-06 17:50:37.417643] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.754 [2024-12-06 17:50:37.417660] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:79976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.754 [2024-12-06 17:50:37.417684] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.754 [2024-12-06 17:50:37.417705] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:79984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.754 [2024-12-06 17:50:37.417720] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.754 [2024-12-06 17:50:37.417750] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:35:10.754 [2024-12-06 17:50:37.417768] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:79992 len:8 PRP1 0x0 PRP2 0x0 00:35:10.754 [2024-12-06 17:50:37.417782] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.754 [2024-12-06 17:50:37.417801] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:35:10.754 [2024-12-06 17:50:37.417813] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:35:10.754 [2024-12-06 17:50:37.417825] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:80000 len:8 PRP1 0x0 PRP2 0x0 00:35:10.754 [2024-12-06 17:50:37.417839] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.754 [2024-12-06 17:50:37.417853] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:35:10.754 [2024-12-06 17:50:37.417865] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:35:10.754 [2024-12-06 17:50:37.417877] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:80008 len:8 PRP1 0x0 PRP2 0x0 00:35:10.754 [2024-12-06 17:50:37.417890] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.754 [2024-12-06 17:50:37.417904] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:35:10.754 [2024-12-06 17:50:37.417916] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:35:10.754 [2024-12-06 17:50:37.417928] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:80016 len:8 PRP1 0x0 PRP2 0x0 00:35:10.754 [2024-12-06 17:50:37.417942] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.754 [2024-12-06 17:50:37.417956] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:35:10.754 [2024-12-06 17:50:37.417967] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:35:10.754 [2024-12-06 17:50:37.417979] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:80024 len:8 PRP1 0x0 PRP2 0x0 00:35:10.754 [2024-12-06 17:50:37.417993] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.754 [2024-12-06 17:50:37.418007] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:35:10.754 [2024-12-06 17:50:37.418019] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:35:10.754 [2024-12-06 17:50:37.418030] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:80032 len:8 PRP1 0x0 PRP2 0x0 00:35:10.754 [2024-12-06 17:50:37.418045] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.754 [2024-12-06 17:50:37.418059] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:35:10.754 [2024-12-06 17:50:37.418071] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:35:10.754 [2024-12-06 17:50:37.418082] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:80040 len:8 PRP1 0x0 PRP2 0x0 00:35:10.754 [2024-12-06 17:50:37.418096] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.754 [2024-12-06 17:50:37.418109] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:35:10.754 [2024-12-06 17:50:37.418125] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:35:10.754 [2024-12-06 17:50:37.418138] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:80048 len:8 PRP1 0x0 PRP2 0x0 00:35:10.754 [2024-12-06 17:50:37.418151] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.754 [2024-12-06 17:50:37.418164] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:35:10.754 [2024-12-06 17:50:37.418176] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:35:10.754 [2024-12-06 17:50:37.418189] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:80056 len:8 PRP1 0x0 PRP2 0x0 00:35:10.754 [2024-12-06 17:50:37.418202] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.754 [2024-12-06 17:50:37.418215] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:35:10.754 [2024-12-06 17:50:37.418227] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:35:10.754 [2024-12-06 17:50:37.418239] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:80064 len:8 PRP1 0x0 PRP2 0x0 00:35:10.754 [2024-12-06 17:50:37.418253] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.754 [2024-12-06 17:50:37.418266] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:35:10.754 [2024-12-06 17:50:37.418278] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:35:10.754 [2024-12-06 17:50:37.418289] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:80072 len:8 PRP1 0x0 PRP2 0x0 00:35:10.754 [2024-12-06 17:50:37.418303] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.754 [2024-12-06 17:50:37.418316] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:35:10.754 [2024-12-06 17:50:37.418328] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:35:10.754 [2024-12-06 17:50:37.418340] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:80080 len:8 PRP1 0x0 PRP2 0x0 00:35:10.754 [2024-12-06 17:50:37.418354] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.754 [2024-12-06 17:50:37.418367] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:35:10.754 [2024-12-06 17:50:37.418379] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:35:10.754 [2024-12-06 17:50:37.418391] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:80088 len:8 PRP1 0x0 PRP2 0x0 00:35:10.754 [2024-12-06 17:50:37.418404] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.754 [2024-12-06 17:50:37.418418] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:35:10.754 [2024-12-06 17:50:37.418429] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:35:10.754 [2024-12-06 17:50:37.418441] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:80096 len:8 PRP1 0x0 PRP2 0x0 00:35:10.754 [2024-12-06 17:50:37.418455] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.754 [2024-12-06 17:50:37.418469] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:35:10.754 [2024-12-06 17:50:37.418481] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:35:10.754 [2024-12-06 17:50:37.418493] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:80104 len:8 PRP1 0x0 PRP2 0x0 00:35:10.754 [2024-12-06 17:50:37.418506] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.754 [2024-12-06 17:50:37.418572] bdev_nvme.c:2056:bdev_nvme_failover_trid: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 1] Start failover from 10.0.0.2:4420 to 10.0.0.2:4421 00:35:10.754 [2024-12-06 17:50:37.418611] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:35:10.754 [2024-12-06 17:50:37.418630] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.754 [2024-12-06 17:50:37.418646] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:35:10.754 [2024-12-06 17:50:37.418659] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.754 [2024-12-06 17:50:37.418683] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:35:10.754 [2024-12-06 17:50:37.418698] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.754 [2024-12-06 17:50:37.418713] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:35:10.754 [2024-12-06 17:50:37.418727] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.754 [2024-12-06 17:50:37.418741] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] in failed state. 00:35:10.754 [2024-12-06 17:50:37.418791] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x155b900 (9): Bad file descriptor 00:35:10.754 [2024-12-06 17:50:37.422081] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 1] resetting controller 00:35:10.754 [2024-12-06 17:50:37.445812] bdev_nvme.c:2286:bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 3] Resetting controller successful. 00:35:10.754 8505.50 IOPS, 33.22 MiB/s [2024-12-06T16:50:52.593Z] 8511.33 IOPS, 33.25 MiB/s [2024-12-06T16:50:52.593Z] 8521.75 IOPS, 33.29 MiB/s [2024-12-06T16:50:52.593Z] [2024-12-06 17:50:41.264025] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:87712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.754 [2024-12-06 17:50:41.264092] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.754 [2024-12-06 17:50:41.264121] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:87720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.754 [2024-12-06 17:50:41.264138] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.264155] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:87728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.264171] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.264187] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:87736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.264201] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.264217] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:87744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.264231] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.264247] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:80 nsid:1 lba:87752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.264276] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.264303] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:108 nsid:1 lba:87760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.264318] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.264334] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:87768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.264348] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.264363] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:87776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.264377] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.264392] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:87784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.264407] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.264422] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:87792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.264436] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.264452] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:87800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.264466] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.264481] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:87808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.264495] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.264511] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:87816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.264525] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.264540] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:87824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.264554] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.264569] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:87832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.264583] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.264599] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:87840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.264613] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.264628] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:87136 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.755 [2024-12-06 17:50:41.264642] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.264658] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:87848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.264702] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.264721] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:107 nsid:1 lba:87856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.264736] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.264751] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:87864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.264766] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.264782] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:87872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.264796] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.264812] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:87880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.264826] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.264842] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:87888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.264857] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.264872] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:87896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.264887] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.264902] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:87904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.264916] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.264932] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:87912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.264946] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.264961] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:99 nsid:1 lba:87920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.264975] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.265005] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:87928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.265019] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.265034] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:87936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.265047] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.265062] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:87944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.265075] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.265091] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:87952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.265108] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.265124] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:87960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.265137] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.265152] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:96 nsid:1 lba:87968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.265166] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.755 [2024-12-06 17:50:41.265181] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:87976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.755 [2024-12-06 17:50:41.265196] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.265211] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:87984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.756 [2024-12-06 17:50:41.265225] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.265240] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:87992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.756 [2024-12-06 17:50:41.265254] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.265269] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:88000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.756 [2024-12-06 17:50:41.265283] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.265298] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:88008 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.756 [2024-12-06 17:50:41.265311] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.265327] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:88016 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.756 [2024-12-06 17:50:41.265340] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.265355] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:114 nsid:1 lba:88024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.756 [2024-12-06 17:50:41.265369] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.265384] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:88032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.756 [2024-12-06 17:50:41.265397] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.265413] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:93 nsid:1 lba:88040 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.756 [2024-12-06 17:50:41.265426] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.265441] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:88048 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.756 [2024-12-06 17:50:41.265455] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.265474] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:88056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.756 [2024-12-06 17:50:41.265489] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.265504] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:88064 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.756 [2024-12-06 17:50:41.265518] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.265533] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:88072 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.756 [2024-12-06 17:50:41.265546] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.265561] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:88080 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.756 [2024-12-06 17:50:41.265575] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.265591] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:88088 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.756 [2024-12-06 17:50:41.265604] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.265619] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:88096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.756 [2024-12-06 17:50:41.265634] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.265650] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:88104 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.756 [2024-12-06 17:50:41.265686] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.265705] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:88112 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.756 [2024-12-06 17:50:41.265720] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.265736] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:88120 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.756 [2024-12-06 17:50:41.265750] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.265765] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:88128 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.756 [2024-12-06 17:50:41.265779] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.265795] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:88136 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.756 [2024-12-06 17:50:41.265809] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.265825] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:88144 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.756 [2024-12-06 17:50:41.265839] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.265855] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:87144 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.756 [2024-12-06 17:50:41.265873] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.265889] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:87152 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.756 [2024-12-06 17:50:41.265904] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.265919] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:87160 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.756 [2024-12-06 17:50:41.265933] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.265949] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:119 nsid:1 lba:87168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.756 [2024-12-06 17:50:41.265963] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.265994] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:87176 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.756 [2024-12-06 17:50:41.266008] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.266023] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:79 nsid:1 lba:87184 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.756 [2024-12-06 17:50:41.266036] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.266052] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:87192 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.756 [2024-12-06 17:50:41.266065] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.266080] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:123 nsid:1 lba:87200 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.756 [2024-12-06 17:50:41.266094] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.266109] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:87208 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.756 [2024-12-06 17:50:41.266123] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.266138] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:87216 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.756 [2024-12-06 17:50:41.266152] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.266168] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:87224 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.756 [2024-12-06 17:50:41.266182] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.266197] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:125 nsid:1 lba:87232 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.756 [2024-12-06 17:50:41.266211] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.266227] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:87240 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.756 [2024-12-06 17:50:41.266256] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.266275] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:87248 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.756 [2024-12-06 17:50:41.266290] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.266306] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:87256 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.756 [2024-12-06 17:50:41.266320] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.266336] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:88152 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.756 [2024-12-06 17:50:41.266350] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.266366] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:72 nsid:1 lba:87264 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.756 [2024-12-06 17:50:41.266381] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.266397] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:87272 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.756 [2024-12-06 17:50:41.266411] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.756 [2024-12-06 17:50:41.266427] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:64 nsid:1 lba:87280 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.266441] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.266457] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:104 nsid:1 lba:87288 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.266471] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.266487] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:87296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.266501] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.266517] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:68 nsid:1 lba:87304 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.266531] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.266546] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:87312 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.266560] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.266576] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:87320 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.266590] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.266606] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:117 nsid:1 lba:87328 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.266620] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.266636] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:87336 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.266651] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.266681] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:87344 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.266697] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.266713] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:87352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.266728] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.266744] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:77 nsid:1 lba:87360 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.266758] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.266774] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:87368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.266788] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.266805] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:87376 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.266819] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.266835] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:115 nsid:1 lba:87384 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.266849] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.266865] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:87392 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.266879] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.266895] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:98 nsid:1 lba:87400 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.266909] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.266925] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:87408 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.266939] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.266955] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:87416 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.266969] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.266985] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:87424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.266999] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.267017] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:87432 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.267031] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.267047] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:87440 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.267069] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.267086] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:87448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.267100] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.267116] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:87 nsid:1 lba:87456 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.267130] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.267146] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:87464 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.267161] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.267177] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:87472 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.267191] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.267207] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:87480 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.267222] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.267238] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:87488 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.267252] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.267268] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:74 nsid:1 lba:87496 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.267282] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.267298] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:87504 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.267311] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.267327] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:76 nsid:1 lba:87512 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.267341] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.267357] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:87520 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.267371] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.267387] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:87528 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.267400] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.267416] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:87536 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.267430] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.267449] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:87544 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.267464] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.267479] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:110 nsid:1 lba:87552 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.267494] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.267509] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:87560 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.267523] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.267539] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:87568 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.267553] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.267568] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:87576 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.267582] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.267599] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:87584 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.267613] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.267628] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:78 nsid:1 lba:87592 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.757 [2024-12-06 17:50:41.267642] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.757 [2024-12-06 17:50:41.267658] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:87600 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:41.267681] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:41.267698] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:87608 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:41.267713] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:41.267728] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:87616 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:41.267742] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:41.267757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:87624 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:41.267771] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:41.267786] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:124 nsid:1 lba:87632 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:41.267800] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:41.267815] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:87640 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:41.267833] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:41.267850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:87648 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:41.267864] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:41.267879] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:87656 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:41.267892] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:41.267908] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:87664 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:41.267922] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:41.267937] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:87672 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:41.267951] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:41.267966] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:87680 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:41.267981] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:41.267996] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:83 nsid:1 lba:87688 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:41.268010] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:41.268026] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:87696 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:41.268039] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:41.268054] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x157e900 is same with the state(6) to be set 00:35:10.758 [2024-12-06 17:50:41.268071] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:35:10.758 [2024-12-06 17:50:41.268083] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:35:10.758 [2024-12-06 17:50:41.268095] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:87704 len:8 PRP1 0x0 PRP2 0x0 00:35:10.758 [2024-12-06 17:50:41.268108] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:41.268171] bdev_nvme.c:2056:bdev_nvme_failover_trid: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 3] Start failover from 10.0.0.2:4421 to 10.0.0.2:4422 00:35:10.758 [2024-12-06 17:50:41.268209] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:35:10.758 [2024-12-06 17:50:41.268228] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:41.268244] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:35:10.758 [2024-12-06 17:50:41.268258] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:41.268272] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:35:10.758 [2024-12-06 17:50:41.268290] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:41.268305] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:35:10.758 [2024-12-06 17:50:41.268319] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:41.268333] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 3] in failed state. 00:35:10.758 [2024-12-06 17:50:41.271637] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 3] resetting controller 00:35:10.758 [2024-12-06 17:50:41.271685] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x155b900 (9): Bad file descriptor 00:35:10.758 8506.60 IOPS, 33.23 MiB/s [2024-12-06T16:50:52.597Z] [2024-12-06 17:50:41.430303] bdev_nvme.c:2286:bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 5] Resetting controller successful. 00:35:10.758 8324.00 IOPS, 32.52 MiB/s [2024-12-06T16:50:52.597Z] 8370.43 IOPS, 32.70 MiB/s [2024-12-06T16:50:52.597Z] 8422.75 IOPS, 32.90 MiB/s [2024-12-06T16:50:52.597Z] 8449.44 IOPS, 33.01 MiB/s [2024-12-06T16:50:52.597Z] [2024-12-06 17:50:45.871007] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:76 nsid:1 lba:54344 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:45.871050] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:45.871080] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:54352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:45.871097] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:45.871114] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:54360 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:45.871128] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:45.871145] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:54368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:45.871160] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:45.871176] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:54376 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:45.871191] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:45.871207] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:54384 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:45.871221] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:45.871238] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:78 nsid:1 lba:54392 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:45.871252] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:45.871268] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:98 nsid:1 lba:54400 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:45.871282] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:45.871298] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:54408 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:45.871312] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:45.871334] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:89 nsid:1 lba:54416 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:45.871349] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:45.871365] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:102 nsid:1 lba:54424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:45.871379] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:45.871394] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:54432 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:45.871409] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:45.871434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:74 nsid:1 lba:54440 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:45.871448] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:45.871464] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:54448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:45.871478] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:45.871494] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:54456 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:45.871509] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:45.871524] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:54464 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:45.871538] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:45.871554] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:54472 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:45.871569] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.758 [2024-12-06 17:50:45.871584] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:110 nsid:1 lba:54480 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.758 [2024-12-06 17:50:45.871598] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.871614] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:54488 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.759 [2024-12-06 17:50:45.871628] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.871643] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:54496 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.759 [2024-12-06 17:50:45.871657] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.871684] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:54504 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.759 [2024-12-06 17:50:45.871700] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.871716] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:54512 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.759 [2024-12-06 17:50:45.871734] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.871750] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:54520 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.759 [2024-12-06 17:50:45.871765] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.871781] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:64 nsid:1 lba:54528 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.759 [2024-12-06 17:50:45.871796] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.871811] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:54536 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.759 [2024-12-06 17:50:45.871825] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.871841] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:54544 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.759 [2024-12-06 17:50:45.871855] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.871871] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:68 nsid:1 lba:54552 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.759 [2024-12-06 17:50:45.871885] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.871901] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:54560 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.759 [2024-12-06 17:50:45.871915] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.871931] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:54568 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.759 [2024-12-06 17:50:45.871945] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.871960] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:112 nsid:1 lba:54576 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.759 [2024-12-06 17:50:45.871974] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.871990] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:54584 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.759 [2024-12-06 17:50:45.872004] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.872020] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:54592 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.759 [2024-12-06 17:50:45.872034] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.872049] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:87 nsid:1 lba:54600 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.759 [2024-12-06 17:50:45.872063] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.872079] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:54608 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.759 [2024-12-06 17:50:45.872093] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.872112] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:54616 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.759 [2024-12-06 17:50:45.872127] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.872143] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:86 nsid:1 lba:54624 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.759 [2024-12-06 17:50:45.872157] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.872173] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:80 nsid:1 lba:54640 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.759 [2024-12-06 17:50:45.872187] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.872202] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:54648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.759 [2024-12-06 17:50:45.872216] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.872232] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:54656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.759 [2024-12-06 17:50:45.872246] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.872262] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:54664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.759 [2024-12-06 17:50:45.872276] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.872292] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:105 nsid:1 lba:54672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.759 [2024-12-06 17:50:45.872306] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.872321] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:90 nsid:1 lba:54680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.759 [2024-12-06 17:50:45.872336] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.872352] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:118 nsid:1 lba:54688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.759 [2024-12-06 17:50:45.872366] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.872381] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:54696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.759 [2024-12-06 17:50:45.872395] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.872411] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:54704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.759 [2024-12-06 17:50:45.872425] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.872441] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:54712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.759 [2024-12-06 17:50:45.872455] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.872472] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:54720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.759 [2024-12-06 17:50:45.872486] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.872505] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:54728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.759 [2024-12-06 17:50:45.872519] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.872535] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:54736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.759 [2024-12-06 17:50:45.872549] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.872565] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:54744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.759 [2024-12-06 17:50:45.872579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.872595] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:54752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.759 [2024-12-06 17:50:45.872609] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.872625] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:119 nsid:1 lba:54760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.759 [2024-12-06 17:50:45.872639] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.872655] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:54768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.759 [2024-12-06 17:50:45.872682] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.872700] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:54776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.759 [2024-12-06 17:50:45.872715] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.872731] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:54784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.759 [2024-12-06 17:50:45.872745] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.759 [2024-12-06 17:50:45.872761] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:54792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.872775] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.872790] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:54800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.872805] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.872820] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:54808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.872834] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.872857] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:54816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.872872] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.872888] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:54824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.872907] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.872923] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:54832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.872937] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.872953] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:54840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.872967] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.872983] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:54848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.872997] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873012] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:54856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873026] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873042] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:54864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873056] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873072] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:54872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873086] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873101] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:54880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873115] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873131] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:54888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873145] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873163] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:54896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873178] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873194] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:54904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873209] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873225] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:54912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873240] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873256] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:54920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873270] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873290] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:54928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873305] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873321] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:54936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873336] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873357] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:54944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873372] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873388] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:54952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873402] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873418] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:54960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873432] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873448] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:54968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873463] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873479] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:54976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873493] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873509] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:54984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873523] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873539] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:54992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873554] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873570] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:97 nsid:1 lba:55000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873584] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873600] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:55008 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873614] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873630] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:55016 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873644] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873660] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:55024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873687] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873704] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:55032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873720] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873736] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:55040 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873750] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873766] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:55048 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873781] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873797] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:107 nsid:1 lba:55056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873811] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873827] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:55064 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873842] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873858] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:55072 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873873] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873889] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:54632 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:35:10.760 [2024-12-06 17:50:45.873903] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873919] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:55080 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.760 [2024-12-06 17:50:45.873933] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.760 [2024-12-06 17:50:45.873958] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:55088 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.873972] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.873988] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:100 nsid:1 lba:55096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874002] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874018] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:70 nsid:1 lba:55104 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874033] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874049] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:108 nsid:1 lba:55112 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874064] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874080] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:55120 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874098] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874114] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:55128 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874130] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874146] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:55136 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874160] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874176] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:55144 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874190] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874206] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:55152 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874220] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874236] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:55160 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874250] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874266] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:55168 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874280] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874296] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:55176 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874310] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874325] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:55184 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874339] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874356] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:55192 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874370] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874386] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:55200 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874400] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874415] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:55208 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874430] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874445] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:55216 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874460] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874479] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:114 nsid:1 lba:55224 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874494] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874510] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:96 nsid:1 lba:55232 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874524] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874540] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:55240 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874554] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874569] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:55248 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874584] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874600] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:55256 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874614] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874630] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:120 nsid:1 lba:55264 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874644] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874659] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:55272 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874681] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874697] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:55280 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874712] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874727] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:55288 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874741] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:77 nsid:1 lba:55296 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874771] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874786] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:88 nsid:1 lba:55304 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874800] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874816] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:55312 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874830] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874845] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:93 nsid:1 lba:55320 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874863] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874879] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:55328 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:35:10.761 [2024-12-06 17:50:45.874893] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874925] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:35:10.761 [2024-12-06 17:50:45.874942] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:55336 len:8 PRP1 0x0 PRP2 0x0 00:35:10.761 [2024-12-06 17:50:45.874956] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.874974] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:35:10.761 [2024-12-06 17:50:45.874987] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:35:10.761 [2024-12-06 17:50:45.874998] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:55344 len:8 PRP1 0x0 PRP2 0x0 00:35:10.761 [2024-12-06 17:50:45.875011] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.875025] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:35:10.761 [2024-12-06 17:50:45.875036] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:35:10.761 [2024-12-06 17:50:45.875048] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:55352 len:8 PRP1 0x0 PRP2 0x0 00:35:10.761 [2024-12-06 17:50:45.875061] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.875074] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:35:10.761 [2024-12-06 17:50:45.875085] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:35:10.761 [2024-12-06 17:50:45.875096] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:55360 len:8 PRP1 0x0 PRP2 0x0 00:35:10.761 [2024-12-06 17:50:45.875109] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.875178] bdev_nvme.c:2056:bdev_nvme_failover_trid: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 5] Start failover from 10.0.0.2:4422 to 10.0.0.2:4420 00:35:10.761 [2024-12-06 17:50:45.875216] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:35:10.761 [2024-12-06 17:50:45.875235] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.761 [2024-12-06 17:50:45.875250] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:35:10.761 [2024-12-06 17:50:45.875263] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.762 [2024-12-06 17:50:45.875277] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:35:10.762 [2024-12-06 17:50:45.875291] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.762 [2024-12-06 17:50:45.875304] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:35:10.762 [2024-12-06 17:50:45.875317] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:10.762 [2024-12-06 17:50:45.875331] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 5] in failed state. 00:35:10.762 [2024-12-06 17:50:45.875393] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x155b900 (9): Bad file descriptor 00:35:10.762 [2024-12-06 17:50:45.878658] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 5] resetting controller 00:35:10.762 [2024-12-06 17:50:45.914240] bdev_nvme.c:2286:bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 6] Resetting controller successful. 00:35:10.762 8426.00 IOPS, 32.91 MiB/s [2024-12-06T16:50:52.601Z] 8443.36 IOPS, 32.98 MiB/s [2024-12-06T16:50:52.601Z] 8478.08 IOPS, 33.12 MiB/s [2024-12-06T16:50:52.601Z] 8496.31 IOPS, 33.19 MiB/s [2024-12-06T16:50:52.601Z] 8505.79 IOPS, 33.23 MiB/s [2024-12-06T16:50:52.601Z] 8523.07 IOPS, 33.29 MiB/s 00:35:10.762 Latency(us) 00:35:10.762 [2024-12-06T16:50:52.601Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:35:10.762 Job: NVMe0n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:35:10.762 Verification LBA range: start 0x0 length 0x4000 00:35:10.762 NVMe0n1 : 15.01 8523.57 33.30 611.46 0.00 13983.82 552.20 16311.18 00:35:10.762 [2024-12-06T16:50:52.601Z] =================================================================================================================== 00:35:10.762 [2024-12-06T16:50:52.601Z] Total : 8523.57 33.30 611.46 0.00 13983.82 552.20 16311.18 00:35:10.762 Received shutdown signal, test time was about 15.000000 seconds 00:35:10.762 00:35:10.762 Latency(us) 00:35:10.762 [2024-12-06T16:50:52.601Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:35:10.762 [2024-12-06T16:50:52.601Z] =================================================================================================================== 00:35:10.762 [2024-12-06T16:50:52.601Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:35:10.762 17:50:51 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@65 -- # grep -c 'Resetting controller successful' 00:35:10.762 17:50:51 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@65 -- # count=3 00:35:10.762 17:50:51 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@67 -- # (( count != 3 )) 00:35:10.762 17:50:51 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@73 -- # bdevperf_pid=373393 00:35:10.762 17:50:51 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 1 -f 00:35:10.762 17:50:51 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@75 -- # waitforlisten 373393 /var/tmp/bdevperf.sock 00:35:10.762 17:50:51 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@835 -- # '[' -z 373393 ']' 00:35:10.762 17:50:51 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:35:10.762 17:50:51 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@840 -- # local max_retries=100 00:35:10.762 17:50:51 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:35:10.762 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:35:10.762 17:50:51 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@844 -- # xtrace_disable 00:35:10.762 17:50:51 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:35:10.762 17:50:51 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:35:10.762 17:50:51 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@868 -- # return 0 00:35:10.762 17:50:51 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@76 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:35:10.762 [2024-12-06 17:50:52.062452] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:35:10.762 17:50:52 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@77 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4422 00:35:10.762 [2024-12-06 17:50:52.347294] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4422 *** 00:35:10.762 17:50:52 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@78 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x failover 00:35:11.021 NVMe0n1 00:35:11.021 17:50:52 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x failover 00:35:11.587 00:35:11.587 17:50:53 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@80 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4422 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x failover 00:35:11.846 00:35:11.846 17:50:53 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:35:11.846 17:50:53 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@82 -- # grep -q NVMe0 00:35:12.104 17:50:53 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:35:12.362 17:50:54 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@87 -- # sleep 3 00:35:15.639 17:50:57 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:35:15.639 17:50:57 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@88 -- # grep -q NVMe0 00:35:15.639 17:50:57 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@90 -- # run_test_pid=374062 00:35:15.639 17:50:57 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:35:15.639 17:50:57 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@92 -- # wait 374062 00:35:17.012 { 00:35:17.012 "results": [ 00:35:17.012 { 00:35:17.012 "job": "NVMe0n1", 00:35:17.012 "core_mask": "0x1", 00:35:17.012 "workload": "verify", 00:35:17.012 "status": "finished", 00:35:17.012 "verify_range": { 00:35:17.012 "start": 0, 00:35:17.012 "length": 16384 00:35:17.012 }, 00:35:17.012 "queue_depth": 128, 00:35:17.012 "io_size": 4096, 00:35:17.012 "runtime": 1.009602, 00:35:17.012 "iops": 8544.951376879206, 00:35:17.012 "mibps": 33.3787163159344, 00:35:17.012 "io_failed": 0, 00:35:17.012 "io_timeout": 0, 00:35:17.012 "avg_latency_us": 14906.731595636438, 00:35:17.012 "min_latency_us": 1067.994074074074, 00:35:17.012 "max_latency_us": 13786.832592592593 00:35:17.012 } 00:35:17.012 ], 00:35:17.012 "core_count": 1 00:35:17.012 } 00:35:17.012 17:50:58 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@94 -- # cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:35:17.012 [2024-12-06 17:50:51.594561] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:35:17.012 [2024-12-06 17:50:51.594662] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid373393 ] 00:35:17.012 [2024-12-06 17:50:51.664352] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:35:17.012 [2024-12-06 17:50:51.708295] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:35:17.012 [2024-12-06 17:50:54.035263] bdev_nvme.c:2056:bdev_nvme_failover_trid: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 7] Start failover from 10.0.0.2:4420 to 10.0.0.2:4421 00:35:17.012 [2024-12-06 17:50:54.035358] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:35:17.012 [2024-12-06 17:50:54.035382] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:17.012 [2024-12-06 17:50:54.035399] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:35:17.012 [2024-12-06 17:50:54.035414] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:17.012 [2024-12-06 17:50:54.035428] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:35:17.012 [2024-12-06 17:50:54.035441] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:17.012 [2024-12-06 17:50:54.035455] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:35:17.012 [2024-12-06 17:50:54.035470] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:17.012 [2024-12-06 17:50:54.035489] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 7] in failed state. 00:35:17.012 [2024-12-06 17:50:54.035536] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 7] resetting controller 00:35:17.012 [2024-12-06 17:50:54.035568] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x23e9900 (9): Bad file descriptor 00:35:17.012 [2024-12-06 17:50:54.078945] bdev_nvme.c:2286:bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 10] Resetting controller successful. 00:35:17.012 Running I/O for 1 seconds... 00:35:17.012 8492.00 IOPS, 33.17 MiB/s 00:35:17.012 Latency(us) 00:35:17.012 [2024-12-06T16:50:58.851Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:35:17.012 Job: NVMe0n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:35:17.012 Verification LBA range: start 0x0 length 0x4000 00:35:17.012 NVMe0n1 : 1.01 8544.95 33.38 0.00 0.00 14906.73 1067.99 13786.83 00:35:17.012 [2024-12-06T16:50:58.851Z] =================================================================================================================== 00:35:17.012 [2024-12-06T16:50:58.851Z] Total : 8544.95 33.38 0.00 0.00 14906.73 1067.99 13786.83 00:35:17.012 17:50:58 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@95 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:35:17.012 17:50:58 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@95 -- # grep -q NVMe0 00:35:17.012 17:50:58 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@98 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe0 -t tcp -a 10.0.0.2 -s 4422 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:35:17.270 17:50:59 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@99 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:35:17.270 17:50:59 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@99 -- # grep -q NVMe0 00:35:17.527 17:50:59 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:35:17.784 17:50:59 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@101 -- # sleep 3 00:35:21.066 17:51:02 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@103 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:35:21.066 17:51:02 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@103 -- # grep -q NVMe0 00:35:21.066 17:51:02 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@108 -- # killprocess 373393 00:35:21.066 17:51:02 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@954 -- # '[' -z 373393 ']' 00:35:21.066 17:51:02 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@958 -- # kill -0 373393 00:35:21.066 17:51:02 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@959 -- # uname 00:35:21.066 17:51:02 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:35:21.066 17:51:02 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 373393 00:35:21.066 17:51:02 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:35:21.066 17:51:02 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:35:21.066 17:51:02 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@972 -- # echo 'killing process with pid 373393' 00:35:21.066 killing process with pid 373393 00:35:21.066 17:51:02 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@973 -- # kill 373393 00:35:21.066 17:51:02 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@978 -- # wait 373393 00:35:21.324 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@110 -- # sync 00:35:21.324 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:35:21.582 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@113 -- # trap - SIGINT SIGTERM EXIT 00:35:21.582 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@115 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:35:21.582 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@116 -- # nvmftestfini 00:35:21.582 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@516 -- # nvmfcleanup 00:35:21.582 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@121 -- # sync 00:35:21.582 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:35:21.582 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@124 -- # set +e 00:35:21.582 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@125 -- # for i in {1..20} 00:35:21.582 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:35:21.582 rmmod nvme_tcp 00:35:21.582 rmmod nvme_fabrics 00:35:21.582 rmmod nvme_keyring 00:35:21.582 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:35:21.582 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@128 -- # set -e 00:35:21.582 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@129 -- # return 0 00:35:21.582 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@517 -- # '[' -n 371131 ']' 00:35:21.582 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@518 -- # killprocess 371131 00:35:21.582 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@954 -- # '[' -z 371131 ']' 00:35:21.582 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@958 -- # kill -0 371131 00:35:21.582 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@959 -- # uname 00:35:21.582 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:35:21.582 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 371131 00:35:21.841 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:35:21.841 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:35:21.841 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@972 -- # echo 'killing process with pid 371131' 00:35:21.841 killing process with pid 371131 00:35:21.841 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@973 -- # kill 371131 00:35:21.841 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@978 -- # wait 371131 00:35:21.841 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:35:21.841 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:35:21.841 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:35:21.841 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@297 -- # iptr 00:35:21.841 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@791 -- # iptables-save 00:35:21.841 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:35:21.841 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@791 -- # iptables-restore 00:35:21.841 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:35:21.841 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@302 -- # remove_spdk_ns 00:35:21.841 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:35:21.841 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:35:21.841 17:51:03 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:35:24.377 00:35:24.377 real 0m35.291s 00:35:24.377 user 2m4.228s 00:35:24.377 sys 0m5.905s 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1130 -- # xtrace_disable 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:35:24.377 ************************************ 00:35:24.377 END TEST nvmf_failover 00:35:24.377 ************************************ 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@26 -- # run_test nvmf_host_discovery /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/discovery.sh --transport=tcp 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:35:24.377 ************************************ 00:35:24.377 START TEST nvmf_host_discovery 00:35:24.377 ************************************ 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/discovery.sh --transport=tcp 00:35:24.377 * Looking for test storage... 00:35:24.377 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1711 -- # lcov --version 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@333 -- # local ver1 ver1_l 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@334 -- # local ver2 ver2_l 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@336 -- # IFS=.-: 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@336 -- # read -ra ver1 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@337 -- # IFS=.-: 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@337 -- # read -ra ver2 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@338 -- # local 'op=<' 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@340 -- # ver1_l=2 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@341 -- # ver2_l=1 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@344 -- # case "$op" in 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@345 -- # : 1 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@364 -- # (( v = 0 )) 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@365 -- # decimal 1 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@353 -- # local d=1 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:35:24.377 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@355 -- # echo 1 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@365 -- # ver1[v]=1 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@366 -- # decimal 2 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@353 -- # local d=2 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@355 -- # echo 2 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@366 -- # ver2[v]=2 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@368 -- # return 0 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:35:24.378 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:24.378 --rc genhtml_branch_coverage=1 00:35:24.378 --rc genhtml_function_coverage=1 00:35:24.378 --rc genhtml_legend=1 00:35:24.378 --rc geninfo_all_blocks=1 00:35:24.378 --rc geninfo_unexecuted_blocks=1 00:35:24.378 00:35:24.378 ' 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:35:24.378 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:24.378 --rc genhtml_branch_coverage=1 00:35:24.378 --rc genhtml_function_coverage=1 00:35:24.378 --rc genhtml_legend=1 00:35:24.378 --rc geninfo_all_blocks=1 00:35:24.378 --rc geninfo_unexecuted_blocks=1 00:35:24.378 00:35:24.378 ' 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:35:24.378 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:24.378 --rc genhtml_branch_coverage=1 00:35:24.378 --rc genhtml_function_coverage=1 00:35:24.378 --rc genhtml_legend=1 00:35:24.378 --rc geninfo_all_blocks=1 00:35:24.378 --rc geninfo_unexecuted_blocks=1 00:35:24.378 00:35:24.378 ' 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:35:24.378 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:24.378 --rc genhtml_branch_coverage=1 00:35:24.378 --rc genhtml_function_coverage=1 00:35:24.378 --rc genhtml_legend=1 00:35:24.378 --rc geninfo_all_blocks=1 00:35:24.378 --rc geninfo_unexecuted_blocks=1 00:35:24.378 00:35:24.378 ' 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@7 -- # uname -s 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@15 -- # shopt -s extglob 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- paths/export.sh@5 -- # export PATH 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@51 -- # : 0 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:35:24.378 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@55 -- # have_pci_nics=0 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@11 -- # '[' tcp == rdma ']' 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@16 -- # DISCOVERY_PORT=8009 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@17 -- # DISCOVERY_NQN=nqn.2014-08.org.nvmexpress.discovery 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@20 -- # NQN=nqn.2016-06.io.spdk:cnode 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@22 -- # HOST_NQN=nqn.2021-12.io.spdk:test 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@23 -- # HOST_SOCK=/tmp/host.sock 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@25 -- # nvmftestinit 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@476 -- # prepare_net_devs 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@438 -- # local -g is_hw=no 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@440 -- # remove_spdk_ns 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@309 -- # xtrace_disable 00:35:24.378 17:51:05 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@315 -- # pci_devs=() 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@315 -- # local -a pci_devs 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@316 -- # pci_net_devs=() 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@317 -- # pci_drivers=() 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@317 -- # local -A pci_drivers 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@319 -- # net_devs=() 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@319 -- # local -ga net_devs 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@320 -- # e810=() 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@320 -- # local -ga e810 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@321 -- # x722=() 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@321 -- # local -ga x722 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@322 -- # mlx=() 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@322 -- # local -ga mlx 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:35:26.283 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:35:26.283 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@418 -- # [[ up == up ]] 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:35:26.283 Found net devices under 0000:0a:00.0: cvl_0_0 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:35:26.283 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@418 -- # [[ up == up ]] 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:35:26.284 Found net devices under 0000:0a:00.1: cvl_0_1 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@442 -- # is_hw=yes 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:35:26.284 17:51:07 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:35:26.284 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:35:26.284 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:35:26.284 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.181 ms 00:35:26.284 00:35:26.284 --- 10.0.0.2 ping statistics --- 00:35:26.284 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:35:26.284 rtt min/avg/max/mdev = 0.181/0.181/0.181/0.000 ms 00:35:26.284 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:35:26.284 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:35:26.284 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.091 ms 00:35:26.284 00:35:26.284 --- 10.0.0.1 ping statistics --- 00:35:26.284 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:35:26.284 rtt min/avg/max/mdev = 0.091/0.091/0.091/0.000 ms 00:35:26.284 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:35:26.284 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@450 -- # return 0 00:35:26.284 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:35:26.284 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:35:26.284 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:35:26.284 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:35:26.284 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:35:26.284 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:35:26.284 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:35:26.284 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@30 -- # nvmfappstart -m 0x2 00:35:26.284 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:35:26.284 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@726 -- # xtrace_disable 00:35:26.284 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:26.284 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@509 -- # nvmfpid=377135 00:35:26.284 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:35:26.284 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@510 -- # waitforlisten 377135 00:35:26.284 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@835 -- # '[' -z 377135 ']' 00:35:26.284 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:35:26.284 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@840 -- # local max_retries=100 00:35:26.284 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:35:26.284 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:35:26.284 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@844 -- # xtrace_disable 00:35:26.284 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:26.284 [2024-12-06 17:51:08.087868] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:35:26.284 [2024-12-06 17:51:08.087976] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:35:26.664 [2024-12-06 17:51:08.160761] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:35:26.664 [2024-12-06 17:51:08.204378] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:35:26.664 [2024-12-06 17:51:08.204441] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:35:26.664 [2024-12-06 17:51:08.204469] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:35:26.664 [2024-12-06 17:51:08.204479] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:35:26.664 [2024-12-06 17:51:08.204489] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:35:26.664 [2024-12-06 17:51:08.205130] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@868 -- # return 0 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@732 -- # xtrace_disable 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@32 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:26.664 [2024-12-06 17:51:08.346051] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@33 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2014-08.org.nvmexpress.discovery -t tcp -a 10.0.0.2 -s 8009 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:26.664 [2024-12-06 17:51:08.354249] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 8009 *** 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@35 -- # rpc_cmd bdev_null_create null0 1000 512 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:26.664 null0 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@36 -- # rpc_cmd bdev_null_create null1 1000 512 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:26.664 null1 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@37 -- # rpc_cmd bdev_wait_for_examine 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@45 -- # hostpid=377299 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -m 0x1 -r /tmp/host.sock 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@46 -- # waitforlisten 377299 /tmp/host.sock 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@835 -- # '[' -z 377299 ']' 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@839 -- # local rpc_addr=/tmp/host.sock 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@840 -- # local max_retries=100 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /tmp/host.sock...' 00:35:26.664 Waiting for process to start up and listen on UNIX domain socket /tmp/host.sock... 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@844 -- # xtrace_disable 00:35:26.664 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:26.969 [2024-12-06 17:51:08.433354] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:35:26.969 [2024-12-06 17:51:08.433431] [ DPDK EAL parameters: nvmf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid377299 ] 00:35:26.969 [2024-12-06 17:51:08.500705] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:35:26.969 [2024-12-06 17:51:08.548000] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:35:26.969 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:35:26.969 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@868 -- # return 0 00:35:26.969 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@48 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; kill $hostpid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:35:26.969 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@50 -- # rpc_cmd -s /tmp/host.sock log_set_flag bdev_nvme 00:35:26.969 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:26.969 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:26.969 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:26.969 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@51 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test 00:35:26.969 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:26.969 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:26.969 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@72 -- # notify_id=0 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@83 -- # get_subsystem_names 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@83 -- # [[ '' == '' ]] 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@84 -- # get_bdev_list 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@84 -- # [[ '' == '' ]] 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@86 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@87 -- # get_subsystem_names 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:35:26.970 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@87 -- # [[ '' == '' ]] 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@88 -- # get_bdev_list 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@88 -- # [[ '' == '' ]] 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@90 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 null0 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@91 -- # get_subsystem_names 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@91 -- # [[ '' == '' ]] 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@92 -- # get_bdev_list 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@92 -- # [[ '' == '' ]] 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@96 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:27.292 [2024-12-06 17:51:08.943840] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@97 -- # get_subsystem_names 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@97 -- # [[ '' == '' ]] 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@98 -- # get_bdev_list 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:35:27.292 17:51:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:27.292 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@98 -- # [[ '' == '' ]] 00:35:27.292 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@99 -- # is_notification_count_eq 0 00:35:27.292 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=0 00:35:27.292 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:35:27.292 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:35:27.292 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:35:27.292 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:35:27.292 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:35:27.292 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_notification_count 00:35:27.292 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 0 00:35:27.292 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:35:27.292 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:27.292 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:27.293 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:27.293 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=0 00:35:27.293 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=0 00:35:27.293 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # (( notification_count == expected_count )) 00:35:27.293 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:35:27.293 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@103 -- # rpc_cmd nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode0 nqn.2021-12.io.spdk:test 00:35:27.293 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:27.293 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:27.293 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:27.293 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@105 -- # waitforcondition '[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:35:27.293 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:35:27.293 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:35:27.293 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:35:27.293 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval '[[' '"$(get_subsystem_names)"' == '"nvme0"' ']]' 00:35:27.293 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_subsystem_names 00:35:27.293 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:35:27.293 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:35:27.293 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:27.293 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:27.293 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:35:27.293 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:35:27.293 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:27.595 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # [[ '' == \n\v\m\e\0 ]] 00:35:27.595 17:51:09 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@924 -- # sleep 1 00:35:27.877 [2024-12-06 17:51:09.683042] bdev_nvme.c:7511:discovery_attach_cb: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr attached 00:35:27.877 [2024-12-06 17:51:09.683092] bdev_nvme.c:7597:discovery_poller: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr connected 00:35:27.877 [2024-12-06 17:51:09.683115] bdev_nvme.c:7474:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:35:28.165 [2024-12-06 17:51:09.769368] bdev_nvme.c:7440:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 new subsystem nvme0 00:35:28.165 [2024-12-06 17:51:09.824275] bdev_nvme.c:5656:nvme_ctrlr_create_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr was created to 10.0.0.2:4420 00:35:28.165 [2024-12-06 17:51:09.825275] bdev_nvme.c:1989:bdev_nvme_create_qpair: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Connecting qpair 0x12a10e0:1 started. 00:35:28.165 [2024-12-06 17:51:09.827047] bdev_nvme.c:7330:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme0 done 00:35:28.165 [2024-12-06 17:51:09.827068] bdev_nvme.c:7289:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 found again 00:35:28.165 [2024-12-06 17:51:09.831490] bdev_nvme.c:1791:bdev_nvme_disconnected_qpair_cb: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpair 0x12a10e0 was disconnected and freed. delete nvme_qpair. 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval '[[' '"$(get_subsystem_names)"' == '"nvme0"' ']]' 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_subsystem_names 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@106 -- # waitforcondition '[[ "$(get_bdev_list)" == "nvme0n1" ]]' 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=[[ "$(get_bdev_list)" == "nvme0n1" ]]' 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval '[[' '"$(get_bdev_list)"' == '"nvme0n1"' ']]' 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_bdev_list 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # [[ nvme0n1 == \n\v\m\e\0\n\1 ]] 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@107 -- # waitforcondition '[[ "$(get_subsystem_paths nvme0)" == "$NVMF_PORT" ]]' 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=[[ "$(get_subsystem_paths nvme0)" == "$NVMF_PORT" ]]' 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval '[[' '"$(get_subsystem_paths' 'nvme0)"' == '"$NVMF_PORT"' ']]' 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_subsystem_paths nvme0 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers -n nvme0 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # jq -r '.[].ctrlrs[].trid.trsvcid' 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # sort -n 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # xargs 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # [[ 4420 == \4\4\2\0 ]] 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@108 -- # is_notification_count_eq 1 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=1 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_notification_count 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 0 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:28.425 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:28.683 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=1 00:35:28.683 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=1 00:35:28.683 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # (( notification_count == expected_count )) 00:35:28.683 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:35:28.683 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@111 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 null1 00:35:28.683 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:28.683 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:28.683 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:28.683 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@113 -- # waitforcondition '[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:35:28.683 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:35:28.683 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:35:28.683 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:35:28.683 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval '[[' '"$(get_bdev_list)"' == '"nvme0n1' 'nvme0n2"' ']]' 00:35:28.683 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_bdev_list 00:35:28.683 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:35:28.683 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:28.683 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:35:28.683 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:28.683 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:35:28.683 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:35:28.942 [2024-12-06 17:51:10.534815] bdev_nvme.c:1989:bdev_nvme_create_qpair: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Connecting qpair 0x12a12c0:1 started. 00:35:28.942 [2024-12-06 17:51:10.543390] bdev_nvme.c:1791:bdev_nvme_disconnected_qpair_cb: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpair 0x12a12c0 was disconnected and freed. delete nvme_qpair. 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@114 -- # is_notification_count_eq 1 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=1 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_notification_count 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 1 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=1 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=2 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # (( notification_count == expected_count )) 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@118 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4421 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:28.942 [2024-12-06 17:51:10.600924] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:35:28.942 [2024-12-06 17:51:10.601632] bdev_nvme.c:7493:discovery_aer_cb: *INFO*: Discovery[10.0.0.2:8009] got aer 00:35:28.942 [2024-12-06 17:51:10.601684] bdev_nvme.c:7474:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@120 -- # waitforcondition '[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval '[[' '"$(get_subsystem_names)"' == '"nvme0"' ']]' 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_subsystem_names 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@121 -- # waitforcondition '[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval '[[' '"$(get_bdev_list)"' == '"nvme0n1' 'nvme0n2"' ']]' 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_bdev_list 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@122 -- # waitforcondition '[[ "$(get_subsystem_paths nvme0)" == "$NVMF_PORT $NVMF_SECOND_PORT" ]]' 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=[[ "$(get_subsystem_paths nvme0)" == "$NVMF_PORT $NVMF_SECOND_PORT" ]]' 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:35:28.942 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:35:28.943 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval '[[' '"$(get_subsystem_paths' 'nvme0)"' == '"$NVMF_PORT' '$NVMF_SECOND_PORT"' ']]' 00:35:28.943 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_subsystem_paths nvme0 00:35:28.943 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers -n nvme0 00:35:28.943 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:28.943 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # jq -r '.[].ctrlrs[].trid.trsvcid' 00:35:28.943 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:28.943 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # sort -n 00:35:28.943 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # xargs 00:35:28.943 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:28.943 [2024-12-06 17:51:10.729426] bdev_nvme.c:7435:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 new path for nvme0 00:35:28.943 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # [[ 4420 == \4\4\2\0\ \4\4\2\1 ]] 00:35:28.943 17:51:10 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@924 -- # sleep 1 00:35:29.201 [2024-12-06 17:51:10.949779] bdev_nvme.c:5656:nvme_ctrlr_create_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 2] ctrlr was created to 10.0.0.2:4421 00:35:29.201 [2024-12-06 17:51:10.949845] bdev_nvme.c:7330:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme0 done 00:35:29.201 [2024-12-06 17:51:10.949863] bdev_nvme.c:7289:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 found again 00:35:29.201 [2024-12-06 17:51:10.949872] bdev_nvme.c:7289:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 found again 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval '[[' '"$(get_subsystem_paths' 'nvme0)"' == '"$NVMF_PORT' '$NVMF_SECOND_PORT"' ']]' 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_subsystem_paths nvme0 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers -n nvme0 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # jq -r '.[].ctrlrs[].trid.trsvcid' 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # sort -n 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # xargs 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # [[ 4420 4421 == \4\4\2\0\ \4\4\2\1 ]] 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@123 -- # is_notification_count_eq 0 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=0 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_notification_count 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 2 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=0 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=2 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # (( notification_count == expected_count )) 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@127 -- # rpc_cmd nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:30.139 [2024-12-06 17:51:11.829221] bdev_nvme.c:7493:discovery_aer_cb: *INFO*: Discovery[10.0.0.2:8009] got aer 00:35:30.139 [2024-12-06 17:51:11.829305] bdev_nvme.c:7474:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:35:30.139 [2024-12-06 17:51:11.829618] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:35:30.139 [2024-12-06 17:51:11.829650] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:30.139 [2024-12-06 17:51:11.829690] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:35:30.139 [2024-12-06 17:51:11.829713] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:30.139 [2024-12-06 17:51:11.829727] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:35:30.139 [2024-12-06 17:51:11.829755] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:30.139 [2024-12-06 17:51:11.829770] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:35:30.139 [2024-12-06 17:51:11.829783] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:35:30.139 [2024-12-06 17:51:11.829796] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1273220 is same with the state(6) to be set 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@129 -- # waitforcondition '[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval '[[' '"$(get_subsystem_names)"' == '"nvme0"' ']]' 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_subsystem_names 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:35:30.139 [2024-12-06 17:51:11.839608] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1273220 (9): Bad file descriptor 00:35:30.139 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:30.139 [2024-12-06 17:51:11.849658] bdev_nvme.c:2549:bdev_nvme_reset_destroy_qpairs: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Delete qpairs for reset. 00:35:30.139 [2024-12-06 17:51:11.849691] bdev_nvme.c:2537:bdev_nvme_reset_destroy_qpair_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpairs were deleted. 00:35:30.139 [2024-12-06 17:51:11.849706] bdev_nvme.c:2133:nvme_ctrlr_disconnect: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start disconnecting ctrlr. 00:35:30.139 [2024-12-06 17:51:11.849716] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:35:30.139 [2024-12-06 17:51:11.849765] bdev_nvme.c:2521:bdev_nvme_reconnect_ctrlr: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start reconnecting ctrlr. 00:35:30.139 [2024-12-06 17:51:11.849928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:35:30.139 [2024-12-06 17:51:11.849958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1273220 with addr=10.0.0.2, port=4420 00:35:30.139 [2024-12-06 17:51:11.849981] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1273220 is same with the state(6) to be set 00:35:30.139 [2024-12-06 17:51:11.850004] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1273220 (9): Bad file descriptor 00:35:30.139 [2024-12-06 17:51:11.850025] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:35:30.139 [2024-12-06 17:51:11.850039] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:35:30.139 [2024-12-06 17:51:11.850056] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:35:30.139 [2024-12-06 17:51:11.850069] bdev_nvme.c:2511:bdev_nvme_reconnect_ctrlr_poll: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr could not be connected. 00:35:30.139 [2024-12-06 17:51:11.850080] bdev_nvme.c:2278:bdev_nvme_reset_ctrlr_complete: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Clear pending resets. 00:35:30.139 [2024-12-06 17:51:11.850088] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:35:30.139 [2024-12-06 17:51:11.859796] bdev_nvme.c:2549:bdev_nvme_reset_destroy_qpairs: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Delete qpairs for reset. 00:35:30.139 [2024-12-06 17:51:11.859817] bdev_nvme.c:2537:bdev_nvme_reset_destroy_qpair_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpairs were deleted. 00:35:30.139 [2024-12-06 17:51:11.859826] bdev_nvme.c:2133:nvme_ctrlr_disconnect: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start disconnecting ctrlr. 00:35:30.139 [2024-12-06 17:51:11.859833] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:35:30.139 [2024-12-06 17:51:11.859872] bdev_nvme.c:2521:bdev_nvme_reconnect_ctrlr: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start reconnecting ctrlr. 00:35:30.139 [2024-12-06 17:51:11.860022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:35:30.139 [2024-12-06 17:51:11.860050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1273220 with addr=10.0.0.2, port=4420 00:35:30.139 [2024-12-06 17:51:11.860066] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1273220 is same with the state(6) to be set 00:35:30.139 [2024-12-06 17:51:11.860088] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1273220 (9): Bad file descriptor 00:35:30.139 [2024-12-06 17:51:11.860122] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:35:30.139 [2024-12-06 17:51:11.860140] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:35:30.139 [2024-12-06 17:51:11.860153] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:35:30.140 [2024-12-06 17:51:11.860166] bdev_nvme.c:2511:bdev_nvme_reconnect_ctrlr_poll: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr could not be connected. 00:35:30.140 [2024-12-06 17:51:11.860175] bdev_nvme.c:2278:bdev_nvme_reset_ctrlr_complete: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Clear pending resets. 00:35:30.140 [2024-12-06 17:51:11.860183] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:35:30.140 [2024-12-06 17:51:11.869907] bdev_nvme.c:2549:bdev_nvme_reset_destroy_qpairs: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Delete qpairs for reset. 00:35:30.140 [2024-12-06 17:51:11.869929] bdev_nvme.c:2537:bdev_nvme_reset_destroy_qpair_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpairs were deleted. 00:35:30.140 [2024-12-06 17:51:11.869938] bdev_nvme.c:2133:nvme_ctrlr_disconnect: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start disconnecting ctrlr. 00:35:30.140 [2024-12-06 17:51:11.869962] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:35:30.140 [2024-12-06 17:51:11.869989] bdev_nvme.c:2521:bdev_nvme_reconnect_ctrlr: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start reconnecting ctrlr. 00:35:30.140 [2024-12-06 17:51:11.870143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:35:30.140 [2024-12-06 17:51:11.870185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1273220 with addr=10.0.0.2, port=4420 00:35:30.140 [2024-12-06 17:51:11.870201] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1273220 is same with the state(6) to be set 00:35:30.140 [2024-12-06 17:51:11.870221] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1273220 (9): Bad file descriptor 00:35:30.140 [2024-12-06 17:51:11.870241] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:35:30.140 [2024-12-06 17:51:11.870254] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:35:30.140 [2024-12-06 17:51:11.870267] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:35:30.140 [2024-12-06 17:51:11.870279] bdev_nvme.c:2511:bdev_nvme_reconnect_ctrlr_poll: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr could not be connected. 00:35:30.140 [2024-12-06 17:51:11.870288] bdev_nvme.c:2278:bdev_nvme_reset_ctrlr_complete: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Clear pending resets. 00:35:30.140 [2024-12-06 17:51:11.870295] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:35:30.140 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:35:30.140 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:35:30.140 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@130 -- # waitforcondition '[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:35:30.140 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:35:30.140 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:35:30.140 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:35:30.140 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval '[[' '"$(get_bdev_list)"' == '"nvme0n1' 'nvme0n2"' ']]' 00:35:30.140 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_bdev_list 00:35:30.140 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:35:30.140 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:35:30.140 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:30.140 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:30.140 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:35:30.140 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:35:30.140 [2024-12-06 17:51:11.880023] bdev_nvme.c:2549:bdev_nvme_reset_destroy_qpairs: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Delete qpairs for reset. 00:35:30.140 [2024-12-06 17:51:11.880055] bdev_nvme.c:2537:bdev_nvme_reset_destroy_qpair_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpairs were deleted. 00:35:30.140 [2024-12-06 17:51:11.880064] bdev_nvme.c:2133:nvme_ctrlr_disconnect: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start disconnecting ctrlr. 00:35:30.140 [2024-12-06 17:51:11.880072] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:35:30.140 [2024-12-06 17:51:11.880111] bdev_nvme.c:2521:bdev_nvme_reconnect_ctrlr: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start reconnecting ctrlr. 00:35:30.140 [2024-12-06 17:51:11.880282] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:35:30.140 [2024-12-06 17:51:11.880310] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1273220 with addr=10.0.0.2, port=4420 00:35:30.140 [2024-12-06 17:51:11.880326] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1273220 is same with the state(6) to be set 00:35:30.140 [2024-12-06 17:51:11.880348] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1273220 (9): Bad file descriptor 00:35:30.140 [2024-12-06 17:51:11.880396] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:35:30.140 [2024-12-06 17:51:11.880415] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:35:30.140 [2024-12-06 17:51:11.880429] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:35:30.140 [2024-12-06 17:51:11.880442] bdev_nvme.c:2511:bdev_nvme_reconnect_ctrlr_poll: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr could not be connected. 00:35:30.140 [2024-12-06 17:51:11.880451] bdev_nvme.c:2278:bdev_nvme_reset_ctrlr_complete: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Clear pending resets. 00:35:30.140 [2024-12-06 17:51:11.880459] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:35:30.140 [2024-12-06 17:51:11.890145] bdev_nvme.c:2549:bdev_nvme_reset_destroy_qpairs: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Delete qpairs for reset. 00:35:30.140 [2024-12-06 17:51:11.890167] bdev_nvme.c:2537:bdev_nvme_reset_destroy_qpair_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpairs were deleted. 00:35:30.140 [2024-12-06 17:51:11.890175] bdev_nvme.c:2133:nvme_ctrlr_disconnect: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start disconnecting ctrlr. 00:35:30.140 [2024-12-06 17:51:11.890183] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:35:30.140 [2024-12-06 17:51:11.890222] bdev_nvme.c:2521:bdev_nvme_reconnect_ctrlr: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start reconnecting ctrlr. 00:35:30.140 [2024-12-06 17:51:11.890359] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:35:30.140 [2024-12-06 17:51:11.890386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1273220 with addr=10.0.0.2, port=4420 00:35:30.140 [2024-12-06 17:51:11.890402] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1273220 is same with the state(6) to be set 00:35:30.140 [2024-12-06 17:51:11.890430] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1273220 (9): Bad file descriptor 00:35:30.140 [2024-12-06 17:51:11.890450] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:35:30.140 [2024-12-06 17:51:11.890463] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:35:30.140 [2024-12-06 17:51:11.890476] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:35:30.140 [2024-12-06 17:51:11.890489] bdev_nvme.c:2511:bdev_nvme_reconnect_ctrlr_poll: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr could not be connected. 00:35:30.140 [2024-12-06 17:51:11.890498] bdev_nvme.c:2278:bdev_nvme_reset_ctrlr_complete: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Clear pending resets. 00:35:30.140 [2024-12-06 17:51:11.890506] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:35:30.140 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:30.140 [2024-12-06 17:51:11.900256] bdev_nvme.c:2549:bdev_nvme_reset_destroy_qpairs: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Delete qpairs for reset. 00:35:30.140 [2024-12-06 17:51:11.900276] bdev_nvme.c:2537:bdev_nvme_reset_destroy_qpair_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpairs were deleted. 00:35:30.140 [2024-12-06 17:51:11.900284] bdev_nvme.c:2133:nvme_ctrlr_disconnect: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start disconnecting ctrlr. 00:35:30.140 [2024-12-06 17:51:11.900291] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:35:30.140 [2024-12-06 17:51:11.900330] bdev_nvme.c:2521:bdev_nvme_reconnect_ctrlr: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start reconnecting ctrlr. 00:35:30.140 [2024-12-06 17:51:11.900457] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:35:30.140 [2024-12-06 17:51:11.900484] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1273220 with addr=10.0.0.2, port=4420 00:35:30.140 [2024-12-06 17:51:11.900500] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1273220 is same with the state(6) to be set 00:35:30.140 [2024-12-06 17:51:11.900521] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1273220 (9): Bad file descriptor 00:35:30.140 [2024-12-06 17:51:11.900554] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:35:30.140 [2024-12-06 17:51:11.900572] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:35:30.140 [2024-12-06 17:51:11.900585] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:35:30.140 [2024-12-06 17:51:11.900597] bdev_nvme.c:2511:bdev_nvme_reconnect_ctrlr_poll: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr could not be connected. 00:35:30.140 [2024-12-06 17:51:11.900607] bdev_nvme.c:2278:bdev_nvme_reset_ctrlr_complete: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Clear pending resets. 00:35:30.140 [2024-12-06 17:51:11.900614] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:35:30.140 [2024-12-06 17:51:11.910364] bdev_nvme.c:2549:bdev_nvme_reset_destroy_qpairs: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Delete qpairs for reset. 00:35:30.140 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:35:30.140 [2024-12-06 17:51:11.910385] bdev_nvme.c:2537:bdev_nvme_reset_destroy_qpair_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpairs were deleted. 00:35:30.140 [2024-12-06 17:51:11.910409] bdev_nvme.c:2133:nvme_ctrlr_disconnect: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start disconnecting ctrlr. 00:35:30.140 [2024-12-06 17:51:11.910417] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:35:30.140 [2024-12-06 17:51:11.910439] bdev_nvme.c:2521:bdev_nvme_reconnect_ctrlr: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start reconnecting ctrlr. 00:35:30.140 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:35:30.140 [2024-12-06 17:51:11.910614] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:35:30.140 [2024-12-06 17:51:11.910642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1273220 with addr=10.0.0.2, port=4420 00:35:30.140 [2024-12-06 17:51:11.910679] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1273220 is same with the state(6) to be set 00:35:30.140 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@131 -- # waitforcondition '[[ "$(get_subsystem_paths nvme0)" == "$NVMF_SECOND_PORT" ]]' 00:35:30.140 [2024-12-06 17:51:11.910701] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1273220 (9): Bad file descriptor 00:35:30.140 [2024-12-06 17:51:11.910721] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:35:30.141 [2024-12-06 17:51:11.910734] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:35:30.141 [2024-12-06 17:51:11.910747] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:35:30.141 [2024-12-06 17:51:11.910759] bdev_nvme.c:2511:bdev_nvme_reconnect_ctrlr_poll: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr could not be connected. 00:35:30.141 [2024-12-06 17:51:11.910768] bdev_nvme.c:2278:bdev_nvme_reset_ctrlr_complete: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Clear pending resets. 00:35:30.141 [2024-12-06 17:51:11.910779] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:35:30.141 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=[[ "$(get_subsystem_paths nvme0)" == "$NVMF_SECOND_PORT" ]]' 00:35:30.141 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:35:30.141 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:35:30.141 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval '[[' '"$(get_subsystem_paths' 'nvme0)"' == '"$NVMF_SECOND_PORT"' ']]' 00:35:30.141 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_subsystem_paths nvme0 00:35:30.141 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers -n nvme0 00:35:30.141 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:30.141 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # jq -r '.[].ctrlrs[].trid.trsvcid' 00:35:30.141 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:30.141 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # sort -n 00:35:30.141 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # xargs 00:35:30.141 [2024-12-06 17:51:11.915111] bdev_nvme.c:7298:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 not found 00:35:30.141 [2024-12-06 17:51:11.915140] bdev_nvme.c:7289:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 found again 00:35:30.141 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:30.141 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # [[ 4421 == \4\4\2\1 ]] 00:35:30.141 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:35:30.141 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@132 -- # is_notification_count_eq 0 00:35:30.141 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=0 00:35:30.141 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:35:30.141 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:35:30.141 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:35:30.141 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:35:30.141 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:35:30.141 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_notification_count 00:35:30.141 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 2 00:35:30.141 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:35:30.141 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:30.141 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:30.141 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:30.400 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=0 00:35:30.400 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=2 00:35:30.400 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # (( notification_count == expected_count )) 00:35:30.400 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:35:30.400 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@134 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_stop_discovery -b nvme 00:35:30.400 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:30.400 17:51:11 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@136 -- # waitforcondition '[[ "$(get_subsystem_names)" == "" ]]' 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=[[ "$(get_subsystem_names)" == "" ]]' 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval '[[' '"$(get_subsystem_names)"' == '""' ']]' 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_subsystem_names 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # [[ '' == '' ]] 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@137 -- # waitforcondition '[[ "$(get_bdev_list)" == "" ]]' 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=[[ "$(get_bdev_list)" == "" ]]' 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval '[[' '"$(get_bdev_list)"' == '""' ']]' 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_bdev_list 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # [[ '' == '' ]] 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@138 -- # is_notification_count_eq 2 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=2 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@919 -- # local max=10 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # (( max-- )) 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # get_notification_count 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 2 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=2 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=4 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@921 -- # (( notification_count == expected_count )) 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@922 -- # return 0 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@141 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:30.401 17:51:12 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:31.336 [2024-12-06 17:51:13.141924] bdev_nvme.c:7511:discovery_attach_cb: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr attached 00:35:31.336 [2024-12-06 17:51:13.141967] bdev_nvme.c:7597:discovery_poller: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr connected 00:35:31.336 [2024-12-06 17:51:13.141988] bdev_nvme.c:7474:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:35:31.593 [2024-12-06 17:51:13.228254] bdev_nvme.c:7440:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 new subsystem nvme0 00:35:31.593 [2024-12-06 17:51:13.327083] bdev_nvme.c:5656:nvme_ctrlr_create_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 3] ctrlr was created to 10.0.0.2:4421 00:35:31.593 [2024-12-06 17:51:13.327820] bdev_nvme.c:1989:bdev_nvme_create_qpair: *INFO*: [nqn.2016-06.io.spdk:cnode0, 3] Connecting qpair 0x126e590:1 started. 00:35:31.593 [2024-12-06 17:51:13.329840] bdev_nvme.c:7330:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme0 done 00:35:31.593 [2024-12-06 17:51:13.329874] bdev_nvme.c:7289:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 found again 00:35:31.593 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:31.593 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@143 -- # NOT rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:35:31.593 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@652 -- # local es=0 00:35:31.593 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:35:31.593 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:35:31.593 [2024-12-06 17:51:13.331455] bdev_nvme.c:1791:bdev_nvme_disconnected_qpair_cb: *INFO*: [nqn.2016-06.io.spdk:cnode0, 3] qpair 0x126e590 was disconnected and freed. delete nvme_qpair. 00:35:31.593 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:35:31.593 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:35:31.593 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:35:31.593 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@655 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:35:31.593 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:31.593 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:31.593 request: 00:35:31.593 { 00:35:31.593 "name": "nvme", 00:35:31.593 "trtype": "tcp", 00:35:31.593 "traddr": "10.0.0.2", 00:35:31.593 "adrfam": "ipv4", 00:35:31.593 "trsvcid": "8009", 00:35:31.593 "hostnqn": "nqn.2021-12.io.spdk:test", 00:35:31.593 "wait_for_attach": true, 00:35:31.593 "method": "bdev_nvme_start_discovery", 00:35:31.593 "req_id": 1 00:35:31.593 } 00:35:31.593 Got JSON-RPC error response 00:35:31.593 response: 00:35:31.593 { 00:35:31.593 "code": -17, 00:35:31.593 "message": "File exists" 00:35:31.593 } 00:35:31.593 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:35:31.593 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@655 -- # es=1 00:35:31.593 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:35:31.593 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@145 -- # get_discovery_ctrlrs 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_discovery_info 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # jq -r '.[].name' 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # sort 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # xargs 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@145 -- # [[ nvme == \n\v\m\e ]] 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@146 -- # get_bdev_list 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@146 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@149 -- # NOT rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@652 -- # local es=0 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@655 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:31.594 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:31.851 request: 00:35:31.851 { 00:35:31.851 "name": "nvme_second", 00:35:31.851 "trtype": "tcp", 00:35:31.851 "traddr": "10.0.0.2", 00:35:31.851 "adrfam": "ipv4", 00:35:31.851 "trsvcid": "8009", 00:35:31.851 "hostnqn": "nqn.2021-12.io.spdk:test", 00:35:31.851 "wait_for_attach": true, 00:35:31.851 "method": "bdev_nvme_start_discovery", 00:35:31.851 "req_id": 1 00:35:31.851 } 00:35:31.851 Got JSON-RPC error response 00:35:31.851 response: 00:35:31.851 { 00:35:31.851 "code": -17, 00:35:31.851 "message": "File exists" 00:35:31.851 } 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@655 -- # es=1 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@151 -- # get_discovery_ctrlrs 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_discovery_info 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # jq -r '.[].name' 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # sort 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # xargs 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@151 -- # [[ nvme == \n\v\m\e ]] 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@152 -- # get_bdev_list 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@152 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@155 -- # NOT rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8010 -f ipv4 -q nqn.2021-12.io.spdk:test -T 3000 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@652 -- # local es=0 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8010 -f ipv4 -q nqn.2021-12.io.spdk:test -T 3000 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@655 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8010 -f ipv4 -q nqn.2021-12.io.spdk:test -T 3000 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:31.851 17:51:13 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:32.781 [2024-12-06 17:51:14.525158] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:35:32.781 [2024-12-06 17:51:14.525219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x12a5570 with addr=10.0.0.2, port=8010 00:35:32.781 [2024-12-06 17:51:14.525243] nvme_tcp.c:2612:nvme_tcp_ctrlr_construct: *ERROR*: failed to create admin qpair 00:35:32.781 [2024-12-06 17:51:14.525257] nvme.c: 842:nvme_probe_internal: *ERROR*: NVMe ctrlr scan failed 00:35:32.781 [2024-12-06 17:51:14.525270] bdev_nvme.c:7579:discovery_poller: *ERROR*: Discovery[10.0.0.2:8010] could not start discovery connect 00:35:33.712 [2024-12-06 17:51:15.527843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:35:33.712 [2024-12-06 17:51:15.527912] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x12a5570 with addr=10.0.0.2, port=8010 00:35:33.712 [2024-12-06 17:51:15.527943] nvme_tcp.c:2612:nvme_tcp_ctrlr_construct: *ERROR*: failed to create admin qpair 00:35:33.712 [2024-12-06 17:51:15.527958] nvme.c: 842:nvme_probe_internal: *ERROR*: NVMe ctrlr scan failed 00:35:33.712 [2024-12-06 17:51:15.527986] bdev_nvme.c:7579:discovery_poller: *ERROR*: Discovery[10.0.0.2:8010] could not start discovery connect 00:35:35.082 [2024-12-06 17:51:16.529868] bdev_nvme.c:7554:discovery_poller: *ERROR*: Discovery[10.0.0.2:8010] timed out while attaching discovery ctrlr 00:35:35.082 request: 00:35:35.082 { 00:35:35.082 "name": "nvme_second", 00:35:35.082 "trtype": "tcp", 00:35:35.082 "traddr": "10.0.0.2", 00:35:35.082 "adrfam": "ipv4", 00:35:35.082 "trsvcid": "8010", 00:35:35.082 "hostnqn": "nqn.2021-12.io.spdk:test", 00:35:35.082 "wait_for_attach": false, 00:35:35.082 "attach_timeout_ms": 3000, 00:35:35.082 "method": "bdev_nvme_start_discovery", 00:35:35.082 "req_id": 1 00:35:35.082 } 00:35:35.082 Got JSON-RPC error response 00:35:35.082 response: 00:35:35.082 { 00:35:35.082 "code": -110, 00:35:35.082 "message": "Connection timed out" 00:35:35.082 } 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@655 -- # es=1 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@157 -- # get_discovery_ctrlrs 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_discovery_info 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@563 -- # xtrace_disable 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # jq -r '.[].name' 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # sort 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # xargs 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@157 -- # [[ nvme == \n\v\m\e ]] 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@159 -- # trap - SIGINT SIGTERM EXIT 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@161 -- # kill 377299 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@162 -- # nvmftestfini 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@516 -- # nvmfcleanup 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@121 -- # sync 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@124 -- # set +e 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@125 -- # for i in {1..20} 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:35:35.082 rmmod nvme_tcp 00:35:35.082 rmmod nvme_fabrics 00:35:35.082 rmmod nvme_keyring 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@128 -- # set -e 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@129 -- # return 0 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@517 -- # '[' -n 377135 ']' 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@518 -- # killprocess 377135 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@954 -- # '[' -z 377135 ']' 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@958 -- # kill -0 377135 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@959 -- # uname 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 377135 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@972 -- # echo 'killing process with pid 377135' 00:35:35.082 killing process with pid 377135 00:35:35.082 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@973 -- # kill 377135 00:35:35.083 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@978 -- # wait 377135 00:35:35.083 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:35:35.083 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:35:35.083 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:35:35.083 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@297 -- # iptr 00:35:35.083 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@791 -- # iptables-save 00:35:35.083 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:35:35.083 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@791 -- # iptables-restore 00:35:35.083 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:35:35.083 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@302 -- # remove_spdk_ns 00:35:35.083 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:35:35.083 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:35:35.083 17:51:16 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:35:37.618 17:51:18 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:35:37.618 00:35:37.618 real 0m13.200s 00:35:37.618 user 0m19.019s 00:35:37.618 sys 0m2.819s 00:35:37.618 17:51:18 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1130 -- # xtrace_disable 00:35:37.618 17:51:18 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:35:37.618 ************************************ 00:35:37.618 END TEST nvmf_host_discovery 00:35:37.618 ************************************ 00:35:37.618 17:51:18 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@27 -- # run_test nvmf_host_multipath_status /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/multipath_status.sh --transport=tcp 00:35:37.618 17:51:18 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:35:37.618 17:51:18 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:35:37.618 17:51:18 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:35:37.618 ************************************ 00:35:37.618 START TEST nvmf_host_multipath_status 00:35:37.618 ************************************ 00:35:37.618 17:51:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/multipath_status.sh --transport=tcp 00:35:37.618 * Looking for test storage... 00:35:37.618 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1711 -- # lcov --version 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@333 -- # local ver1 ver1_l 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@334 -- # local ver2 ver2_l 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@336 -- # IFS=.-: 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@336 -- # read -ra ver1 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@337 -- # IFS=.-: 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@337 -- # read -ra ver2 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@338 -- # local 'op=<' 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@340 -- # ver1_l=2 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@341 -- # ver2_l=1 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@344 -- # case "$op" in 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@345 -- # : 1 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@364 -- # (( v = 0 )) 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@365 -- # decimal 1 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@353 -- # local d=1 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@355 -- # echo 1 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@365 -- # ver1[v]=1 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@366 -- # decimal 2 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@353 -- # local d=2 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@355 -- # echo 2 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@366 -- # ver2[v]=2 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@368 -- # return 0 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:35:37.618 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:37.618 --rc genhtml_branch_coverage=1 00:35:37.618 --rc genhtml_function_coverage=1 00:35:37.618 --rc genhtml_legend=1 00:35:37.618 --rc geninfo_all_blocks=1 00:35:37.618 --rc geninfo_unexecuted_blocks=1 00:35:37.618 00:35:37.618 ' 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:35:37.618 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:37.618 --rc genhtml_branch_coverage=1 00:35:37.618 --rc genhtml_function_coverage=1 00:35:37.618 --rc genhtml_legend=1 00:35:37.618 --rc geninfo_all_blocks=1 00:35:37.618 --rc geninfo_unexecuted_blocks=1 00:35:37.618 00:35:37.618 ' 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:35:37.618 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:37.618 --rc genhtml_branch_coverage=1 00:35:37.618 --rc genhtml_function_coverage=1 00:35:37.618 --rc genhtml_legend=1 00:35:37.618 --rc geninfo_all_blocks=1 00:35:37.618 --rc geninfo_unexecuted_blocks=1 00:35:37.618 00:35:37.618 ' 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:35:37.618 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:37.618 --rc genhtml_branch_coverage=1 00:35:37.618 --rc genhtml_function_coverage=1 00:35:37.618 --rc genhtml_legend=1 00:35:37.618 --rc geninfo_all_blocks=1 00:35:37.618 --rc geninfo_unexecuted_blocks=1 00:35:37.618 00:35:37.618 ' 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@7 -- # uname -s 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:35:37.618 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@15 -- # shopt -s extglob 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- paths/export.sh@5 -- # export PATH 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@51 -- # : 0 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:35:37.619 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@55 -- # have_pci_nics=0 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@12 -- # MALLOC_BDEV_SIZE=64 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@15 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@16 -- # bpf_sh=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/bpftrace.sh 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@18 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@21 -- # NQN=nqn.2016-06.io.spdk:cnode1 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@31 -- # nvmftestinit 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@476 -- # prepare_net_devs 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@438 -- # local -g is_hw=no 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@440 -- # remove_spdk_ns 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@309 -- # xtrace_disable 00:35:37.619 17:51:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@315 -- # pci_devs=() 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@315 -- # local -a pci_devs 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@316 -- # pci_net_devs=() 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@317 -- # pci_drivers=() 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@317 -- # local -A pci_drivers 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@319 -- # net_devs=() 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@319 -- # local -ga net_devs 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@320 -- # e810=() 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@320 -- # local -ga e810 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@321 -- # x722=() 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@321 -- # local -ga x722 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@322 -- # mlx=() 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@322 -- # local -ga mlx 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:35:39.521 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:35:39.521 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:35:39.521 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@418 -- # [[ up == up ]] 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:35:39.522 Found net devices under 0000:0a:00.0: cvl_0_0 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@418 -- # [[ up == up ]] 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:35:39.522 Found net devices under 0000:0a:00.1: cvl_0_1 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@442 -- # is_hw=yes 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:35:39.522 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:35:39.780 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:35:39.780 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:35:39.780 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:35:39.780 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:35:39.780 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:35:39.780 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:35:39.780 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:35:39.780 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:35:39.780 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:35:39.780 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:35:39.780 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:35:39.780 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.189 ms 00:35:39.780 00:35:39.780 --- 10.0.0.2 ping statistics --- 00:35:39.780 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:35:39.780 rtt min/avg/max/mdev = 0.189/0.189/0.189/0.000 ms 00:35:39.780 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:35:39.781 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:35:39.781 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.098 ms 00:35:39.781 00:35:39.781 --- 10.0.0.1 ping statistics --- 00:35:39.781 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:35:39.781 rtt min/avg/max/mdev = 0.098/0.098/0.098/0.000 ms 00:35:39.781 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:35:39.781 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@450 -- # return 0 00:35:39.781 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:35:39.781 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:35:39.781 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:35:39.781 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:35:39.781 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:35:39.781 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:35:39.781 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:35:39.781 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@33 -- # nvmfappstart -m 0x3 00:35:39.781 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:35:39.781 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@726 -- # xtrace_disable 00:35:39.781 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:35:39.781 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@509 -- # nvmfpid=380474 00:35:39.781 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x3 00:35:39.781 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@510 -- # waitforlisten 380474 00:35:39.781 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@835 -- # '[' -z 380474 ']' 00:35:39.781 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:35:39.781 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@840 -- # local max_retries=100 00:35:39.781 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:35:39.781 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:35:39.781 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@844 -- # xtrace_disable 00:35:39.781 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:35:39.781 [2024-12-06 17:51:21.565698] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:35:39.781 [2024-12-06 17:51:21.565788] [ DPDK EAL parameters: nvmf -c 0x3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:35:40.039 [2024-12-06 17:51:21.639848] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:35:40.039 [2024-12-06 17:51:21.683051] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:35:40.039 [2024-12-06 17:51:21.683118] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:35:40.039 [2024-12-06 17:51:21.683145] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:35:40.039 [2024-12-06 17:51:21.683156] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:35:40.039 [2024-12-06 17:51:21.683165] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:35:40.039 [2024-12-06 17:51:21.684532] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:35:40.039 [2024-12-06 17:51:21.684547] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:35:40.039 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:35:40.039 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@868 -- # return 0 00:35:40.039 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:35:40.039 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@732 -- # xtrace_disable 00:35:40.039 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:35:40.039 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:35:40.039 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@34 -- # nvmfapp_pid=380474 00:35:40.039 17:51:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:35:40.297 [2024-12-06 17:51:22.079890] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:35:40.297 17:51:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc0 00:35:40.555 Malloc0 00:35:40.555 17:51:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@39 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -r -m 2 00:35:40.813 17:51:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:35:41.379 17:51:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:35:41.379 [2024-12-06 17:51:23.181023] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:35:41.379 17:51:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:35:41.636 [2024-12-06 17:51:23.441612] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:35:41.637 17:51:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@45 -- # bdevperf_pid=380759 00:35:41.637 17:51:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 90 00:35:41.637 17:51:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@47 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:35:41.637 17:51:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@48 -- # waitforlisten 380759 /var/tmp/bdevperf.sock 00:35:41.637 17:51:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@835 -- # '[' -z 380759 ']' 00:35:41.637 17:51:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:35:41.637 17:51:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@840 -- # local max_retries=100 00:35:41.637 17:51:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:35:41.637 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:35:41.637 17:51:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@844 -- # xtrace_disable 00:35:41.637 17:51:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:35:42.203 17:51:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:35:42.203 17:51:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@868 -- # return 0 00:35:42.203 17:51:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_set_options -r -1 00:35:42.203 17:51:24 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x multipath -l -1 -o 10 00:35:42.768 Nvme0n1 00:35:42.768 17:51:24 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x multipath -l -1 -o 10 00:35:43.334 Nvme0n1 00:35:43.334 17:51:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@78 -- # sleep 2 00:35:43.334 17:51:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@76 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -t 120 -s /var/tmp/bdevperf.sock perform_tests 00:35:45.232 17:51:27 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@90 -- # set_ANA_state optimized optimized 00:35:45.232 17:51:27 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n optimized 00:35:45.798 17:51:27 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n optimized 00:35:46.056 17:51:27 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@91 -- # sleep 1 00:35:46.991 17:51:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@92 -- # check_status true false true true true true 00:35:46.991 17:51:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:35:46.991 17:51:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:35:46.991 17:51:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:35:47.250 17:51:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:35:47.250 17:51:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current false 00:35:47.250 17:51:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:35:47.250 17:51:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:35:47.508 17:51:29 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:35:47.508 17:51:29 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:35:47.508 17:51:29 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:35:47.508 17:51:29 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:35:47.766 17:51:29 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:35:47.766 17:51:29 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:35:47.766 17:51:29 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:35:47.766 17:51:29 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:35:48.024 17:51:29 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:35:48.024 17:51:29 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:35:48.024 17:51:29 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:35:48.024 17:51:29 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:35:48.283 17:51:29 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:35:48.283 17:51:29 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:35:48.283 17:51:29 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:35:48.283 17:51:29 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:35:48.540 17:51:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:35:48.540 17:51:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@94 -- # set_ANA_state non_optimized optimized 00:35:48.540 17:51:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:35:48.798 17:51:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n optimized 00:35:49.056 17:51:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@95 -- # sleep 1 00:35:49.989 17:51:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@96 -- # check_status false true true true true true 00:35:49.989 17:51:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current false 00:35:49.989 17:51:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:35:49.989 17:51:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:35:50.555 17:51:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:35:50.555 17:51:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current true 00:35:50.555 17:51:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:35:50.555 17:51:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:35:50.555 17:51:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:35:50.555 17:51:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:35:50.555 17:51:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:35:50.555 17:51:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:35:50.813 17:51:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:35:50.813 17:51:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:35:50.813 17:51:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:35:50.813 17:51:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:35:51.071 17:51:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:35:51.071 17:51:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:35:51.329 17:51:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:35:51.329 17:51:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:35:51.588 17:51:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:35:51.588 17:51:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:35:51.588 17:51:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:35:51.588 17:51:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:35:51.847 17:51:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:35:51.847 17:51:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@100 -- # set_ANA_state non_optimized non_optimized 00:35:51.847 17:51:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:35:52.105 17:51:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n non_optimized 00:35:52.373 17:51:34 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@101 -- # sleep 1 00:35:53.316 17:51:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@102 -- # check_status true false true true true true 00:35:53.316 17:51:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:35:53.316 17:51:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:35:53.316 17:51:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:35:53.574 17:51:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:35:53.574 17:51:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current false 00:35:53.574 17:51:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:35:53.574 17:51:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:35:53.832 17:51:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:35:53.832 17:51:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:35:53.832 17:51:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:35:53.832 17:51:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:35:54.090 17:51:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:35:54.090 17:51:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:35:54.090 17:51:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:35:54.090 17:51:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:35:54.347 17:51:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:35:54.347 17:51:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:35:54.347 17:51:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:35:54.347 17:51:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:35:54.604 17:51:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:35:54.604 17:51:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:35:54.604 17:51:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:35:54.604 17:51:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:35:54.861 17:51:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:35:54.861 17:51:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@104 -- # set_ANA_state non_optimized inaccessible 00:35:54.861 17:51:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:35:55.118 17:51:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n inaccessible 00:35:55.681 17:51:37 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@105 -- # sleep 1 00:35:56.610 17:51:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@106 -- # check_status true false true true true false 00:35:56.610 17:51:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:35:56.610 17:51:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:35:56.610 17:51:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:35:56.868 17:51:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:35:56.868 17:51:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current false 00:35:56.868 17:51:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:35:56.868 17:51:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:35:57.126 17:51:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:35:57.126 17:51:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:35:57.126 17:51:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:35:57.126 17:51:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:35:57.384 17:51:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:35:57.384 17:51:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:35:57.384 17:51:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:35:57.384 17:51:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:35:57.643 17:51:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:35:57.643 17:51:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:35:57.643 17:51:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:35:57.643 17:51:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:35:57.901 17:51:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:35:57.901 17:51:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible false 00:35:57.901 17:51:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:35:57.901 17:51:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:35:58.158 17:51:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:35:58.158 17:51:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@108 -- # set_ANA_state inaccessible inaccessible 00:35:58.159 17:51:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n inaccessible 00:35:58.417 17:51:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n inaccessible 00:35:58.675 17:51:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@109 -- # sleep 1 00:36:00.046 17:51:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@110 -- # check_status false false true true false false 00:36:00.046 17:51:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current false 00:36:00.046 17:51:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:00.046 17:51:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:36:00.046 17:51:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:36:00.046 17:51:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current false 00:36:00.046 17:51:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:00.046 17:51:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:36:00.304 17:51:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:36:00.304 17:51:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:36:00.304 17:51:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:00.304 17:51:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:36:00.562 17:51:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:36:00.562 17:51:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:36:00.562 17:51:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:00.562 17:51:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:36:00.820 17:51:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:36:00.820 17:51:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible false 00:36:00.820 17:51:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:00.820 17:51:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:36:01.078 17:51:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:36:01.078 17:51:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible false 00:36:01.078 17:51:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:01.078 17:51:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:36:01.337 17:51:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:36:01.337 17:51:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@112 -- # set_ANA_state inaccessible optimized 00:36:01.337 17:51:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n inaccessible 00:36:01.594 17:51:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n optimized 00:36:01.852 17:51:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@113 -- # sleep 1 00:36:02.786 17:51:44 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@114 -- # check_status false true true true false true 00:36:02.786 17:51:44 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current false 00:36:02.786 17:51:44 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:02.786 17:51:44 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:36:03.352 17:51:44 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:36:03.352 17:51:44 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current true 00:36:03.352 17:51:44 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:03.352 17:51:44 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:36:03.352 17:51:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:36:03.352 17:51:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:36:03.352 17:51:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:03.352 17:51:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:36:03.610 17:51:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:36:03.610 17:51:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:36:03.867 17:51:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:03.867 17:51:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:36:04.125 17:51:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:36:04.125 17:51:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible false 00:36:04.125 17:51:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:04.125 17:51:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:36:04.383 17:51:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:36:04.383 17:51:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:36:04.383 17:51:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:04.383 17:51:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:36:04.641 17:51:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:36:04.641 17:51:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@116 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_set_multipath_policy -b Nvme0n1 -p active_active 00:36:04.899 17:51:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@119 -- # set_ANA_state optimized optimized 00:36:04.899 17:51:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n optimized 00:36:05.156 17:51:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n optimized 00:36:05.414 17:51:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@120 -- # sleep 1 00:36:06.348 17:51:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@121 -- # check_status true true true true true true 00:36:06.348 17:51:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:36:06.348 17:51:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:06.348 17:51:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:36:06.606 17:51:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:36:06.606 17:51:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current true 00:36:06.606 17:51:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:06.606 17:51:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:36:06.864 17:51:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:36:06.864 17:51:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:36:06.864 17:51:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:06.864 17:51:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:36:07.430 17:51:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:36:07.430 17:51:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:36:07.430 17:51:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:07.430 17:51:48 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:36:07.688 17:51:49 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:36:07.688 17:51:49 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:36:07.688 17:51:49 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:07.688 17:51:49 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:36:07.947 17:51:49 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:36:07.947 17:51:49 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:36:07.947 17:51:49 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:07.947 17:51:49 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:36:08.205 17:51:49 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:36:08.205 17:51:49 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@123 -- # set_ANA_state non_optimized optimized 00:36:08.205 17:51:49 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:36:08.464 17:51:50 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n optimized 00:36:08.722 17:51:50 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@124 -- # sleep 1 00:36:09.656 17:51:51 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@125 -- # check_status false true true true true true 00:36:09.656 17:51:51 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current false 00:36:09.656 17:51:51 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:09.656 17:51:51 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:36:09.914 17:51:51 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:36:09.914 17:51:51 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current true 00:36:09.914 17:51:51 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:09.914 17:51:51 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:36:10.172 17:51:51 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:36:10.172 17:51:51 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:36:10.172 17:51:51 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:10.172 17:51:51 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:36:10.430 17:51:52 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:36:10.430 17:51:52 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:36:10.430 17:51:52 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:10.430 17:51:52 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:36:10.689 17:51:52 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:36:10.689 17:51:52 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:36:10.689 17:51:52 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:10.689 17:51:52 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:36:10.947 17:51:52 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:36:10.947 17:51:52 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:36:10.947 17:51:52 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:10.947 17:51:52 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:36:11.205 17:51:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:36:11.205 17:51:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@129 -- # set_ANA_state non_optimized non_optimized 00:36:11.205 17:51:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:36:11.771 17:51:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n non_optimized 00:36:11.771 17:51:53 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@130 -- # sleep 1 00:36:13.147 17:51:54 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@131 -- # check_status true true true true true true 00:36:13.147 17:51:54 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:36:13.147 17:51:54 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:13.147 17:51:54 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:36:13.147 17:51:54 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:36:13.147 17:51:54 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current true 00:36:13.147 17:51:54 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:13.147 17:51:54 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:36:13.405 17:51:55 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:36:13.405 17:51:55 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:36:13.405 17:51:55 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:13.405 17:51:55 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:36:13.661 17:51:55 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:36:13.661 17:51:55 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:36:13.661 17:51:55 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:13.661 17:51:55 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:36:13.919 17:51:55 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:36:13.919 17:51:55 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:36:13.919 17:51:55 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:13.919 17:51:55 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:36:14.176 17:51:55 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:36:14.176 17:51:55 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:36:14.176 17:51:55 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:14.176 17:51:55 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:36:14.434 17:51:56 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:36:14.434 17:51:56 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@133 -- # set_ANA_state non_optimized inaccessible 00:36:14.434 17:51:56 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:36:14.691 17:51:56 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n inaccessible 00:36:15.255 17:51:56 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@134 -- # sleep 1 00:36:16.183 17:51:57 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@135 -- # check_status true false true true true false 00:36:16.183 17:51:57 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:36:16.183 17:51:57 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:16.183 17:51:57 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:36:16.439 17:51:58 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:36:16.439 17:51:58 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current false 00:36:16.439 17:51:58 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:16.439 17:51:58 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:36:16.696 17:51:58 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:36:16.696 17:51:58 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:36:16.696 17:51:58 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:16.696 17:51:58 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:36:16.953 17:51:58 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:36:16.953 17:51:58 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:36:16.953 17:51:58 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:16.953 17:51:58 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:36:17.210 17:51:58 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:36:17.210 17:51:58 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:36:17.210 17:51:58 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:17.210 17:51:58 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:36:17.467 17:51:59 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:36:17.467 17:51:59 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible false 00:36:17.467 17:51:59 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:36:17.467 17:51:59 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:36:17.725 17:51:59 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:36:17.725 17:51:59 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@137 -- # killprocess 380759 00:36:17.725 17:51:59 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@954 -- # '[' -z 380759 ']' 00:36:17.725 17:51:59 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@958 -- # kill -0 380759 00:36:17.725 17:51:59 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@959 -- # uname 00:36:17.725 17:51:59 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:36:17.725 17:51:59 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 380759 00:36:17.725 17:51:59 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@960 -- # process_name=reactor_2 00:36:17.725 17:51:59 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@964 -- # '[' reactor_2 = sudo ']' 00:36:17.725 17:51:59 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@972 -- # echo 'killing process with pid 380759' 00:36:17.725 killing process with pid 380759 00:36:17.725 17:51:59 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@973 -- # kill 380759 00:36:17.725 17:51:59 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@978 -- # wait 380759 00:36:17.725 { 00:36:17.725 "results": [ 00:36:17.725 { 00:36:17.725 "job": "Nvme0n1", 00:36:17.725 "core_mask": "0x4", 00:36:17.725 "workload": "verify", 00:36:17.725 "status": "terminated", 00:36:17.725 "verify_range": { 00:36:17.725 "start": 0, 00:36:17.725 "length": 16384 00:36:17.725 }, 00:36:17.725 "queue_depth": 128, 00:36:17.725 "io_size": 4096, 00:36:17.725 "runtime": 34.251692, 00:36:17.725 "iops": 7987.371835528592, 00:36:17.725 "mibps": 31.200671232533562, 00:36:17.725 "io_failed": 0, 00:36:17.725 "io_timeout": 0, 00:36:17.725 "avg_latency_us": 15997.745278119948, 00:36:17.725 "min_latency_us": 458.1451851851852, 00:36:17.725 "max_latency_us": 4026531.84 00:36:17.725 } 00:36:17.725 ], 00:36:17.725 "core_count": 1 00:36:17.725 } 00:36:17.995 17:51:59 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@139 -- # wait 380759 00:36:17.995 17:51:59 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@141 -- # cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:36:17.995 [2024-12-06 17:51:23.509237] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:36:17.995 [2024-12-06 17:51:23.509310] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid380759 ] 00:36:17.995 [2024-12-06 17:51:23.577870] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:36:17.995 [2024-12-06 17:51:23.627515] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:36:17.995 Running I/O for 90 seconds... 00:36:17.995 8347.00 IOPS, 32.61 MiB/s [2024-12-06T16:51:59.834Z] 8427.00 IOPS, 32.92 MiB/s [2024-12-06T16:51:59.834Z] 8436.67 IOPS, 32.96 MiB/s [2024-12-06T16:51:59.834Z] 8481.50 IOPS, 33.13 MiB/s [2024-12-06T16:51:59.834Z] 8466.80 IOPS, 33.07 MiB/s [2024-12-06T16:51:59.834Z] 8483.17 IOPS, 33.14 MiB/s [2024-12-06T16:51:59.834Z] 8491.00 IOPS, 33.17 MiB/s [2024-12-06T16:51:59.834Z] 8504.12 IOPS, 33.22 MiB/s [2024-12-06T16:51:59.834Z] 8518.00 IOPS, 33.27 MiB/s [2024-12-06T16:51:59.834Z] 8502.60 IOPS, 33.21 MiB/s [2024-12-06T16:51:59.834Z] 8490.91 IOPS, 33.17 MiB/s [2024-12-06T16:51:59.834Z] 8494.08 IOPS, 33.18 MiB/s [2024-12-06T16:51:59.834Z] 8486.46 IOPS, 33.15 MiB/s [2024-12-06T16:51:59.834Z] 8483.71 IOPS, 33.14 MiB/s [2024-12-06T16:51:59.834Z] [2024-12-06 17:51:40.162250] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:93 nsid:1 lba:96208 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:17.995 [2024-12-06 17:51:40.162313] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:93 cdw0:0 sqhd:000a p:0 m:0 dnr:0 00:36:17.995 [2024-12-06 17:51:40.162376] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:96232 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.995 [2024-12-06 17:51:40.162398] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:000b p:0 m:0 dnr:0 00:36:17.995 [2024-12-06 17:51:40.162423] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:70 nsid:1 lba:96240 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.995 [2024-12-06 17:51:40.162440] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:70 cdw0:0 sqhd:000c p:0 m:0 dnr:0 00:36:17.995 [2024-12-06 17:51:40.162463] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:96248 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.995 [2024-12-06 17:51:40.162481] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:11 cdw0:0 sqhd:000d p:0 m:0 dnr:0 00:36:17.995 [2024-12-06 17:51:40.162520] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:96256 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.995 [2024-12-06 17:51:40.162536] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:72 cdw0:0 sqhd:000e p:0 m:0 dnr:0 00:36:17.995 [2024-12-06 17:51:40.162575] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:96264 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.995 [2024-12-06 17:51:40.162592] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:48 cdw0:0 sqhd:000f p:0 m:0 dnr:0 00:36:17.995 [2024-12-06 17:51:40.162615] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:96272 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.995 [2024-12-06 17:51:40.162632] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:71 cdw0:0 sqhd:0010 p:0 m:0 dnr:0 00:36:17.995 [2024-12-06 17:51:40.162673] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:96280 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.162692] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:60 cdw0:0 sqhd:0011 p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.162715] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:96288 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.162732] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:49 cdw0:0 sqhd:0012 p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.162755] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:96296 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.162782] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:125 cdw0:0 sqhd:0013 p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.162805] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:96304 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.162822] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:87 cdw0:0 sqhd:0014 p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.162845] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:96312 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.162862] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:9 cdw0:0 sqhd:0015 p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.162884] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:96320 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.162901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:84 cdw0:0 sqhd:0016 p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.162923] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:96328 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.162939] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:0017 p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.162962] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:96336 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.162979] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:44 cdw0:0 sqhd:0018 p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.163001] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:96344 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.163017] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:65 cdw0:0 sqhd:0019 p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.163039] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:96352 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.163057] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:85 cdw0:0 sqhd:001a p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.163080] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:96360 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.163097] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:001b p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.163134] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:96368 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.163150] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:73 cdw0:0 sqhd:001c p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.163187] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:96376 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.163204] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:001d p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.163226] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:96384 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.163243] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:22 cdw0:0 sqhd:001e p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.163265] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:96392 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.163285] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:001f p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.163309] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:96400 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.163326] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:76 cdw0:0 sqhd:0020 p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.163348] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:96 nsid:1 lba:96408 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.163364] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:96 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.163387] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:96416 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.163404] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:68 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.163426] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:96424 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.163442] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:53 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.163464] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:96432 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.163481] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:20 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.163503] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:99 nsid:1 lba:96440 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.163520] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:99 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.163542] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:96448 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.163559] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:25 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.163581] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:96456 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.163598] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:32 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.163621] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:96464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.163639] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:123 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.163662] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:96472 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.163687] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:0029 p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.163720] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:96480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.163737] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:002a p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.163759] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:96488 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.163775] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:002b p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.163802] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:96496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.163819] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:27 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.163842] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:118 nsid:1 lba:96504 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.163858] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:118 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.163900] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:96512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.163927] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.163965] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:96520 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.163983] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:51 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.164005] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:96528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.164022] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:0030 p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.164045] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:96536 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.164062] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:30 cdw0:0 sqhd:0031 p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.164534] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:96544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.164558] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:14 cdw0:0 sqhd:0032 p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.164588] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:96552 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.164607] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:50 cdw0:0 sqhd:0033 p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.164632] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:96560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.164648] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:59 cdw0:0 sqhd:0034 p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.164682] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:96568 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.164712] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:0035 p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.164737] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:96576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.996 [2024-12-06 17:51:40.164754] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:13 cdw0:0 sqhd:0036 p:0 m:0 dnr:0 00:36:17.996 [2024-12-06 17:51:40.164779] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:96584 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.164795] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:43 cdw0:0 sqhd:0037 p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.164826] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:96592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.164844] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:38 cdw0:0 sqhd:0038 p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.164884] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:88 nsid:1 lba:96600 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.164901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:88 cdw0:0 sqhd:0039 p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.164925] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:96608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.164941] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:117 cdw0:0 sqhd:003a p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.164972] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:96616 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.164988] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:47 cdw0:0 sqhd:003b p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.165012] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:96624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.165043] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:3 cdw0:0 sqhd:003c p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.165069] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:96632 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.165085] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:003d p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.165109] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:96640 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.165126] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:003e p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.165151] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:96648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.165167] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:39 cdw0:0 sqhd:003f p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.165192] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:96656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.165208] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:0040 p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.165233] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:96664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.165249] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:91 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.165274] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:96672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.165290] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:12 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.165314] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:105 nsid:1 lba:96680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.165346] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:105 cdw0:0 sqhd:0043 p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.165371] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:96688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.165410] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:0044 p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.165437] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:96696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.165454] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:116 cdw0:0 sqhd:0045 p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.165478] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:74 nsid:1 lba:96704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.165495] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:74 cdw0:0 sqhd:0046 p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.165520] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:96712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.165536] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:24 cdw0:0 sqhd:0047 p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.165561] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:97 nsid:1 lba:96720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.165578] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:97 cdw0:0 sqhd:0048 p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.165603] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:96728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.165620] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:82 cdw0:0 sqhd:0049 p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.165645] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:96736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.165661] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:17 cdw0:0 sqhd:004a p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.165695] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:96744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.165713] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:83 cdw0:0 sqhd:004b p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.165737] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:120 nsid:1 lba:96752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.165754] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:120 cdw0:0 sqhd:004c p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.165778] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:69 nsid:1 lba:96760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.165795] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:69 cdw0:0 sqhd:004d p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.165819] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:96768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.165836] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:8 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.165862] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:96776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.165878] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:62 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.165903] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:96784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.165924] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:5 cdw0:0 sqhd:0050 p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.165950] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:96792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.165966] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:35 cdw0:0 sqhd:0051 p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.165996] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:96800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.166024] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:36 cdw0:0 sqhd:0052 p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.166054] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:96808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.166072] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:7 cdw0:0 sqhd:0053 p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.166097] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:108 nsid:1 lba:96816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.166113] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:108 cdw0:0 sqhd:0054 p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.166138] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:96824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.166154] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:45 cdw0:0 sqhd:0055 p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.166179] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:96832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.166195] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:104 cdw0:0 sqhd:0056 p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.166220] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:96840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.166238] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:33 cdw0:0 sqhd:0057 p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.166263] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:96848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.166279] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:0058 p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.166320] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:96856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.166336] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:0059 p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.166359] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:96864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.166375] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:124 cdw0:0 sqhd:005a p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.166399] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:96872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.166415] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:89 cdw0:0 sqhd:005b p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.166439] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:96880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.166454] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:16 cdw0:0 sqhd:005c p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.166482] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:113 nsid:1 lba:96888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.997 [2024-12-06 17:51:40.166499] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:113 cdw0:0 sqhd:005d p:0 m:0 dnr:0 00:36:17.997 [2024-12-06 17:51:40.166523] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:96896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.166538] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:29 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.166562] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:96904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.166578] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:2 cdw0:0 sqhd:005f p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.166602] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:96912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.166618] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:86 cdw0:0 sqhd:0060 p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.166641] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:80 nsid:1 lba:96920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.166681] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:80 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.166708] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:96928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.166742] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:112 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.166770] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:96936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.166787] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:56 cdw0:0 sqhd:0063 p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.166812] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:96944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.166828] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:37 cdw0:0 sqhd:0064 p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.166852] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:100 nsid:1 lba:96952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.166869] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:100 cdw0:0 sqhd:0065 p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.166894] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:96960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.166910] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:54 cdw0:0 sqhd:0066 p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.166934] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:107 nsid:1 lba:96968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.166951] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:107 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.166976] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:96976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.167007] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:78 cdw0:0 sqhd:0068 p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.167037] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:96984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.167054] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:0069 p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.167095] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:96216 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:17.998 [2024-12-06 17:51:40.167117] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:006a p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.167286] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:96224 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:17.998 [2024-12-06 17:51:40.167308] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:30 cdw0:0 sqhd:006b p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.167340] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:96992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.167358] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:006c p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.167386] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:97000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.167404] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:51 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.167433] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:97008 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.167450] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.167478] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:118 nsid:1 lba:97016 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.167495] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:118 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.167524] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:97024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.167541] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:27 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.167568] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:97032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.167585] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:0071 p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.167628] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:97040 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.167645] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:0072 p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.167681] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:97048 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.167699] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:0073 p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.167736] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:97056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.167752] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:123 cdw0:0 sqhd:0074 p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.167780] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:97064 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.167801] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:32 cdw0:0 sqhd:0075 p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.167829] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:97072 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.167846] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:25 cdw0:0 sqhd:0076 p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.167873] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:99 nsid:1 lba:97080 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.167889] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:99 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.167917] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:97088 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.167934] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:20 cdw0:0 sqhd:0078 p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.167962] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:97096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.167978] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:53 cdw0:0 sqhd:0079 p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.168005] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:97104 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.168022] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:68 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.168050] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:96 nsid:1 lba:97112 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.168067] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:96 cdw0:0 sqhd:007b p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.168094] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:97120 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.168110] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:76 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.168138] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:97128 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.168169] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:007d p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.168198] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:97136 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.168223] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:22 cdw0:0 sqhd:007e p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.168261] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:97144 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.168279] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.168307] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:97152 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.168324] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:73 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.168352] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:97160 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.168373] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.168401] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:97168 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.168418] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:85 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.168446] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:97176 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.998 [2024-12-06 17:51:40.168463] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:65 cdw0:0 sqhd:0003 p:0 m:0 dnr:0 00:36:17.998 [2024-12-06 17:51:40.168491] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:97184 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:40.168523] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:44 cdw0:0 sqhd:0004 p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:40.168551] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:97192 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:40.168567] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:0005 p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:40.168595] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:97200 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:40.168611] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:84 cdw0:0 sqhd:0006 p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:40.168638] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:97208 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:40.168655] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:9 cdw0:0 sqhd:0007 p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:40.168693] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:97216 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:40.168711] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:87 cdw0:0 sqhd:0008 p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:40.168740] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:97224 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:40.168756] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:125 cdw0:0 sqhd:0009 p:0 m:0 dnr:0 00:36:17.999 8447.60 IOPS, 33.00 MiB/s [2024-12-06T16:51:59.838Z] 7919.62 IOPS, 30.94 MiB/s [2024-12-06T16:51:59.838Z] 7453.76 IOPS, 29.12 MiB/s [2024-12-06T16:51:59.838Z] 7039.67 IOPS, 27.50 MiB/s [2024-12-06T16:51:59.838Z] 6691.53 IOPS, 26.14 MiB/s [2024-12-06T16:51:59.838Z] 6783.10 IOPS, 26.50 MiB/s [2024-12-06T16:51:59.838Z] 6865.38 IOPS, 26.82 MiB/s [2024-12-06T16:51:59.838Z] 6980.64 IOPS, 27.27 MiB/s [2024-12-06T16:51:59.838Z] 7155.87 IOPS, 27.95 MiB/s [2024-12-06T16:51:59.838Z] 7326.67 IOPS, 28.62 MiB/s [2024-12-06T16:51:59.838Z] 7477.72 IOPS, 29.21 MiB/s [2024-12-06T16:51:59.838Z] 7520.04 IOPS, 29.38 MiB/s [2024-12-06T16:51:59.838Z] 7559.04 IOPS, 29.53 MiB/s [2024-12-06T16:51:59.838Z] 7599.64 IOPS, 29.69 MiB/s [2024-12-06T16:51:59.838Z] 7688.21 IOPS, 30.03 MiB/s [2024-12-06T16:51:59.838Z] 7798.40 IOPS, 30.46 MiB/s [2024-12-06T16:51:59.838Z] 7907.71 IOPS, 30.89 MiB/s [2024-12-06T16:51:59.838Z] [2024-12-06 17:51:56.774444] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:39256 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:17.999 [2024-12-06 17:51:56.774512] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:47 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.774562] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:39304 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:56.774581] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:126 cdw0:0 sqhd:0029 p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.774616] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:39320 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:56.774633] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:002a p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.774655] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:39336 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:56.774697] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:002b p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.774732] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:39352 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:56.774750] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.774772] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:39368 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:56.774788] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:25 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.774810] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:39384 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:56.774826] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:8 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.774847] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:39400 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:56.774863] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:43 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.774884] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:39416 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:56.774900] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:0030 p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.774921] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:39432 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:56.774937] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:34 cdw0:0 sqhd:0031 p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.774959] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:39448 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:56.774979] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:112 cdw0:0 sqhd:0032 p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.775000] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:39464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:56.775016] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:21 cdw0:0 sqhd:0033 p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.775038] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:39480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:56.775054] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:0034 p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.775075] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:39496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:56.775091] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:62 cdw0:0 sqhd:0035 p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.775113] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:39512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:56.775134] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:0036 p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.775157] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:89 nsid:1 lba:39224 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:17.999 [2024-12-06 17:51:56.775173] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:89 cdw0:0 sqhd:0037 p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.775196] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:39248 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:17.999 [2024-12-06 17:51:56.775212] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:91 cdw0:0 sqhd:0038 p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.777805] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:39528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:56.777836] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:66 cdw0:0 sqhd:0039 p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.777866] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:39544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:56.777885] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:73 cdw0:0 sqhd:003a p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.777907] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:39560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:56.777925] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:1 cdw0:0 sqhd:003b p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.777947] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:39576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:56.777965] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:003c p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.777987] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:39592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:56.778004] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:123 cdw0:0 sqhd:003d p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.778032] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:39608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:56.778049] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:003e p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.778071] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:39624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:56.778088] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:003f p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.778110] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:39640 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:56.778127] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:0040 p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.778149] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:39656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:56.778165] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.778187] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:39672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:56.778209] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:45 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.778238] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:39688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:56.778255] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:76 cdw0:0 sqhd:0043 p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.778507] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:39704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:56.778529] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:0044 p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.778561] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:39720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:17.999 [2024-12-06 17:51:56.778579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:0045 p:0 m:0 dnr:0 00:36:17.999 [2024-12-06 17:51:56.778602] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:39736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.778620] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:0046 p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.778642] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:108 nsid:1 lba:39752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.778658] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:108 cdw0:0 sqhd:0047 p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.778695] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:39768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.778712] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:103 cdw0:0 sqhd:0048 p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.778742] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:39784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.778759] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:41 cdw0:0 sqhd:0049 p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.778781] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:39800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.778798] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:28 cdw0:0 sqhd:004a p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.778820] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:74 nsid:1 lba:39816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.778837] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:74 cdw0:0 sqhd:004b p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.778859] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:39832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.778875] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:004c p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.778897] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:39848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.778914] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:83 cdw0:0 sqhd:004d p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.778935] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:39864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.778952] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:92 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.778989] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:39880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.779006] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.779027] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:39896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.779044] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:0050 p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.779066] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:39912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.779098] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:0051 p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.779120] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:39928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.779135] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:0052 p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.779157] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:39944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.779175] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:37 cdw0:0 sqhd:0053 p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.779197] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:39272 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.000 [2024-12-06 17:51:56.779213] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:32 cdw0:0 sqhd:0054 p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.779234] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:39960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.779249] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:111 cdw0:0 sqhd:0055 p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.779271] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:39976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.779286] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:91 cdw0:0 sqhd:0056 p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.779308] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:39992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.779324] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:89 cdw0:0 sqhd:0057 p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.779345] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:40008 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.779360] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:0058 p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.779381] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:40024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.779397] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:62 cdw0:0 sqhd:0059 p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.779418] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:40040 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.779434] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:005a p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.779460] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:40056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.779476] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:21 cdw0:0 sqhd:005b p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.779516] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:40072 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.779536] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:112 cdw0:0 sqhd:005c p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.779560] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:40088 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.779577] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:34 cdw0:0 sqhd:005d p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.779599] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:40104 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.779615] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.779641] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:40120 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.779658] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:43 cdw0:0 sqhd:005f p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.779690] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:40136 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.779716] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:8 cdw0:0 sqhd:0060 p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.779739] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:40152 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.779755] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:25 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.779777] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:40168 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.779793] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.779815] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:40184 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.779831] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:0063 p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.779854] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:39256 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.000 [2024-12-06 17:51:56.779870] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:0064 p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.781170] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:39320 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.781195] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:126 cdw0:0 sqhd:0065 p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.781223] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:39352 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.000 [2024-12-06 17:51:56.781242] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:47 cdw0:0 sqhd:0066 p:0 m:0 dnr:0 00:36:18.000 [2024-12-06 17:51:56.781265] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:39384 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.001 [2024-12-06 17:51:56.781287] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:44 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.781310] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:93 nsid:1 lba:39416 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.001 [2024-12-06 17:51:56.781328] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:93 cdw0:0 sqhd:0068 p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.781351] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:39448 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.001 [2024-12-06 17:51:56.781367] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:24 cdw0:0 sqhd:0069 p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.781389] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:39480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.001 [2024-12-06 17:51:56.781405] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:3 cdw0:0 sqhd:006a p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.781427] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:39512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.001 [2024-12-06 17:51:56.781443] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:16 cdw0:0 sqhd:006b p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.781465] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:39248 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.001 [2024-12-06 17:51:56.781481] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:30 cdw0:0 sqhd:006c p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.781504] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:39280 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.001 [2024-12-06 17:51:56.781520] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:84 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.781543] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:40208 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.001 [2024-12-06 17:51:56.781574] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:56 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.781596] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:40224 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.001 [2024-12-06 17:51:56.781611] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:5 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.781633] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:116 nsid:1 lba:39296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.001 [2024-12-06 17:51:56.781648] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:116 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.781694] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:107 nsid:1 lba:39328 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.001 [2024-12-06 17:51:56.781717] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:107 cdw0:0 sqhd:0071 p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.781740] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:39360 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.001 [2024-12-06 17:51:56.781756] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:0072 p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.781778] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:39392 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.001 [2024-12-06 17:51:56.781799] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:58 cdw0:0 sqhd:0073 p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.781822] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:39424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.001 [2024-12-06 17:51:56.781838] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:100 cdw0:0 sqhd:0074 p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.781860] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:39456 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.001 [2024-12-06 17:51:56.781876] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:52 cdw0:0 sqhd:0075 p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.781899] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:39488 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.001 [2024-12-06 17:51:56.781915] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:0076 p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.781936] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:40240 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.001 [2024-12-06 17:51:56.781952] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:65 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.781974] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:40256 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.001 [2024-12-06 17:51:56.782013] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:0078 p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.782035] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:79 nsid:1 lba:39536 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.001 [2024-12-06 17:51:56.782051] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:0079 p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.782072] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:39568 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.001 [2024-12-06 17:51:56.782087] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.782109] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:39600 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.001 [2024-12-06 17:51:56.782125] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:70 cdw0:0 sqhd:007b p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.782146] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:39632 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.001 [2024-12-06 17:51:56.782161] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:7 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.782182] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:115 nsid:1 lba:39664 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.001 [2024-12-06 17:51:56.782198] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:115 cdw0:0 sqhd:007d p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.782219] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:39544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.001 [2024-12-06 17:51:56.782235] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:76 cdw0:0 sqhd:007e p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.782256] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:39576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.001 [2024-12-06 17:51:56.782271] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:45 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.782302] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:39608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.001 [2024-12-06 17:51:56.782335] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.782362] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:39640 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.001 [2024-12-06 17:51:56.782378] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.782401] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:39672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.001 [2024-12-06 17:51:56.782416] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.783760] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:39712 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.001 [2024-12-06 17:51:56.783790] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:0003 p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.783819] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:39744 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.001 [2024-12-06 17:51:56.783837] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:0004 p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.783860] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:39776 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.001 [2024-12-06 17:51:56.783877] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:0005 p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.783899] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:39808 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.001 [2024-12-06 17:51:56.783916] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:25 cdw0:0 sqhd:0006 p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.783937] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:39840 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.001 [2024-12-06 17:51:56.783953] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:8 cdw0:0 sqhd:0007 p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.783976] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:39872 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.001 [2024-12-06 17:51:56.783992] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:43 cdw0:0 sqhd:0008 p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.784014] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:39904 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.001 [2024-12-06 17:51:56.784030] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:0009 p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.784052] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:39936 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.001 [2024-12-06 17:51:56.784068] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:34 cdw0:0 sqhd:000a p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.784090] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:112 nsid:1 lba:39968 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.001 [2024-12-06 17:51:56.784106] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:112 cdw0:0 sqhd:000b p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.784150] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:40000 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.001 [2024-12-06 17:51:56.784166] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:21 cdw0:0 sqhd:000c p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.784188] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:64 nsid:1 lba:40032 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.001 [2024-12-06 17:51:56.784203] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:000d p:0 m:0 dnr:0 00:36:18.001 [2024-12-06 17:51:56.784225] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:40064 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.002 [2024-12-06 17:51:56.784240] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:62 cdw0:0 sqhd:000e p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.784261] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:40096 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.002 [2024-12-06 17:51:56.784277] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:000f p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.784298] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:89 nsid:1 lba:40128 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.002 [2024-12-06 17:51:56.784314] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:89 cdw0:0 sqhd:0010 p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.784335] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:40160 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.002 [2024-12-06 17:51:56.784366] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:91 cdw0:0 sqhd:0011 p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.784388] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:40264 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.002 [2024-12-06 17:51:56.784404] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:111 cdw0:0 sqhd:0012 p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.784426] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:39304 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.002 [2024-12-06 17:51:56.784442] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:32 cdw0:0 sqhd:0013 p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.784464] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:39368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.002 [2024-12-06 17:51:56.784480] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:37 cdw0:0 sqhd:0014 p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.784502] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:106 nsid:1 lba:39432 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.002 [2024-12-06 17:51:56.784518] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:0015 p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.784539] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:39496 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.002 [2024-12-06 17:51:56.784555] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:0016 p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.784577] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:39720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.002 [2024-12-06 17:51:56.784593] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:0017 p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.784615] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:39752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.002 [2024-12-06 17:51:56.784638] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:0018 p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.784661] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:39784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.002 [2024-12-06 17:51:56.784687] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:92 cdw0:0 sqhd:0019 p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.784710] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:39816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.002 [2024-12-06 17:51:56.784727] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:83 cdw0:0 sqhd:001a p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.784749] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:39848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.002 [2024-12-06 17:51:56.784765] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:001b p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.784786] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:74 nsid:1 lba:39880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.002 [2024-12-06 17:51:56.784802] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:74 cdw0:0 sqhd:001c p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.784825] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:39912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.002 [2024-12-06 17:51:56.784841] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:28 cdw0:0 sqhd:001d p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.784862] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:39944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.002 [2024-12-06 17:51:56.784878] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:41 cdw0:0 sqhd:001e p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.784900] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:39960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.002 [2024-12-06 17:51:56.784917] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:103 cdw0:0 sqhd:001f p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.784938] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:108 nsid:1 lba:39992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.002 [2024-12-06 17:51:56.784954] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:108 cdw0:0 sqhd:0020 p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.784976] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:40024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.002 [2024-12-06 17:51:56.784992] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.785014] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:40056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.002 [2024-12-06 17:51:56.785030] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.785052] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:40088 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.002 [2024-12-06 17:51:56.785068] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.785090] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:40120 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.002 [2024-12-06 17:51:56.785110] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.785133] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:40152 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.002 [2024-12-06 17:51:56.785150] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:123 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.785171] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:40184 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.002 [2024-12-06 17:51:56.785187] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.785209] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:40216 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.002 [2024-12-06 17:51:56.785226] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.785248] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:40248 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.002 [2024-12-06 17:51:56.785264] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.785286] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:39352 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.002 [2024-12-06 17:51:56.785302] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:0029 p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.785324] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:39416 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.002 [2024-12-06 17:51:56.785340] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:45 cdw0:0 sqhd:002a p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.785377] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:39480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.002 [2024-12-06 17:51:56.785393] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:76 cdw0:0 sqhd:002b p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.785414] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:115 nsid:1 lba:39248 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.002 [2024-12-06 17:51:56.785430] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:115 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.785451] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:40208 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.002 [2024-12-06 17:51:56.785467] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:7 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.785488] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:39296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.002 [2024-12-06 17:51:56.785504] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:70 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.785526] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:39360 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.002 [2024-12-06 17:51:56.785542] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.785563] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:79 nsid:1 lba:39424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.002 [2024-12-06 17:51:56.785578] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:0030 p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.785604] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:39488 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.002 [2024-12-06 17:51:56.785621] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:0031 p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.785659] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:40256 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.002 [2024-12-06 17:51:56.785684] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:65 cdw0:0 sqhd:0032 p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.785708] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:39568 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.002 [2024-12-06 17:51:56.785725] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:0033 p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.785747] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:39632 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.002 [2024-12-06 17:51:56.785764] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:52 cdw0:0 sqhd:0034 p:0 m:0 dnr:0 00:36:18.002 [2024-12-06 17:51:56.785786] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:100 nsid:1 lba:39544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.003 [2024-12-06 17:51:56.785802] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:100 cdw0:0 sqhd:0035 p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.785824] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:39608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.003 [2024-12-06 17:51:56.785841] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:58 cdw0:0 sqhd:0036 p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.786537] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:39672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.003 [2024-12-06 17:51:56.786561] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:0037 p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.788929] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:107 nsid:1 lba:39560 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.003 [2024-12-06 17:51:56.788970] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:107 cdw0:0 sqhd:0038 p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.789016] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:116 nsid:1 lba:39624 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.003 [2024-12-06 17:51:56.789035] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:116 cdw0:0 sqhd:0039 p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.789059] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:40280 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.003 [2024-12-06 17:51:56.789091] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:5 cdw0:0 sqhd:003a p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.789114] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:40296 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.003 [2024-12-06 17:51:56.789131] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:56 cdw0:0 sqhd:003b p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.789153] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:40312 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.003 [2024-12-06 17:51:56.789170] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:84 cdw0:0 sqhd:003c p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.789198] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:40328 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.003 [2024-12-06 17:51:56.789215] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:30 cdw0:0 sqhd:003d p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.789237] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:40344 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.003 [2024-12-06 17:51:56.789254] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:16 cdw0:0 sqhd:003e p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.789276] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:40360 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.003 [2024-12-06 17:51:56.789297] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:3 cdw0:0 sqhd:003f p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.789320] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:40376 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.003 [2024-12-06 17:51:56.789337] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:24 cdw0:0 sqhd:0040 p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.789359] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:93 nsid:1 lba:39656 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.003 [2024-12-06 17:51:56.789376] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:93 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.789398] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:39744 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.003 [2024-12-06 17:51:56.789414] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:58 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.789436] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:39808 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.003 [2024-12-06 17:51:56.789452] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:100 cdw0:0 sqhd:0043 p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.789474] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:39872 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.003 [2024-12-06 17:51:56.789490] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:52 cdw0:0 sqhd:0044 p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.789512] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:39936 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.003 [2024-12-06 17:51:56.789528] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:0045 p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.789550] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:40000 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.003 [2024-12-06 17:51:56.789567] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:65 cdw0:0 sqhd:0046 p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.789588] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:40064 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.003 [2024-12-06 17:51:56.789604] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:0047 p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.789626] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:79 nsid:1 lba:40128 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.003 [2024-12-06 17:51:56.789642] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:0048 p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.789679] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:40264 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.003 [2024-12-06 17:51:56.789698] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:0049 p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.789722] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:39368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.003 [2024-12-06 17:51:56.789738] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:70 cdw0:0 sqhd:004a p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.789761] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:39496 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.003 [2024-12-06 17:51:56.789777] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:7 cdw0:0 sqhd:004b p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.789799] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:39752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.003 [2024-12-06 17:51:56.789815] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:115 cdw0:0 sqhd:004c p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.789837] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:39816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.003 [2024-12-06 17:51:56.789853] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:76 cdw0:0 sqhd:004d p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.789875] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:39880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.003 [2024-12-06 17:51:56.789891] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:45 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.789913] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:39944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.003 [2024-12-06 17:51:56.789929] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.789967] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:39992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.003 [2024-12-06 17:51:56.789982] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:0050 p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.790004] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:40056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.003 [2024-12-06 17:51:56.790019] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:0051 p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.790041] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:40120 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.003 [2024-12-06 17:51:56.790056] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:0052 p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.790078] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:40184 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.003 [2024-12-06 17:51:56.790093] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:123 cdw0:0 sqhd:0053 p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.790115] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:40248 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.003 [2024-12-06 17:51:56.790131] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:0054 p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.790152] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:39416 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.003 [2024-12-06 17:51:56.790172] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:0055 p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.790195] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:39248 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.003 [2024-12-06 17:51:56.790226] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:0056 p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.790251] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:110 nsid:1 lba:39296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.003 [2024-12-06 17:51:56.790267] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:0057 p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.790289] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:39424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.003 [2024-12-06 17:51:56.790306] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:108 cdw0:0 sqhd:0058 p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.790328] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:40256 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.003 [2024-12-06 17:51:56.790344] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:103 cdw0:0 sqhd:0059 p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.790366] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:39632 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.003 [2024-12-06 17:51:56.790383] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:41 cdw0:0 sqhd:005a p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.790405] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:39608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.003 [2024-12-06 17:51:56.790423] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:28 cdw0:0 sqhd:005b p:0 m:0 dnr:0 00:36:18.003 [2024-12-06 17:51:56.790965] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:39704 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.004 [2024-12-06 17:51:56.790989] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:005c p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.791016] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:74 nsid:1 lba:39768 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.004 [2024-12-06 17:51:56.791034] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:74 cdw0:0 sqhd:005d p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.791057] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:39832 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.004 [2024-12-06 17:51:56.791074] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.791097] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:83 nsid:1 lba:39896 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.004 [2024-12-06 17:51:56.791113] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:83 cdw0:0 sqhd:005f p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.791135] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:39976 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.004 [2024-12-06 17:51:56.791152] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:92 cdw0:0 sqhd:0060 p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.791174] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:40040 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.004 [2024-12-06 17:51:56.791196] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.791219] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:40104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.004 [2024-12-06 17:51:56.791236] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.791258] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:40168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.004 [2024-12-06 17:51:56.791275] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:0063 p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.791297] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:40400 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.004 [2024-12-06 17:51:56.791314] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:0064 p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.791336] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:40416 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.004 [2024-12-06 17:51:56.791353] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:37 cdw0:0 sqhd:0065 p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.791375] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:40432 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.004 [2024-12-06 17:51:56.791392] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:32 cdw0:0 sqhd:0066 p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.791414] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:40448 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.004 [2024-12-06 17:51:56.791439] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:111 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.791461] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:40464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.004 [2024-12-06 17:51:56.791477] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:91 cdw0:0 sqhd:0068 p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.791500] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:40480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.004 [2024-12-06 17:51:56.791517] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:89 cdw0:0 sqhd:0069 p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.791539] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:40496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.004 [2024-12-06 17:51:56.791555] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:006a p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.791577] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:40512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.004 [2024-12-06 17:51:56.791593] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:62 cdw0:0 sqhd:006b p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.791616] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:40528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.004 [2024-12-06 17:51:56.791632] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:006c p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.791654] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:40544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.004 [2024-12-06 17:51:56.791679] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:21 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.791723] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:40560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.004 [2024-12-06 17:51:56.791744] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:112 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.791768] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:39384 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.004 [2024-12-06 17:51:56.791784] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:34 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.791806] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:39512 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.004 [2024-12-06 17:51:56.791822] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.792520] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:40240 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.004 [2024-12-06 17:51:56.792544] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:43 cdw0:0 sqhd:0071 p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.792571] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:39640 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.004 [2024-12-06 17:51:56.792589] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:8 cdw0:0 sqhd:0072 p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.792617] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:40576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.004 [2024-12-06 17:51:56.792635] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:25 cdw0:0 sqhd:0073 p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.792658] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:40592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.004 [2024-12-06 17:51:56.792686] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:0074 p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.792710] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:40608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.004 [2024-12-06 17:51:56.792727] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:0075 p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.792749] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:40624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.004 [2024-12-06 17:51:56.792766] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:0076 p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.792789] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:40272 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.004 [2024-12-06 17:51:56.792805] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:44 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.792827] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:40304 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.004 [2024-12-06 17:51:56.792843] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:47 cdw0:0 sqhd:0078 p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.792866] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:40336 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.004 [2024-12-06 17:51:56.792882] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:126 cdw0:0 sqhd:0079 p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.792910] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:40368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.004 [2024-12-06 17:51:56.792927] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:1 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.792949] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:39624 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.004 [2024-12-06 17:51:56.792966] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:28 cdw0:0 sqhd:007b p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.792988] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:40296 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.004 [2024-12-06 17:51:56.793005] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:41 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.793027] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:40328 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.004 [2024-12-06 17:51:56.793043] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:103 cdw0:0 sqhd:007d p:0 m:0 dnr:0 00:36:18.004 [2024-12-06 17:51:56.793066] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:108 nsid:1 lba:40360 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.004 [2024-12-06 17:51:56.793082] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:108 cdw0:0 sqhd:007e p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.793105] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:110 nsid:1 lba:39656 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.005 [2024-12-06 17:51:56.793137] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.793159] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:39808 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.005 [2024-12-06 17:51:56.793175] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.793211] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:39936 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.005 [2024-12-06 17:51:56.793227] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.793247] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:40064 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.005 [2024-12-06 17:51:56.793262] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.793283] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:40264 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.005 [2024-12-06 17:51:56.793299] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:123 cdw0:0 sqhd:0003 p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.793319] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:39496 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.005 [2024-12-06 17:51:56.793335] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:0004 p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.793356] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:39816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.005 [2024-12-06 17:51:56.793371] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:0005 p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.793394] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:39944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.005 [2024-12-06 17:51:56.793429] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:0006 p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.793453] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:40056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.005 [2024-12-06 17:51:56.793469] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:0007 p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.793507] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:40184 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.005 [2024-12-06 17:51:56.793524] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:45 cdw0:0 sqhd:0008 p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.793546] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:39416 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.005 [2024-12-06 17:51:56.793562] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:76 cdw0:0 sqhd:0009 p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.793584] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:115 nsid:1 lba:39296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.005 [2024-12-06 17:51:56.793600] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:115 cdw0:0 sqhd:000a p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.793623] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:40256 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.005 [2024-12-06 17:51:56.793639] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:7 cdw0:0 sqhd:000b p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.793662] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:70 nsid:1 lba:39608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.005 [2024-12-06 17:51:56.793689] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:70 cdw0:0 sqhd:000c p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.795430] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:39768 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.005 [2024-12-06 17:51:56.795456] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:000d p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.795483] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:39896 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.005 [2024-12-06 17:51:56.795501] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:34 cdw0:0 sqhd:000e p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.795523] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:112 nsid:1 lba:40040 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.005 [2024-12-06 17:51:56.795540] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:112 cdw0:0 sqhd:000f p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.795562] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:40168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.005 [2024-12-06 17:51:56.795579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:21 cdw0:0 sqhd:0010 p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.795601] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:40416 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.005 [2024-12-06 17:51:56.795617] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:0011 p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.795638] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:40448 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.005 [2024-12-06 17:51:56.795660] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:62 cdw0:0 sqhd:0012 p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.795694] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:40480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.005 [2024-12-06 17:51:56.795711] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:0013 p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.795733] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:40512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.005 [2024-12-06 17:51:56.795749] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:89 cdw0:0 sqhd:0014 p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.795771] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:40544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.005 [2024-12-06 17:51:56.795788] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:91 cdw0:0 sqhd:0015 p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.795810] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:39384 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.005 [2024-12-06 17:51:56.795826] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:111 cdw0:0 sqhd:0016 p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.795848] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:39720 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.005 [2024-12-06 17:51:56.795864] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:32 cdw0:0 sqhd:0017 p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.795886] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:39848 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.005 [2024-12-06 17:51:56.795902] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:37 cdw0:0 sqhd:0018 p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.795924] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:106 nsid:1 lba:39960 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.005 [2024-12-06 17:51:56.795940] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:0019 p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.795962] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:40088 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.005 [2024-12-06 17:51:56.795992] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:001a p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.796014] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:40648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.005 [2024-12-06 17:51:56.796030] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:001b p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.796050] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:39352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.005 [2024-12-06 17:51:56.796065] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:001c p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.796085] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:40208 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.005 [2024-12-06 17:51:56.796100] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:92 cdw0:0 sqhd:001d p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.796121] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:83 nsid:1 lba:39672 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.005 [2024-12-06 17:51:56.796136] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:83 cdw0:0 sqhd:001e p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.796161] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:39640 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.005 [2024-12-06 17:51:56.796176] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:70 cdw0:0 sqhd:001f p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.796197] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:40592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.005 [2024-12-06 17:51:56.796213] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:7 cdw0:0 sqhd:0020 p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.796233] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:40624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.005 [2024-12-06 17:51:56.796248] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:115 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.796268] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:76 nsid:1 lba:40304 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.005 [2024-12-06 17:51:56.796284] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:76 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.796304] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:40368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.005 [2024-12-06 17:51:56.796319] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:45 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.796340] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:40296 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.005 [2024-12-06 17:51:56.796355] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:36:18.005 [2024-12-06 17:51:56.796376] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:40360 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.005 [2024-12-06 17:51:56.796391] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.796412] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:39808 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.006 [2024-12-06 17:51:56.796427] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.796448] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:40064 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.006 [2024-12-06 17:51:56.796463] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.796484] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:123 nsid:1 lba:39496 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.006 [2024-12-06 17:51:56.796499] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:123 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.796520] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:39944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.006 [2024-12-06 17:51:56.796535] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:0029 p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.796556] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:40184 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.006 [2024-12-06 17:51:56.796572] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:002a p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.796596] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:39296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.006 [2024-12-06 17:51:56.796612] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:002b p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.796633] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:39608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.006 [2024-12-06 17:51:56.796662] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.798849] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:40408 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.006 [2024-12-06 17:51:56.798874] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:108 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.798918] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:40664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.006 [2024-12-06 17:51:56.798935] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:103 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.798958] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:40680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.006 [2024-12-06 17:51:56.798974] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:41 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.798996] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:40696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.006 [2024-12-06 17:51:56.799028] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:28 cdw0:0 sqhd:0030 p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.799051] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:40712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.006 [2024-12-06 17:51:56.799066] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:1 cdw0:0 sqhd:0031 p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.799109] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:40728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.006 [2024-12-06 17:51:56.799127] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:126 cdw0:0 sqhd:0032 p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.799150] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:40744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.006 [2024-12-06 17:51:56.799166] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:47 cdw0:0 sqhd:0033 p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.799188] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:40760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.006 [2024-12-06 17:51:56.799205] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:44 cdw0:0 sqhd:0034 p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.799227] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:40776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.006 [2024-12-06 17:51:56.799243] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:0035 p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.799265] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:40792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.006 [2024-12-06 17:51:56.799281] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:0036 p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.799304] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:40808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.006 [2024-12-06 17:51:56.799325] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:0037 p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.799348] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:40824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.006 [2024-12-06 17:51:56.799365] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:25 cdw0:0 sqhd:0038 p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.799386] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:40440 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.006 [2024-12-06 17:51:56.799403] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:8 cdw0:0 sqhd:0039 p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.799441] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:40472 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.006 [2024-12-06 17:51:56.799456] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:43 cdw0:0 sqhd:003a p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.799492] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:40504 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.006 [2024-12-06 17:51:56.799507] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:003b p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.799529] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:74 nsid:1 lba:40536 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.006 [2024-12-06 17:51:56.799544] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:74 cdw0:0 sqhd:003c p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.799564] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:40832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.006 [2024-12-06 17:51:56.799579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:003d p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.799600] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:40848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.006 [2024-12-06 17:51:56.799615] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:003e p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.799636] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:110 nsid:1 lba:39896 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.006 [2024-12-06 17:51:56.799673] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:003f p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.799700] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:40168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.006 [2024-12-06 17:51:56.799716] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:0040 p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.799739] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:40448 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.006 [2024-12-06 17:51:56.799755] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.799777] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:40512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.006 [2024-12-06 17:51:56.799794] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.799816] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:123 nsid:1 lba:39384 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.006 [2024-12-06 17:51:56.799836] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:123 cdw0:0 sqhd:0043 p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.799859] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:39848 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.006 [2024-12-06 17:51:56.799876] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:0044 p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.799897] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:40088 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.006 [2024-12-06 17:51:56.799913] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:0045 p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.799935] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:39352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.006 [2024-12-06 17:51:56.799951] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:0046 p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.799974] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:39672 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.006 [2024-12-06 17:51:56.800005] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:0047 p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.801163] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:40592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.006 [2024-12-06 17:51:56.801187] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:45 cdw0:0 sqhd:0048 p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.801230] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:76 nsid:1 lba:40304 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.006 [2024-12-06 17:51:56.801248] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:76 cdw0:0 sqhd:0049 p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.801270] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:40296 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.006 [2024-12-06 17:51:56.801287] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:115 cdw0:0 sqhd:004a p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.801309] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:39808 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.006 [2024-12-06 17:51:56.801325] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:7 cdw0:0 sqhd:004b p:0 m:0 dnr:0 00:36:18.006 [2024-12-06 17:51:56.801348] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:39496 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.006 [2024-12-06 17:51:56.801364] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:70 cdw0:0 sqhd:004c p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.801386] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:40184 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.007 [2024-12-06 17:51:56.801402] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:83 cdw0:0 sqhd:004d p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.801424] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:39608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.007 [2024-12-06 17:51:56.801441] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:92 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.801463] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:40584 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.007 [2024-12-06 17:51:56.801479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.801509] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:40616 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.007 [2024-12-06 17:51:56.801530] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:0050 p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.801554] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:40280 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.007 [2024-12-06 17:51:56.801571] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:0051 p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.801593] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:106 nsid:1 lba:40344 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.007 [2024-12-06 17:51:56.801610] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:0052 p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.801632] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:39752 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.007 [2024-12-06 17:51:56.801648] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:37 cdw0:0 sqhd:0053 p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.801679] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:39992 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.007 [2024-12-06 17:51:56.801698] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:32 cdw0:0 sqhd:0054 p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.801721] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:40856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.007 [2024-12-06 17:51:56.801737] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:111 cdw0:0 sqhd:0055 p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.801758] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:40872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.007 [2024-12-06 17:51:56.801774] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:91 cdw0:0 sqhd:0056 p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.801796] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:40888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.007 [2024-12-06 17:51:56.801813] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:89 cdw0:0 sqhd:0057 p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.801835] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:40400 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.007 [2024-12-06 17:51:56.801851] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:0058 p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.801873] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:40464 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.007 [2024-12-06 17:51:56.801889] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:62 cdw0:0 sqhd:0059 p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.801911] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:64 nsid:1 lba:40528 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.007 [2024-12-06 17:51:56.801927] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:005a p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.801965] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:40904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.007 [2024-12-06 17:51:56.801981] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:21 cdw0:0 sqhd:005b p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.802022] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:40920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.007 [2024-12-06 17:51:56.802041] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:112 cdw0:0 sqhd:005c p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.802064] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:40936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.007 [2024-12-06 17:51:56.802080] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:34 cdw0:0 sqhd:005d p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.802102] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:40952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.007 [2024-12-06 17:51:56.802118] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.802140] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:40968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.007 [2024-12-06 17:51:56.802156] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:005f p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.802178] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:40984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.007 [2024-12-06 17:51:56.802194] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:0060 p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.802216] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:41000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.007 [2024-12-06 17:51:56.802231] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:65 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.802254] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:40640 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.007 [2024-12-06 17:51:56.802271] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.803087] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:40664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.007 [2024-12-06 17:51:56.803110] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:0063 p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.803137] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:40696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.007 [2024-12-06 17:51:56.803154] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:0064 p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.803176] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:40728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.007 [2024-12-06 17:51:56.803191] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:0065 p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.803212] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:40760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.007 [2024-12-06 17:51:56.803228] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:0066 p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.803249] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:40792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.007 [2024-12-06 17:51:56.803264] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:123 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.803285] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:40824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.007 [2024-12-06 17:51:56.803305] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:0068 p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.803327] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:40472 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.007 [2024-12-06 17:51:56.803342] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:0069 p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.803363] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:40536 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.007 [2024-12-06 17:51:56.803379] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:006a p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.803399] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:40848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.007 [2024-12-06 17:51:56.803414] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:006b p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.803434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:40168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.007 [2024-12-06 17:51:56.803470] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:006c p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.803494] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:40512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.007 [2024-12-06 17:51:56.803511] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.803532] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:74 nsid:1 lba:39848 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.007 [2024-12-06 17:51:56.803549] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:74 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.803572] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:39352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.007 [2024-12-06 17:51:56.803588] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.804022] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:40576 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.007 [2024-12-06 17:51:56.804047] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:43 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.804075] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:40328 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.007 [2024-12-06 17:51:56.804108] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:8 cdw0:0 sqhd:0071 p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.804130] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:39816 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.007 [2024-12-06 17:51:56.804161] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:25 cdw0:0 sqhd:0072 p:0 m:0 dnr:0 00:36:18.007 [2024-12-06 17:51:56.804183] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:39416 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.007 [2024-12-06 17:51:56.804199] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:0073 p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.804220] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:41016 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.008 [2024-12-06 17:51:56.804240] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:0074 p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.804262] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:41032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.008 [2024-12-06 17:51:56.804278] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:0075 p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.804299] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:41048 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.008 [2024-12-06 17:51:56.804314] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:44 cdw0:0 sqhd:0076 p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.804352] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:41064 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.008 [2024-12-06 17:51:56.804367] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:47 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.804404] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:41080 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.008 [2024-12-06 17:51:56.804421] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:126 cdw0:0 sqhd:0078 p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.804449] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:41096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.008 [2024-12-06 17:51:56.804466] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:1 cdw0:0 sqhd:0079 p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.804488] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:40304 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.008 [2024-12-06 17:51:56.804504] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.804526] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:39808 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.008 [2024-12-06 17:51:56.804543] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:65 cdw0:0 sqhd:007b p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.804565] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:40184 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.008 [2024-12-06 17:51:56.804581] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.804602] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:79 nsid:1 lba:40584 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.008 [2024-12-06 17:51:56.804619] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:007d p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.804641] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:40280 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.008 [2024-12-06 17:51:56.804657] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:007e p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.804693] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:39752 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.008 [2024-12-06 17:51:56.804710] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:34 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.804732] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:40856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.008 [2024-12-06 17:51:56.804749] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:112 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.804776] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:40888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.008 [2024-12-06 17:51:56.804793] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:21 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.804815] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:64 nsid:1 lba:40464 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.008 [2024-12-06 17:51:56.804831] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.804853] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:40904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.008 [2024-12-06 17:51:56.804870] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:62 cdw0:0 sqhd:0003 p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.804892] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:40936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.008 [2024-12-06 17:51:56.804908] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:0004 p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.804930] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:40968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.008 [2024-12-06 17:51:56.804946] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:89 cdw0:0 sqhd:0005 p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.804969] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:41000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.008 [2024-12-06 17:51:56.804986] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:91 cdw0:0 sqhd:0006 p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.806333] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:40656 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.008 [2024-12-06 17:51:56.806359] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:111 cdw0:0 sqhd:0007 p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.806386] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:40688 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.008 [2024-12-06 17:51:56.806404] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:32 cdw0:0 sqhd:0008 p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.806427] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:40720 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.008 [2024-12-06 17:51:56.806444] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:37 cdw0:0 sqhd:0009 p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.806466] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:106 nsid:1 lba:40752 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.008 [2024-12-06 17:51:56.806483] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:000a p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.806505] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:40784 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.008 [2024-12-06 17:51:56.806522] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:000b p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.806544] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:40816 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.008 [2024-12-06 17:51:56.806560] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:000c p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.806588] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:40416 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.008 [2024-12-06 17:51:56.806605] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:000d p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.806627] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:40544 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.008 [2024-12-06 17:51:56.806644] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:92 cdw0:0 sqhd:000e p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.806673] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:40696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.008 [2024-12-06 17:51:56.806691] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:000f p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.806759] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:74 nsid:1 lba:40760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.008 [2024-12-06 17:51:56.806777] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:74 cdw0:0 sqhd:0010 p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.806799] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:40824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.008 [2024-12-06 17:51:56.806815] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:0011 p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.806836] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:40536 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.008 [2024-12-06 17:51:56.806851] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:0012 p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.806873] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:110 nsid:1 lba:40168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.008 [2024-12-06 17:51:56.806888] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:0013 p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.806910] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:39848 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.008 [2024-12-06 17:51:56.806926] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:0014 p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.806947] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:41112 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.008 [2024-12-06 17:51:56.806963] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:0015 p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.807001] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:40624 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.008 [2024-12-06 17:51:56.807016] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:0016 p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.807037] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:123 nsid:1 lba:39944 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.008 [2024-12-06 17:51:56.807052] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:123 cdw0:0 sqhd:0017 p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.807072] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:40328 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.008 [2024-12-06 17:51:56.807088] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:91 cdw0:0 sqhd:0018 p:0 m:0 dnr:0 00:36:18.008 [2024-12-06 17:51:56.807113] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:89 nsid:1 lba:39416 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.008 [2024-12-06 17:51:56.807129] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:89 cdw0:0 sqhd:0019 p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.807149] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:41032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.009 [2024-12-06 17:51:56.807164] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:001a p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.807185] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:41064 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.009 [2024-12-06 17:51:56.807200] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:62 cdw0:0 sqhd:001b p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.807221] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:41096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.009 [2024-12-06 17:51:56.807236] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:001c p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.807256] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:39808 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.009 [2024-12-06 17:51:56.807271] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:21 cdw0:0 sqhd:001d p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.807292] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:112 nsid:1 lba:40584 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.009 [2024-12-06 17:51:56.807307] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:112 cdw0:0 sqhd:001e p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.807327] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:39752 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.009 [2024-12-06 17:51:56.807343] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:34 cdw0:0 sqhd:001f p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.807363] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:40888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.009 [2024-12-06 17:51:56.807379] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:0020 p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.807400] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:40904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.009 [2024-12-06 17:51:56.807415] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.807436] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:40968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.009 [2024-12-06 17:51:56.807451] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.809072] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:40864 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.009 [2024-12-06 17:51:56.809123] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:65 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.809156] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:40896 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.009 [2024-12-06 17:51:56.809194] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.809218] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:41128 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.009 [2024-12-06 17:51:56.809240] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:1 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.809263] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:41144 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.009 [2024-12-06 17:51:56.809279] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:126 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.809301] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:41160 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.009 [2024-12-06 17:51:56.809317] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:47 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.809338] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:41176 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.009 [2024-12-06 17:51:56.809354] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:44 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.809375] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:41192 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.009 [2024-12-06 17:51:56.809390] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:0029 p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.809412] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:41208 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.009 [2024-12-06 17:51:56.809428] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:002a p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.809449] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:41224 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.009 [2024-12-06 17:51:56.809465] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:002b p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.809987] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:41240 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.009 [2024-12-06 17:51:56.810012] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:25 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.810048] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:41256 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.009 [2024-12-06 17:51:56.810082] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:8 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.810104] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:41272 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.009 [2024-12-06 17:51:56.810120] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:43 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.810157] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:40688 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.009 [2024-12-06 17:51:56.810173] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.810194] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:79 nsid:1 lba:40752 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.009 [2024-12-06 17:51:56.810209] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:0030 p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.810250] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:40816 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.009 [2024-12-06 17:51:56.810272] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:0031 p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.810295] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:40544 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.009 [2024-12-06 17:51:56.810311] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:34 cdw0:0 sqhd:0032 p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.810333] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:40760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.009 [2024-12-06 17:51:56.810349] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:112 cdw0:0 sqhd:0033 p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.810370] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:40536 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.009 [2024-12-06 17:51:56.810385] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:21 cdw0:0 sqhd:0034 p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.810407] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:64 nsid:1 lba:39848 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.009 [2024-12-06 17:51:56.810423] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:0035 p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.810444] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:40624 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.009 [2024-12-06 17:51:56.810459] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:62 cdw0:0 sqhd:0036 p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.810481] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:40328 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.009 [2024-12-06 17:51:56.810497] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:0037 p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.810518] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:41032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.009 [2024-12-06 17:51:56.810547] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:89 cdw0:0 sqhd:0038 p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.810569] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:41096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.009 [2024-12-06 17:51:56.810584] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:91 cdw0:0 sqhd:0039 p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.810605] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:123 nsid:1 lba:40584 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.009 [2024-12-06 17:51:56.810620] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:123 cdw0:0 sqhd:003a p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.810641] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:40888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.009 [2024-12-06 17:51:56.810656] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:003b p:0 m:0 dnr:0 00:36:18.009 [2024-12-06 17:51:56.810703] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:40968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.009 [2024-12-06 17:51:56.810721] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:003c p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.810742] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:40928 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.010 [2024-12-06 17:51:56.810758] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:003d p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.810784] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:110 nsid:1 lba:40960 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.010 [2024-12-06 17:51:56.810800] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:003e p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.810821] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:40992 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.010 [2024-12-06 17:51:56.810837] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:003f p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.810858] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:40680 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.010 [2024-12-06 17:51:56.810873] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:0040 p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.810894] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:74 nsid:1 lba:40744 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.010 [2024-12-06 17:51:56.810910] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:74 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.810932] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:40808 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.010 [2024-12-06 17:51:56.810947] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.810968] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:40448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.010 [2024-12-06 17:51:56.810998] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:92 cdw0:0 sqhd:0043 p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.811020] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:41296 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.010 [2024-12-06 17:51:56.811036] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:0044 p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.811075] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:41312 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.010 [2024-12-06 17:51:56.811091] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:0045 p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.811113] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:41328 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.010 [2024-12-06 17:51:56.811129] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:0046 p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.811151] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:41344 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.010 [2024-12-06 17:51:56.811168] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:0047 p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.811191] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:41360 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.010 [2024-12-06 17:51:56.811207] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:37 cdw0:0 sqhd:0048 p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.812490] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:41040 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.010 [2024-12-06 17:51:56.812516] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:32 cdw0:0 sqhd:0049 p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.812550] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:41072 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.010 [2024-12-06 17:51:56.812569] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:111 cdw0:0 sqhd:004a p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.812592] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:41104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.010 [2024-12-06 17:51:56.812608] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:004b p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.812630] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:40296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.010 [2024-12-06 17:51:56.812647] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:004c p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.812687] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:40872 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.010 [2024-12-06 17:51:56.812706] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:004d p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.812728] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:40896 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.010 [2024-12-06 17:51:56.812744] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.812766] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:41144 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.010 [2024-12-06 17:51:56.812783] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.812805] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:41176 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.010 [2024-12-06 17:51:56.812822] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:0050 p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.812845] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:41208 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.010 [2024-12-06 17:51:56.812861] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:44 cdw0:0 sqhd:0051 p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.812884] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:40920 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.010 [2024-12-06 17:51:56.812901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:47 cdw0:0 sqhd:0052 p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.812923] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:40984 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.010 [2024-12-06 17:51:56.812940] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:126 cdw0:0 sqhd:0053 p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.812962] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:41376 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.010 [2024-12-06 17:51:56.812978] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:1 cdw0:0 sqhd:0054 p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.813000] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:41392 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.010 [2024-12-06 17:51:56.813016] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:0055 p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.813038] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:41256 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.010 [2024-12-06 17:51:56.813059] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:37 cdw0:0 sqhd:0056 p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.813082] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:106 nsid:1 lba:40688 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.010 [2024-12-06 17:51:56.813099] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:0057 p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.813121] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:40816 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.010 [2024-12-06 17:51:56.813137] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:0058 p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.813159] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:40760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.010 [2024-12-06 17:51:56.813175] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:0059 p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.813197] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:39848 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.010 [2024-12-06 17:51:56.813213] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:005a p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.813236] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:40328 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.010 [2024-12-06 17:51:56.813253] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:92 cdw0:0 sqhd:005b p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.813275] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:41096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.010 [2024-12-06 17:51:56.813291] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:005c p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.813315] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:74 nsid:1 lba:40888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.010 [2024-12-06 17:51:56.813332] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:74 cdw0:0 sqhd:005d p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.813356] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:40928 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.010 [2024-12-06 17:51:56.813372] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.813395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:40992 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.010 [2024-12-06 17:51:56.813412] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:005f p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.813434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:110 nsid:1 lba:40744 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.010 [2024-12-06 17:51:56.813450] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:0060 p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.813490] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:40448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.010 [2024-12-06 17:51:56.813507] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.813529] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:41312 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.010 [2024-12-06 17:51:56.813549] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:36:18.010 [2024-12-06 17:51:56.813574] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:41344 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.010 [2024-12-06 17:51:56.813590] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:0063 p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.815441] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:123 nsid:1 lba:40664 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.011 [2024-12-06 17:51:56.815464] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:123 cdw0:0 sqhd:0064 p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.815506] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:40792 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.011 [2024-12-06 17:51:56.815523] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:91 cdw0:0 sqhd:0065 p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.815544] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:89 nsid:1 lba:40512 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.011 [2024-12-06 17:51:56.815560] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:89 cdw0:0 sqhd:0066 p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.815581] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:41408 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.011 [2024-12-06 17:51:56.815596] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.815616] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:41424 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.011 [2024-12-06 17:51:56.815631] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:62 cdw0:0 sqhd:0068 p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.815682] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:41440 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.011 [2024-12-06 17:51:56.815701] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:0069 p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.815747] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:41456 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.011 [2024-12-06 17:51:56.815763] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:21 cdw0:0 sqhd:006a p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.815785] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:41472 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.011 [2024-12-06 17:51:56.815802] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:112 cdw0:0 sqhd:006b p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.815824] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:41488 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.011 [2024-12-06 17:51:56.815841] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:34 cdw0:0 sqhd:006c p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.815863] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:41504 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.011 [2024-12-06 17:51:56.815878] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.815900] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:41520 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.011 [2024-12-06 17:51:56.815917] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.815945] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:41048 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.011 [2024-12-06 17:51:56.815962] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.815984] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:41072 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.011 [2024-12-06 17:51:56.816001] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.816023] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:40296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.011 [2024-12-06 17:51:56.816040] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:0071 p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.816062] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:40896 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.011 [2024-12-06 17:51:56.816078] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:0072 p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.816099] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:41176 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.011 [2024-12-06 17:51:56.816116] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:0073 p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.816138] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:40920 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.011 [2024-12-06 17:51:56.816154] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:0074 p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.816176] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:41376 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.011 [2024-12-06 17:51:56.816193] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:0075 p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.816215] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:74 nsid:1 lba:41256 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.011 [2024-12-06 17:51:56.816245] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:74 cdw0:0 sqhd:0076 p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.816267] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:40816 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.011 [2024-12-06 17:51:56.816282] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.816303] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:39848 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.011 [2024-12-06 17:51:56.816318] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:92 cdw0:0 sqhd:0078 p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.816346] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:41096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.011 [2024-12-06 17:51:56.816361] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:0079 p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.816381] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:40928 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.011 [2024-12-06 17:51:56.816396] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.816421] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:40744 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.011 [2024-12-06 17:51:56.816437] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:007b p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.816458] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:41312 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.011 [2024-12-06 17:51:56.816474] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.816999] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:40184 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.011 [2024-12-06 17:51:56.817022] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:37 cdw0:0 sqhd:007d p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.817050] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:40936 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.011 [2024-12-06 17:51:56.817068] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:007e p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.817090] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:41528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.011 [2024-12-06 17:51:56.817107] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.817144] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:41544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.011 [2024-12-06 17:51:56.817160] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:126 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.817182] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:41560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.011 [2024-12-06 17:51:56.817197] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:47 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.817218] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:41576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.011 [2024-12-06 17:51:56.817233] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:44 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.817254] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:41136 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.011 [2024-12-06 17:51:56.817269] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:0003 p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.817290] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:41168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.011 [2024-12-06 17:51:56.817305] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:0004 p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.817326] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:41200 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.011 [2024-12-06 17:51:56.817341] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:0005 p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.817361] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:41232 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.011 [2024-12-06 17:51:56.817397] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:0006 p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.817421] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:41264 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.011 [2024-12-06 17:51:56.817444] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:0007 p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.817467] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:40696 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.011 [2024-12-06 17:51:56.817500] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:0008 p:0 m:0 dnr:0 00:36:18.011 [2024-12-06 17:51:56.818605] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:41112 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.011 [2024-12-06 17:51:56.818630] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:111 cdw0:0 sqhd:0009 p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.818689] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:41584 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.012 [2024-12-06 17:51:56.818708] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:32 cdw0:0 sqhd:000a p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.818732] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:41600 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.012 [2024-12-06 17:51:56.818749] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:43 cdw0:0 sqhd:000b p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.818771] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:41616 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.012 [2024-12-06 17:51:56.818787] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:8 cdw0:0 sqhd:000c p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.818809] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:41632 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.012 [2024-12-06 17:51:56.818826] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:25 cdw0:0 sqhd:000d p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.818847] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:41648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.012 [2024-12-06 17:51:56.818864] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:65 cdw0:0 sqhd:000e p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.818885] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:40904 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.012 [2024-12-06 17:51:56.818902] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:000f p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.818924] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:106 nsid:1 lba:40792 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.012 [2024-12-06 17:51:56.818940] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:0010 p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.818962] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:41408 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.012 [2024-12-06 17:51:56.818978] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:0011 p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.819000] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:41440 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.012 [2024-12-06 17:51:56.819016] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:0012 p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.819038] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:41472 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.012 [2024-12-06 17:51:56.819074] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:0013 p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.819096] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:41504 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.012 [2024-12-06 17:51:56.819112] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:92 cdw0:0 sqhd:0014 p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.819132] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:41048 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.012 [2024-12-06 17:51:56.819148] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:0015 p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.819168] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:74 nsid:1 lba:40296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.012 [2024-12-06 17:51:56.819184] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:74 cdw0:0 sqhd:0016 p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.819204] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:41176 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.012 [2024-12-06 17:51:56.819219] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:0017 p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.819240] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:41376 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.012 [2024-12-06 17:51:56.819256] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:0018 p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.819276] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:110 nsid:1 lba:40816 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.012 [2024-12-06 17:51:56.819291] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:0019 p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.819312] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:41096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.012 [2024-12-06 17:51:56.819327] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:001a p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.819347] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:40744 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.012 [2024-12-06 17:51:56.819363] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:001b p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.819383] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:41288 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.012 [2024-12-06 17:51:56.819398] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:001c p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.819418] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:41320 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.012 [2024-12-06 17:51:56.819434] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:001d p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.819454] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:79 nsid:1 lba:41352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.012 [2024-12-06 17:51:56.819469] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:001e p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.819489] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:40936 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.012 [2024-12-06 17:51:56.819504] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:001f p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.819529] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:41544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.012 [2024-12-06 17:51:56.819545] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:0020 p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.819565] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:41576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.012 [2024-12-06 17:51:56.819580] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.819601] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:41168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.012 [2024-12-06 17:51:56.819616] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.819636] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:41232 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.012 [2024-12-06 17:51:56.819679] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.819704] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:40696 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.012 [2024-12-06 17:51:56.819737] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.821487] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:41160 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.012 [2024-12-06 17:51:56.821529] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:44 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.821556] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:41224 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.012 [2024-12-06 17:51:56.821573] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:47 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.821594] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:41672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.012 [2024-12-06 17:51:56.821609] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:126 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.821631] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:41688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.012 [2024-12-06 17:51:56.821662] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:1 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.821703] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:41704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.012 [2024-12-06 17:51:56.821720] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:0029 p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.821741] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:41720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.012 [2024-12-06 17:51:56.821758] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:37 cdw0:0 sqhd:002a p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.821785] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:41736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.012 [2024-12-06 17:51:56.821802] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:002b p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.821830] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:41752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.012 [2024-12-06 17:51:56.821847] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:34 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.821869] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:112 nsid:1 lba:41368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.012 [2024-12-06 17:51:56.821885] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:112 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.821907] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:41240 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.012 [2024-12-06 17:51:56.821923] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:21 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:36:18.012 [2024-12-06 17:51:56.821944] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:64 nsid:1 lba:41032 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.012 [2024-12-06 17:51:56.821960] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.821982] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:41584 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.013 [2024-12-06 17:51:56.821998] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:0030 p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.822020] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:41616 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.013 [2024-12-06 17:51:56.822036] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:0031 p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.822058] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:41648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.013 [2024-12-06 17:51:56.822075] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:0032 p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.822096] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:40792 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.013 [2024-12-06 17:51:56.822112] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:0033 p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.822134] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:41440 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.013 [2024-12-06 17:51:56.822150] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:0034 p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.822172] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:41504 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.013 [2024-12-06 17:51:56.822189] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:0035 p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.822211] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:79 nsid:1 lba:40296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.013 [2024-12-06 17:51:56.822227] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:0036 p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.822249] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:41376 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.013 [2024-12-06 17:51:56.822265] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:0037 p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.822287] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:41096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.013 [2024-12-06 17:51:56.822307] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:0038 p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.822346] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:41288 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.013 [2024-12-06 17:51:56.822361] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:0039 p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.822382] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:41352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.013 [2024-12-06 17:51:56.822397] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:003a p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.822418] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:41544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.013 [2024-12-06 17:51:56.822433] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:003b p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.822454] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:41168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.013 [2024-12-06 17:51:56.822469] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:003c p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.822490] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:40696 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.013 [2024-12-06 17:51:56.822506] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:003d p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.823792] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:74 nsid:1 lba:41328 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.013 [2024-12-06 17:51:56.823817] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:74 cdw0:0 sqhd:003e p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.823845] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:41768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.013 [2024-12-06 17:51:56.823863] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:003f p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.823886] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:41784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.013 [2024-12-06 17:51:56.823903] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:92 cdw0:0 sqhd:0040 p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.823925] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:41800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.013 [2024-12-06 17:51:56.823941] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.823964] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:41816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.013 [2024-12-06 17:51:56.823981] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.824002] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:41832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.013 [2024-12-06 17:51:56.824019] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:0043 p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.824041] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:106 nsid:1 lba:41416 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.013 [2024-12-06 17:51:56.824062] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:0044 p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.824085] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:41448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.013 [2024-12-06 17:51:56.824102] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:0045 p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.824124] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:41480 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.013 [2024-12-06 17:51:56.824140] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:65 cdw0:0 sqhd:0046 p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.824162] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:41512 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.013 [2024-12-06 17:51:56.824178] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:25 cdw0:0 sqhd:0047 p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.824200] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:41208 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.013 [2024-12-06 17:51:56.824234] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:8 cdw0:0 sqhd:0048 p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.824257] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:40760 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.013 [2024-12-06 17:51:56.824272] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:43 cdw0:0 sqhd:0049 p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.824309] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:41344 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.013 [2024-12-06 17:51:56.824324] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:32 cdw0:0 sqhd:004a p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.824345] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:41848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.013 [2024-12-06 17:51:56.824359] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:111 cdw0:0 sqhd:004b p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.824380] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:41224 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.013 [2024-12-06 17:51:56.824395] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:004c p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.824416] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:41688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.013 [2024-12-06 17:51:56.824431] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:004d p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.824451] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:41720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.013 [2024-12-06 17:51:56.824467] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.824488] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:41752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.013 [2024-12-06 17:51:56.824503] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.824523] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:41240 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.013 [2024-12-06 17:51:56.824538] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:0050 p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.824563] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:41584 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.013 [2024-12-06 17:51:56.824579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:0051 p:0 m:0 dnr:0 00:36:18.013 [2024-12-06 17:51:56.824599] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:41648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.014 [2024-12-06 17:51:56.824614] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:0052 p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.824634] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:41440 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.014 [2024-12-06 17:51:56.824674] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:0053 p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.824700] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:40296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.014 [2024-12-06 17:51:56.824732] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:0054 p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.824755] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:41096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.014 [2024-12-06 17:51:56.824771] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:0055 p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.824793] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:41352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.014 [2024-12-06 17:51:56.824809] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:0056 p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.824832] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:41168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.014 [2024-12-06 17:51:56.824848] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:0057 p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.826472] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:41536 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.014 [2024-12-06 17:51:56.826510] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:0058 p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.826537] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:41568 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.014 [2024-12-06 17:51:56.826568] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:0059 p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.826590] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:41864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.014 [2024-12-06 17:51:56.826605] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:005a p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.826626] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:41880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.014 [2024-12-06 17:51:56.826662] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:21 cdw0:0 sqhd:005b p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.826704] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:41896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.014 [2024-12-06 17:51:56.826721] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:112 cdw0:0 sqhd:005c p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.826748] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:41912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.014 [2024-12-06 17:51:56.826770] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:34 cdw0:0 sqhd:005d p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.826801] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:41928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.014 [2024-12-06 17:51:56.826819] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.826841] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:41944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.014 [2024-12-06 17:51:56.826857] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:37 cdw0:0 sqhd:005f p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.826879] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:41960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.014 [2024-12-06 17:51:56.826895] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:0060 p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.826917] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:41976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.014 [2024-12-06 17:51:56.826934] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:1 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.826956] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:41992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.014 [2024-12-06 17:51:56.826972] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:126 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.827008] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:41768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.014 [2024-12-06 17:51:56.827025] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:0063 p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.827047] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:41800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.014 [2024-12-06 17:51:56.827078] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:0064 p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.827099] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:41832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.014 [2024-12-06 17:51:56.827115] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:0065 p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.827135] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:41448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.014 [2024-12-06 17:51:56.827150] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:0066 p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.827171] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:79 nsid:1 lba:41512 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.014 [2024-12-06 17:51:56.827186] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.827223] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:40760 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.014 [2024-12-06 17:51:56.827239] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:0068 p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.827266] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:41848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.014 [2024-12-06 17:51:56.827282] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:0069 p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.827305] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:41688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.014 [2024-12-06 17:51:56.827321] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:006a p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.827344] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:41752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.014 [2024-12-06 17:51:56.827361] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:006b p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.827846] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:41584 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.014 [2024-12-06 17:51:56.827869] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:006c p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.827896] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:41440 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.014 [2024-12-06 17:51:56.827914] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.827937] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:41096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.014 [2024-12-06 17:51:56.827954] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.827976] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:41168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.014 [2024-12-06 17:51:56.827992] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:111 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.828015] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:41608 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.014 [2024-12-06 17:51:56.828033] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:32 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.828072] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:41640 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.014 [2024-12-06 17:51:56.828088] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:43 cdw0:0 sqhd:0071 p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.828110] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:41424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.014 [2024-12-06 17:51:56.828140] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:8 cdw0:0 sqhd:0072 p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.828162] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:41488 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.014 [2024-12-06 17:51:56.828177] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:25 cdw0:0 sqhd:0073 p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.828197] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:41256 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.014 [2024-12-06 17:51:56.828212] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:65 cdw0:0 sqhd:0074 p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.828233] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:42008 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.014 [2024-12-06 17:51:56.828253] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:0075 p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.828275] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:42024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.014 [2024-12-06 17:51:56.828291] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:0076 p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.829012] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:41560 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.014 [2024-12-06 17:51:56.829050] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.829093] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:42040 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.014 [2024-12-06 17:51:56.829131] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:0078 p:0 m:0 dnr:0 00:36:18.014 [2024-12-06 17:51:56.829155] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:42056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.015 [2024-12-06 17:51:56.829187] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:0079 p:0 m:0 dnr:0 00:36:18.015 [2024-12-06 17:51:56.829209] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:42072 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.015 [2024-12-06 17:51:56.829225] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:92 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:36:18.015 [2024-12-06 17:51:56.829246] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:42088 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.015 [2024-12-06 17:51:56.829261] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:007b p:0 m:0 dnr:0 00:36:18.015 [2024-12-06 17:51:56.829282] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:74 nsid:1 lba:42104 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.015 [2024-12-06 17:51:56.829298] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:74 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:36:18.015 [2024-12-06 17:51:56.829319] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:41664 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.015 [2024-12-06 17:51:56.829334] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:47 cdw0:0 sqhd:007d p:0 m:0 dnr:0 00:36:18.015 [2024-12-06 17:51:56.829371] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:41696 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.015 [2024-12-06 17:51:56.829387] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:44 cdw0:0 sqhd:007e p:0 m:0 dnr:0 00:36:18.015 [2024-12-06 17:51:56.829407] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:41728 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.015 [2024-12-06 17:51:56.829422] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:62 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:36:18.015 [2024-12-06 17:51:56.829443] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:41760 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.015 [2024-12-06 17:51:56.829457] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:36:18.015 [2024-12-06 17:51:56.829478] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:41568 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.015 [2024-12-06 17:51:56.829498] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:36:18.015 [2024-12-06 17:51:56.829520] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:41880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.015 [2024-12-06 17:51:56.829551] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:36:18.015 [2024-12-06 17:51:56.829575] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:41912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.015 [2024-12-06 17:51:56.829591] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:0003 p:0 m:0 dnr:0 00:36:18.015 [2024-12-06 17:51:56.829613] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:41944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.015 [2024-12-06 17:51:56.829629] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:0004 p:0 m:0 dnr:0 00:36:18.015 [2024-12-06 17:51:56.829651] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:41976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.015 [2024-12-06 17:51:56.829676] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:0005 p:0 m:0 dnr:0 00:36:18.015 [2024-12-06 17:51:56.829702] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:41768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.015 [2024-12-06 17:51:56.829719] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:0006 p:0 m:0 dnr:0 00:36:18.015 [2024-12-06 17:51:56.829741] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:41832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.015 [2024-12-06 17:51:56.829758] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:0007 p:0 m:0 dnr:0 00:36:18.015 [2024-12-06 17:51:56.829780] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:41512 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.015 [2024-12-06 17:51:56.829796] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:0008 p:0 m:0 dnr:0 00:36:18.015 [2024-12-06 17:51:56.829818] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:41848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.015 [2024-12-06 17:51:56.829834] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:0009 p:0 m:0 dnr:0 00:36:18.015 [2024-12-06 17:51:56.829856] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:41752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:36:18.015 [2024-12-06 17:51:56.829872] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:126 cdw0:0 sqhd:000a p:0 m:0 dnr:0 00:36:18.015 [2024-12-06 17:51:56.829894] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:41632 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.015 [2024-12-06 17:51:56.829910] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:1 cdw0:0 sqhd:000b p:0 m:0 dnr:0 00:36:18.015 [2024-12-06 17:51:56.829933] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:41472 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:36:18.015 [2024-12-06 17:51:56.829964] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:000c p:0 m:0 dnr:0 00:36:18.015 7959.75 IOPS, 31.09 MiB/s [2024-12-06T16:51:59.854Z] 7973.67 IOPS, 31.15 MiB/s [2024-12-06T16:51:59.854Z] 7988.24 IOPS, 31.20 MiB/s [2024-12-06T16:51:59.854Z] Received shutdown signal, test time was about 34.252490 seconds 00:36:18.015 00:36:18.015 Latency(us) 00:36:18.015 [2024-12-06T16:51:59.854Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:36:18.015 Job: Nvme0n1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:36:18.015 Verification LBA range: start 0x0 length 0x4000 00:36:18.015 Nvme0n1 : 34.25 7987.37 31.20 0.00 0.00 15997.75 458.15 4026531.84 00:36:18.015 [2024-12-06T16:51:59.854Z] =================================================================================================================== 00:36:18.015 [2024-12-06T16:51:59.854Z] Total : 7987.37 31.20 0.00 0.00 15997.75 458.15 4026531.84 00:36:18.015 17:51:59 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@143 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:36:18.273 17:51:59 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@145 -- # trap - SIGINT SIGTERM EXIT 00:36:18.273 17:51:59 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@147 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:36:18.273 17:51:59 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@148 -- # nvmftestfini 00:36:18.273 17:51:59 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@516 -- # nvmfcleanup 00:36:18.273 17:51:59 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@121 -- # sync 00:36:18.273 17:51:59 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:36:18.273 17:51:59 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@124 -- # set +e 00:36:18.273 17:51:59 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@125 -- # for i in {1..20} 00:36:18.273 17:51:59 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:36:18.273 rmmod nvme_tcp 00:36:18.273 rmmod nvme_fabrics 00:36:18.273 rmmod nvme_keyring 00:36:18.273 17:52:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:36:18.273 17:52:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@128 -- # set -e 00:36:18.273 17:52:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@129 -- # return 0 00:36:18.273 17:52:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@517 -- # '[' -n 380474 ']' 00:36:18.273 17:52:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@518 -- # killprocess 380474 00:36:18.273 17:52:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@954 -- # '[' -z 380474 ']' 00:36:18.273 17:52:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@958 -- # kill -0 380474 00:36:18.273 17:52:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@959 -- # uname 00:36:18.273 17:52:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:36:18.273 17:52:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 380474 00:36:18.273 17:52:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:36:18.273 17:52:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:36:18.273 17:52:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@972 -- # echo 'killing process with pid 380474' 00:36:18.273 killing process with pid 380474 00:36:18.273 17:52:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@973 -- # kill 380474 00:36:18.274 17:52:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@978 -- # wait 380474 00:36:18.533 17:52:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:36:18.534 17:52:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:36:18.534 17:52:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:36:18.534 17:52:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@297 -- # iptr 00:36:18.534 17:52:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@791 -- # iptables-save 00:36:18.534 17:52:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:36:18.534 17:52:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@791 -- # iptables-restore 00:36:18.534 17:52:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:36:18.534 17:52:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@302 -- # remove_spdk_ns 00:36:18.534 17:52:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:36:18.534 17:52:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:36:18.534 17:52:00 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:36:21.073 17:52:02 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:36:21.073 00:36:21.073 real 0m43.355s 00:36:21.073 user 2m11.893s 00:36:21.073 sys 0m10.727s 00:36:21.073 17:52:02 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1130 -- # xtrace_disable 00:36:21.073 17:52:02 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:36:21.074 ************************************ 00:36:21.074 END TEST nvmf_host_multipath_status 00:36:21.074 ************************************ 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@28 -- # run_test nvmf_discovery_remove_ifc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/discovery_remove_ifc.sh --transport=tcp 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:36:21.074 ************************************ 00:36:21.074 START TEST nvmf_discovery_remove_ifc 00:36:21.074 ************************************ 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/discovery_remove_ifc.sh --transport=tcp 00:36:21.074 * Looking for test storage... 00:36:21.074 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1711 -- # lcov --version 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@336 -- # IFS=.-: 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@336 -- # read -ra ver1 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@337 -- # IFS=.-: 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@337 -- # read -ra ver2 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@338 -- # local 'op=<' 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@340 -- # ver1_l=2 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@341 -- # ver2_l=1 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@344 -- # case "$op" in 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@345 -- # : 1 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@364 -- # (( v = 0 )) 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@365 -- # decimal 1 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@353 -- # local d=1 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@355 -- # echo 1 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@365 -- # ver1[v]=1 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@366 -- # decimal 2 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@353 -- # local d=2 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@355 -- # echo 2 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@366 -- # ver2[v]=2 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@368 -- # return 0 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:36:21.074 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:21.074 --rc genhtml_branch_coverage=1 00:36:21.074 --rc genhtml_function_coverage=1 00:36:21.074 --rc genhtml_legend=1 00:36:21.074 --rc geninfo_all_blocks=1 00:36:21.074 --rc geninfo_unexecuted_blocks=1 00:36:21.074 00:36:21.074 ' 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:36:21.074 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:21.074 --rc genhtml_branch_coverage=1 00:36:21.074 --rc genhtml_function_coverage=1 00:36:21.074 --rc genhtml_legend=1 00:36:21.074 --rc geninfo_all_blocks=1 00:36:21.074 --rc geninfo_unexecuted_blocks=1 00:36:21.074 00:36:21.074 ' 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:36:21.074 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:21.074 --rc genhtml_branch_coverage=1 00:36:21.074 --rc genhtml_function_coverage=1 00:36:21.074 --rc genhtml_legend=1 00:36:21.074 --rc geninfo_all_blocks=1 00:36:21.074 --rc geninfo_unexecuted_blocks=1 00:36:21.074 00:36:21.074 ' 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:36:21.074 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:21.074 --rc genhtml_branch_coverage=1 00:36:21.074 --rc genhtml_function_coverage=1 00:36:21.074 --rc genhtml_legend=1 00:36:21.074 --rc geninfo_all_blocks=1 00:36:21.074 --rc geninfo_unexecuted_blocks=1 00:36:21.074 00:36:21.074 ' 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@12 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@7 -- # uname -s 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@15 -- # shopt -s extglob 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- paths/export.sh@5 -- # export PATH 00:36:21.074 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@51 -- # : 0 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:36:21.075 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@55 -- # have_pci_nics=0 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@14 -- # '[' tcp == rdma ']' 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@19 -- # discovery_port=8009 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@20 -- # discovery_nqn=nqn.2014-08.org.nvmexpress.discovery 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@23 -- # nqn=nqn.2016-06.io.spdk:cnode 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@25 -- # host_nqn=nqn.2021-12.io.spdk:test 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@26 -- # host_sock=/tmp/host.sock 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@39 -- # nvmftestinit 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@476 -- # prepare_net_devs 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@438 -- # local -g is_hw=no 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@440 -- # remove_spdk_ns 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@309 -- # xtrace_disable 00:36:21.075 17:52:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@315 -- # pci_devs=() 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@315 -- # local -a pci_devs 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@316 -- # pci_net_devs=() 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@317 -- # pci_drivers=() 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@317 -- # local -A pci_drivers 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@319 -- # net_devs=() 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@319 -- # local -ga net_devs 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@320 -- # e810=() 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@320 -- # local -ga e810 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@321 -- # x722=() 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@321 -- # local -ga x722 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@322 -- # mlx=() 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@322 -- # local -ga mlx 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:36:22.978 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:36:22.978 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@418 -- # [[ up == up ]] 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:36:22.978 Found net devices under 0000:0a:00.0: cvl_0_0 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@418 -- # [[ up == up ]] 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:36:22.978 Found net devices under 0000:0a:00.1: cvl_0_1 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@442 -- # is_hw=yes 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:36:22.978 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:36:22.979 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:36:22.979 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:36:22.979 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:36:22.979 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:36:22.979 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:36:22.979 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:36:22.979 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:36:22.979 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:36:22.979 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:36:22.979 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:36:22.979 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:36:22.979 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:36:22.979 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:36:22.979 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:36:23.238 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:36:23.238 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:36:23.238 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:36:23.238 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:36:23.238 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.180 ms 00:36:23.238 00:36:23.238 --- 10.0.0.2 ping statistics --- 00:36:23.238 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:36:23.238 rtt min/avg/max/mdev = 0.180/0.180/0.180/0.000 ms 00:36:23.238 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:36:23.238 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:36:23.238 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.110 ms 00:36:23.238 00:36:23.238 --- 10.0.0.1 ping statistics --- 00:36:23.238 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:36:23.238 rtt min/avg/max/mdev = 0.110/0.110/0.110/0.000 ms 00:36:23.238 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:36:23.238 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@450 -- # return 0 00:36:23.238 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:36:23.238 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:36:23.238 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:36:23.238 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:36:23.238 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:36:23.238 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:36:23.238 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:36:23.238 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@40 -- # nvmfappstart -m 0x2 00:36:23.238 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:36:23.238 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@726 -- # xtrace_disable 00:36:23.238 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:36:23.238 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@509 -- # nvmfpid=387226 00:36:23.238 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@510 -- # waitforlisten 387226 00:36:23.238 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@835 -- # '[' -z 387226 ']' 00:36:23.238 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:36:23.238 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:36:23.238 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@840 -- # local max_retries=100 00:36:23.238 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:36:23.238 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:36:23.238 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@844 -- # xtrace_disable 00:36:23.238 17:52:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:36:23.238 [2024-12-06 17:52:04.907489] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:36:23.238 [2024-12-06 17:52:04.907575] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:36:23.238 [2024-12-06 17:52:04.978232] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:36:23.238 [2024-12-06 17:52:05.020176] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:36:23.238 [2024-12-06 17:52:05.020238] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:36:23.238 [2024-12-06 17:52:05.020266] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:36:23.238 [2024-12-06 17:52:05.020277] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:36:23.239 [2024-12-06 17:52:05.020285] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:36:23.239 [2024-12-06 17:52:05.020875] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:36:23.496 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:36:23.496 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@868 -- # return 0 00:36:23.496 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:36:23.496 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@732 -- # xtrace_disable 00:36:23.496 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:36:23.496 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:36:23.496 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@43 -- # rpc_cmd 00:36:23.496 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:23.496 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:36:23.496 [2024-12-06 17:52:05.165049] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:36:23.496 [2024-12-06 17:52:05.173246] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 8009 *** 00:36:23.496 null0 00:36:23.496 [2024-12-06 17:52:05.205171] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:36:23.496 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:23.496 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@59 -- # hostpid=387261 00:36:23.496 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -m 0x1 -r /tmp/host.sock --wait-for-rpc -L bdev_nvme 00:36:23.496 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@60 -- # waitforlisten 387261 /tmp/host.sock 00:36:23.496 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@835 -- # '[' -z 387261 ']' 00:36:23.496 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@839 -- # local rpc_addr=/tmp/host.sock 00:36:23.496 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@840 -- # local max_retries=100 00:36:23.496 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /tmp/host.sock...' 00:36:23.496 Waiting for process to start up and listen on UNIX domain socket /tmp/host.sock... 00:36:23.496 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@844 -- # xtrace_disable 00:36:23.496 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:36:23.496 [2024-12-06 17:52:05.269689] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:36:23.496 [2024-12-06 17:52:05.269772] [ DPDK EAL parameters: nvmf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid387261 ] 00:36:23.754 [2024-12-06 17:52:05.334381] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:36:23.754 [2024-12-06 17:52:05.379390] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:36:23.754 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:36:23.754 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@868 -- # return 0 00:36:23.754 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@62 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; killprocess $hostpid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:36:23.754 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@65 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_set_options -e 1 00:36:23.754 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:23.754 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:36:23.754 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:23.754 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@66 -- # rpc_cmd -s /tmp/host.sock framework_start_init 00:36:23.754 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:23.754 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:36:23.754 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:23.754 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@69 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test --ctrlr-loss-timeout-sec 2 --reconnect-delay-sec 1 --fast-io-fail-timeout-sec 1 --wait-for-attach 00:36:23.754 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:23.754 17:52:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:36:25.128 [2024-12-06 17:52:06.644260] bdev_nvme.c:7511:discovery_attach_cb: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr attached 00:36:25.128 [2024-12-06 17:52:06.644286] bdev_nvme.c:7597:discovery_poller: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr connected 00:36:25.128 [2024-12-06 17:52:06.644307] bdev_nvme.c:7474:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:36:25.128 [2024-12-06 17:52:06.770705] bdev_nvme.c:7440:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 new subsystem nvme0 00:36:25.128 [2024-12-06 17:52:06.831363] bdev_nvme.c:5656:nvme_ctrlr_create_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr was created to 10.0.0.2:4420 00:36:25.128 [2024-12-06 17:52:06.832326] bdev_nvme.c:1989:bdev_nvme_create_qpair: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Connecting qpair 0x1f24c90:1 started. 00:36:25.128 [2024-12-06 17:52:06.834010] bdev_nvme.c:8307:bdev_nvme_readv: *DEBUG*: read 8 blocks with offset 0 00:36:25.128 [2024-12-06 17:52:06.834065] bdev_nvme.c:8307:bdev_nvme_readv: *DEBUG*: read 1 blocks with offset 0 00:36:25.128 [2024-12-06 17:52:06.834099] bdev_nvme.c:8307:bdev_nvme_readv: *DEBUG*: read 64 blocks with offset 0 00:36:25.128 [2024-12-06 17:52:06.834126] bdev_nvme.c:7330:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme0 done 00:36:25.128 [2024-12-06 17:52:06.834150] bdev_nvme.c:7289:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 found again 00:36:25.128 17:52:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:25.128 17:52:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@72 -- # wait_for_bdev nvme0n1 00:36:25.128 17:52:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:36:25.128 17:52:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:36:25.128 17:52:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:25.128 17:52:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:36:25.128 17:52:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:36:25.128 17:52:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:36:25.128 17:52:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:36:25.128 [2024-12-06 17:52:06.840948] bdev_nvme.c:1791:bdev_nvme_disconnected_qpair_cb: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpair 0x1f24c90 was disconnected and freed. delete nvme_qpair. 00:36:25.128 17:52:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:25.128 17:52:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != \n\v\m\e\0\n\1 ]] 00:36:25.128 17:52:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@75 -- # ip netns exec cvl_0_0_ns_spdk ip addr del 10.0.0.2/24 dev cvl_0_0 00:36:25.128 17:52:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@76 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 down 00:36:25.128 17:52:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@79 -- # wait_for_bdev '' 00:36:25.128 17:52:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:36:25.128 17:52:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:36:25.128 17:52:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:36:25.128 17:52:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:25.128 17:52:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:36:25.128 17:52:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:36:25.128 17:52:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:36:25.128 17:52:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:25.386 17:52:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:36:25.386 17:52:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:36:26.318 17:52:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:36:26.318 17:52:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:36:26.318 17:52:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:36:26.318 17:52:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:26.318 17:52:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:36:26.318 17:52:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:36:26.319 17:52:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:36:26.319 17:52:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:26.319 17:52:08 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:36:26.319 17:52:08 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:36:27.252 17:52:09 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:36:27.252 17:52:09 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:36:27.252 17:52:09 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:36:27.252 17:52:09 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:36:27.252 17:52:09 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:27.252 17:52:09 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:36:27.252 17:52:09 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:36:27.252 17:52:09 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:27.252 17:52:09 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:36:27.252 17:52:09 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:36:28.627 17:52:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:36:28.627 17:52:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:36:28.627 17:52:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:28.627 17:52:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:36:28.627 17:52:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:36:28.627 17:52:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:36:28.627 17:52:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:36:28.627 17:52:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:28.627 17:52:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:36:28.627 17:52:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:36:29.561 17:52:11 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:36:29.561 17:52:11 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:36:29.561 17:52:11 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:36:29.561 17:52:11 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:29.561 17:52:11 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:36:29.561 17:52:11 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:36:29.561 17:52:11 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:36:29.561 17:52:11 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:29.561 17:52:11 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:36:29.561 17:52:11 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:36:30.494 17:52:12 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:36:30.494 17:52:12 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:36:30.494 17:52:12 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:36:30.494 17:52:12 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:30.494 17:52:12 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:36:30.494 17:52:12 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:36:30.494 17:52:12 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:36:30.494 17:52:12 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:30.494 17:52:12 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:36:30.494 17:52:12 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:36:30.494 [2024-12-06 17:52:12.275520] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 421:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 110: Connection timed out 00:36:30.494 [2024-12-06 17:52:12.275605] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:36:30.494 [2024-12-06 17:52:12.275628] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:36:30.494 [2024-12-06 17:52:12.275660] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:36:30.494 [2024-12-06 17:52:12.275682] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:36:30.494 [2024-12-06 17:52:12.275697] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:36:30.494 [2024-12-06 17:52:12.275710] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:36:30.494 [2024-12-06 17:52:12.275723] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:36:30.494 [2024-12-06 17:52:12.275735] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:36:30.494 [2024-12-06 17:52:12.275750] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: KEEP ALIVE (18) qid:0 cid:4 nsid:0 cdw10:00000000 cdw11:00000000 00:36:30.494 [2024-12-06 17:52:12.275763] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:36:30.494 [2024-12-06 17:52:12.275776] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1f01510 is same with the state(6) to be set 00:36:30.494 [2024-12-06 17:52:12.285554] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1f01510 (9): Bad file descriptor 00:36:30.494 [2024-12-06 17:52:12.295596] bdev_nvme.c:2549:bdev_nvme_reset_destroy_qpairs: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Delete qpairs for reset. 00:36:30.495 [2024-12-06 17:52:12.295619] bdev_nvme.c:2537:bdev_nvme_reset_destroy_qpair_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpairs were deleted. 00:36:30.495 [2024-12-06 17:52:12.295632] bdev_nvme.c:2133:nvme_ctrlr_disconnect: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start disconnecting ctrlr. 00:36:30.495 [2024-12-06 17:52:12.295641] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:36:30.495 [2024-12-06 17:52:12.295701] bdev_nvme.c:2521:bdev_nvme_reconnect_ctrlr: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start reconnecting ctrlr. 00:36:31.425 17:52:13 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:36:31.425 17:52:13 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:36:31.425 17:52:13 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:36:31.425 17:52:13 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:31.425 17:52:13 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:36:31.425 17:52:13 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:36:31.425 17:52:13 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:36:31.682 [2024-12-06 17:52:13.352699] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 110 00:36:31.682 [2024-12-06 17:52:13.352764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1f01510 with addr=10.0.0.2, port=4420 00:36:31.682 [2024-12-06 17:52:13.352787] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1f01510 is same with the state(6) to be set 00:36:31.682 [2024-12-06 17:52:13.352824] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1f01510 (9): Bad file descriptor 00:36:31.682 [2024-12-06 17:52:13.353257] bdev_nvme.c:3172:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] Unable to perform failover, already in progress. 00:36:31.682 [2024-12-06 17:52:13.353305] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:36:31.682 [2024-12-06 17:52:13.353322] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:36:31.682 [2024-12-06 17:52:13.353336] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:36:31.682 [2024-12-06 17:52:13.353350] bdev_nvme.c:2511:bdev_nvme_reconnect_ctrlr_poll: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr could not be connected. 00:36:31.682 [2024-12-06 17:52:13.353361] bdev_nvme.c:2278:bdev_nvme_reset_ctrlr_complete: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Clear pending resets. 00:36:31.682 [2024-12-06 17:52:13.353369] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:36:31.682 [2024-12-06 17:52:13.353383] bdev_nvme.c:2133:nvme_ctrlr_disconnect: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start disconnecting ctrlr. 00:36:31.682 [2024-12-06 17:52:13.353391] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:36:31.682 17:52:13 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:31.682 17:52:13 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:36:31.682 17:52:13 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:36:32.674 [2024-12-06 17:52:14.355884] bdev_nvme.c:2521:bdev_nvme_reconnect_ctrlr: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start reconnecting ctrlr. 00:36:32.674 [2024-12-06 17:52:14.355931] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:36:32.674 [2024-12-06 17:52:14.355975] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:36:32.674 [2024-12-06 17:52:14.355989] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:36:32.674 [2024-12-06 17:52:14.356003] nvme_ctrlr.c:1098:nvme_ctrlr_fail: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] already in failed state 00:36:32.674 [2024-12-06 17:52:14.356017] bdev_nvme.c:2511:bdev_nvme_reconnect_ctrlr_poll: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr could not be connected. 00:36:32.674 [2024-12-06 17:52:14.356029] bdev_nvme.c:2278:bdev_nvme_reset_ctrlr_complete: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Clear pending resets. 00:36:32.674 [2024-12-06 17:52:14.356037] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:36:32.674 [2024-12-06 17:52:14.356081] bdev_nvme.c:7262:remove_discovery_entry: *INFO*: Discovery[10.0.0.2:8009] Remove discovery entry: nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 00:36:32.674 [2024-12-06 17:52:14.356143] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:36:32.674 [2024-12-06 17:52:14.356164] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:36:32.674 [2024-12-06 17:52:14.356183] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:36:32.675 [2024-12-06 17:52:14.356197] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:36:32.675 [2024-12-06 17:52:14.356211] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:36:32.675 [2024-12-06 17:52:14.356224] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:36:32.675 [2024-12-06 17:52:14.356237] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:36:32.675 [2024-12-06 17:52:14.356250] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:36:32.675 [2024-12-06 17:52:14.356264] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: KEEP ALIVE (18) qid:0 cid:4 nsid:0 cdw10:00000000 cdw11:00000000 00:36:32.675 [2024-12-06 17:52:14.356283] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:36:32.675 [2024-12-06 17:52:14.356297] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2014-08.org.nvmexpress.discovery, 1] in failed state. 00:36:32.675 [2024-12-06 17:52:14.356346] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1ef0c60 (9): Bad file descriptor 00:36:32.675 [2024-12-06 17:52:14.357336] nvme_fabric.c: 214:nvme_fabric_prop_get_cmd_async: *ERROR*: Failed to send Property Get fabrics command 00:36:32.675 [2024-12-06 17:52:14.357359] nvme_ctrlr.c:1217:nvme_ctrlr_shutdown_async: *ERROR*: [nqn.2014-08.org.nvmexpress.discovery, 1] Failed to read the CC register 00:36:32.675 17:52:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:36:32.675 17:52:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:36:32.675 17:52:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:36:32.675 17:52:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:32.675 17:52:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:36:32.675 17:52:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:36:32.675 17:52:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:36:32.675 17:52:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:32.675 17:52:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ '' != '' ]] 00:36:32.675 17:52:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@82 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:36:32.675 17:52:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@83 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:36:32.675 17:52:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@86 -- # wait_for_bdev nvme1n1 00:36:32.675 17:52:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:36:32.675 17:52:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:36:32.675 17:52:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:36:32.675 17:52:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:32.675 17:52:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:36:32.675 17:52:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:36:32.675 17:52:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:36:32.675 17:52:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:32.675 17:52:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ '' != \n\v\m\e\1\n\1 ]] 00:36:32.675 17:52:14 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:36:34.046 17:52:15 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:36:34.046 17:52:15 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:36:34.046 17:52:15 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:36:34.046 17:52:15 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:34.046 17:52:15 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:36:34.046 17:52:15 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:36:34.046 17:52:15 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:36:34.046 17:52:15 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:34.046 17:52:15 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ '' != \n\v\m\e\1\n\1 ]] 00:36:34.046 17:52:15 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:36:34.611 [2024-12-06 17:52:16.408362] bdev_nvme.c:7511:discovery_attach_cb: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr attached 00:36:34.611 [2024-12-06 17:52:16.408401] bdev_nvme.c:7597:discovery_poller: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr connected 00:36:34.611 [2024-12-06 17:52:16.408424] bdev_nvme.c:7474:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:36:34.869 [2024-12-06 17:52:16.536835] bdev_nvme.c:7440:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 new subsystem nvme1 00:36:34.869 17:52:16 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:36:34.869 17:52:16 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:36:34.869 17:52:16 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:36:34.869 17:52:16 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:34.869 17:52:16 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:36:34.869 17:52:16 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:36:34.869 17:52:16 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:36:34.870 17:52:16 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:34.870 17:52:16 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ '' != \n\v\m\e\1\n\1 ]] 00:36:34.870 17:52:16 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:36:35.127 [2024-12-06 17:52:16.759181] bdev_nvme.c:5656:nvme_ctrlr_create_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 2] ctrlr was created to 10.0.0.2:4420 00:36:35.127 [2024-12-06 17:52:16.760237] bdev_nvme.c:1989:bdev_nvme_create_qpair: *INFO*: [nqn.2016-06.io.spdk:cnode0, 2] Connecting qpair 0x1f2f2f0:1 started. 00:36:35.127 [2024-12-06 17:52:16.761597] bdev_nvme.c:8307:bdev_nvme_readv: *DEBUG*: read 8 blocks with offset 0 00:36:35.127 [2024-12-06 17:52:16.761641] bdev_nvme.c:8307:bdev_nvme_readv: *DEBUG*: read 1 blocks with offset 0 00:36:35.127 [2024-12-06 17:52:16.761695] bdev_nvme.c:8307:bdev_nvme_readv: *DEBUG*: read 64 blocks with offset 0 00:36:35.127 [2024-12-06 17:52:16.761720] bdev_nvme.c:7330:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme1 done 00:36:35.127 [2024-12-06 17:52:16.761735] bdev_nvme.c:7289:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 found again 00:36:35.127 [2024-12-06 17:52:16.767281] bdev_nvme.c:1791:bdev_nvme_disconnected_qpair_cb: *INFO*: [nqn.2016-06.io.spdk:cnode0, 2] qpair 0x1f2f2f0 was disconnected and freed. delete nvme_qpair. 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme1n1 != \n\v\m\e\1\n\1 ]] 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@88 -- # trap - SIGINT SIGTERM EXIT 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@90 -- # killprocess 387261 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@954 -- # '[' -z 387261 ']' 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@958 -- # kill -0 387261 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@959 -- # uname 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 387261 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 387261' 00:36:36.063 killing process with pid 387261 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@973 -- # kill 387261 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@978 -- # wait 387261 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@91 -- # nvmftestfini 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@516 -- # nvmfcleanup 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@121 -- # sync 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@124 -- # set +e 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@125 -- # for i in {1..20} 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:36:36.063 rmmod nvme_tcp 00:36:36.063 rmmod nvme_fabrics 00:36:36.063 rmmod nvme_keyring 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@128 -- # set -e 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@129 -- # return 0 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@517 -- # '[' -n 387226 ']' 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@518 -- # killprocess 387226 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@954 -- # '[' -z 387226 ']' 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@958 -- # kill -0 387226 00:36:36.063 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@959 -- # uname 00:36:36.322 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:36:36.322 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 387226 00:36:36.322 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:36:36.322 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:36:36.322 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@972 -- # echo 'killing process with pid 387226' 00:36:36.322 killing process with pid 387226 00:36:36.322 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@973 -- # kill 387226 00:36:36.322 17:52:17 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@978 -- # wait 387226 00:36:36.322 17:52:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:36:36.322 17:52:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:36:36.322 17:52:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:36:36.322 17:52:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@297 -- # iptr 00:36:36.322 17:52:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@791 -- # iptables-save 00:36:36.322 17:52:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:36:36.322 17:52:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@791 -- # iptables-restore 00:36:36.322 17:52:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:36:36.322 17:52:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@302 -- # remove_spdk_ns 00:36:36.322 17:52:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:36:36.322 17:52:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:36:36.322 17:52:18 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:36:38.859 00:36:38.859 real 0m17.768s 00:36:38.859 user 0m25.592s 00:36:38.859 sys 0m3.107s 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1130 -- # xtrace_disable 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:36:38.859 ************************************ 00:36:38.859 END TEST nvmf_discovery_remove_ifc 00:36:38.859 ************************************ 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@29 -- # run_test nvmf_identify_kernel_target /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/identify_kernel_nvmf.sh --transport=tcp 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:36:38.859 ************************************ 00:36:38.859 START TEST nvmf_identify_kernel_target 00:36:38.859 ************************************ 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/identify_kernel_nvmf.sh --transport=tcp 00:36:38.859 * Looking for test storage... 00:36:38.859 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1711 -- # lcov --version 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@333 -- # local ver1 ver1_l 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@334 -- # local ver2 ver2_l 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@336 -- # IFS=.-: 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@336 -- # read -ra ver1 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@337 -- # IFS=.-: 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@337 -- # read -ra ver2 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@338 -- # local 'op=<' 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@340 -- # ver1_l=2 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@341 -- # ver2_l=1 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@344 -- # case "$op" in 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@345 -- # : 1 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@364 -- # (( v = 0 )) 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@365 -- # decimal 1 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@353 -- # local d=1 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@355 -- # echo 1 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@365 -- # ver1[v]=1 00:36:38.859 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@366 -- # decimal 2 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@353 -- # local d=2 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@355 -- # echo 2 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@366 -- # ver2[v]=2 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@368 -- # return 0 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:36:38.860 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:38.860 --rc genhtml_branch_coverage=1 00:36:38.860 --rc genhtml_function_coverage=1 00:36:38.860 --rc genhtml_legend=1 00:36:38.860 --rc geninfo_all_blocks=1 00:36:38.860 --rc geninfo_unexecuted_blocks=1 00:36:38.860 00:36:38.860 ' 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:36:38.860 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:38.860 --rc genhtml_branch_coverage=1 00:36:38.860 --rc genhtml_function_coverage=1 00:36:38.860 --rc genhtml_legend=1 00:36:38.860 --rc geninfo_all_blocks=1 00:36:38.860 --rc geninfo_unexecuted_blocks=1 00:36:38.860 00:36:38.860 ' 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:36:38.860 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:38.860 --rc genhtml_branch_coverage=1 00:36:38.860 --rc genhtml_function_coverage=1 00:36:38.860 --rc genhtml_legend=1 00:36:38.860 --rc geninfo_all_blocks=1 00:36:38.860 --rc geninfo_unexecuted_blocks=1 00:36:38.860 00:36:38.860 ' 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:36:38.860 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:38.860 --rc genhtml_branch_coverage=1 00:36:38.860 --rc genhtml_function_coverage=1 00:36:38.860 --rc genhtml_legend=1 00:36:38.860 --rc geninfo_all_blocks=1 00:36:38.860 --rc geninfo_unexecuted_blocks=1 00:36:38.860 00:36:38.860 ' 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@7 -- # uname -s 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@15 -- # shopt -s extglob 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- paths/export.sh@5 -- # export PATH 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@51 -- # : 0 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:36:38.860 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@55 -- # have_pci_nics=0 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@11 -- # nvmftestinit 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@476 -- # prepare_net_devs 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@438 -- # local -g is_hw=no 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@440 -- # remove_spdk_ns 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@309 -- # xtrace_disable 00:36:38.860 17:52:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@10 -- # set +x 00:36:40.761 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:36:40.761 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@315 -- # pci_devs=() 00:36:40.761 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@315 -- # local -a pci_devs 00:36:40.761 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@316 -- # pci_net_devs=() 00:36:40.761 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:36:40.761 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@317 -- # pci_drivers=() 00:36:40.761 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@317 -- # local -A pci_drivers 00:36:40.761 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@319 -- # net_devs=() 00:36:40.761 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@319 -- # local -ga net_devs 00:36:40.761 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@320 -- # e810=() 00:36:40.761 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@320 -- # local -ga e810 00:36:40.761 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@321 -- # x722=() 00:36:40.761 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@321 -- # local -ga x722 00:36:40.761 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@322 -- # mlx=() 00:36:40.761 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@322 -- # local -ga mlx 00:36:40.761 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:36:40.761 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:36:40.761 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:36:40.761 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:36:40.761 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:36:40.761 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:36:40.761 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:36:40.761 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:36:40.762 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:36:40.762 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@418 -- # [[ up == up ]] 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:36:40.762 Found net devices under 0000:0a:00.0: cvl_0_0 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@418 -- # [[ up == up ]] 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:36:40.762 Found net devices under 0000:0a:00.1: cvl_0_1 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@442 -- # is_hw=yes 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:36:40.762 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:36:41.034 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:36:41.034 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:36:41.034 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:36:41.034 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:36:41.034 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:36:41.034 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.224 ms 00:36:41.034 00:36:41.034 --- 10.0.0.2 ping statistics --- 00:36:41.034 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:36:41.034 rtt min/avg/max/mdev = 0.224/0.224/0.224/0.000 ms 00:36:41.034 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:36:41.034 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:36:41.034 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.187 ms 00:36:41.034 00:36:41.034 --- 10.0.0.1 ping statistics --- 00:36:41.034 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:36:41.034 rtt min/avg/max/mdev = 0.187/0.187/0.187/0.000 ms 00:36:41.035 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:36:41.035 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@450 -- # return 0 00:36:41.035 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:36:41.035 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:36:41.035 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:36:41.035 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:36:41.035 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:36:41.035 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:36:41.035 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:36:41.035 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@13 -- # trap 'nvmftestfini || :; clean_kernel_target' EXIT 00:36:41.035 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@15 -- # get_main_ns_ip 00:36:41.035 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@769 -- # local ip 00:36:41.035 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@770 -- # ip_candidates=() 00:36:41.035 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@770 -- # local -A ip_candidates 00:36:41.035 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:36:41.035 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:36:41.035 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:36:41.035 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:36:41.035 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:36:41.035 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:36:41.035 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:36:41.035 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@15 -- # target_ip=10.0.0.1 00:36:41.035 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@16 -- # configure_kernel_target nqn.2016-06.io.spdk:testnqn 10.0.0.1 00:36:41.035 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@660 -- # local kernel_name=nqn.2016-06.io.spdk:testnqn kernel_target_ip=10.0.0.1 00:36:41.035 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@662 -- # nvmet=/sys/kernel/config/nvmet 00:36:41.035 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@663 -- # kernel_subsystem=/sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:36:41.036 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@664 -- # kernel_namespace=/sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:36:41.036 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@665 -- # kernel_port=/sys/kernel/config/nvmet/ports/1 00:36:41.036 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@667 -- # local block nvme 00:36:41.036 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@669 -- # [[ ! -e /sys/module/nvmet ]] 00:36:41.036 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@670 -- # modprobe nvmet 00:36:41.036 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@673 -- # [[ -e /sys/kernel/config/nvmet ]] 00:36:41.036 17:52:22 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@675 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:36:41.973 Waiting for block devices as requested 00:36:42.231 0000:88:00.0 (8086 0a54): vfio-pci -> nvme 00:36:42.231 0000:00:04.7 (8086 0e27): vfio-pci -> ioatdma 00:36:42.490 0000:00:04.6 (8086 0e26): vfio-pci -> ioatdma 00:36:42.490 0000:00:04.5 (8086 0e25): vfio-pci -> ioatdma 00:36:42.490 0000:00:04.4 (8086 0e24): vfio-pci -> ioatdma 00:36:42.748 0000:00:04.3 (8086 0e23): vfio-pci -> ioatdma 00:36:42.748 0000:00:04.2 (8086 0e22): vfio-pci -> ioatdma 00:36:42.748 0000:00:04.1 (8086 0e21): vfio-pci -> ioatdma 00:36:42.748 0000:00:04.0 (8086 0e20): vfio-pci -> ioatdma 00:36:43.007 0000:80:04.7 (8086 0e27): vfio-pci -> ioatdma 00:36:43.007 0000:80:04.6 (8086 0e26): vfio-pci -> ioatdma 00:36:43.007 0000:80:04.5 (8086 0e25): vfio-pci -> ioatdma 00:36:43.007 0000:80:04.4 (8086 0e24): vfio-pci -> ioatdma 00:36:43.267 0000:80:04.3 (8086 0e23): vfio-pci -> ioatdma 00:36:43.267 0000:80:04.2 (8086 0e22): vfio-pci -> ioatdma 00:36:43.267 0000:80:04.1 (8086 0e21): vfio-pci -> ioatdma 00:36:43.267 0000:80:04.0 (8086 0e20): vfio-pci -> ioatdma 00:36:43.528 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@678 -- # for block in /sys/block/nvme* 00:36:43.528 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@679 -- # [[ -e /sys/block/nvme0n1 ]] 00:36:43.528 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@680 -- # is_block_zoned nvme0n1 00:36:43.528 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1650 -- # local device=nvme0n1 00:36:43.528 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:36:43.528 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:36:43.528 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@681 -- # block_in_use nvme0n1 00:36:43.528 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@381 -- # local block=nvme0n1 pt 00:36:43.528 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@390 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdk-gpt.py nvme0n1 00:36:43.528 No valid GPT data, bailing 00:36:43.528 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:36:43.528 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@394 -- # pt= 00:36:43.528 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@395 -- # return 1 00:36:43.528 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@681 -- # nvme=/dev/nvme0n1 00:36:43.528 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@684 -- # [[ -b /dev/nvme0n1 ]] 00:36:43.528 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@686 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:36:43.528 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@687 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:36:43.528 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@688 -- # mkdir /sys/kernel/config/nvmet/ports/1 00:36:43.528 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@693 -- # echo SPDK-nqn.2016-06.io.spdk:testnqn 00:36:43.528 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@695 -- # echo 1 00:36:43.528 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@696 -- # echo /dev/nvme0n1 00:36:43.528 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@697 -- # echo 1 00:36:43.528 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@699 -- # echo 10.0.0.1 00:36:43.528 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@700 -- # echo tcp 00:36:43.528 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@701 -- # echo 4420 00:36:43.528 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@702 -- # echo ipv4 00:36:43.528 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@705 -- # ln -s /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn /sys/kernel/config/nvmet/ports/1/subsystems/ 00:36:43.528 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@708 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -a 10.0.0.1 -t tcp -s 4420 00:36:43.528 00:36:43.528 Discovery Log Number of Records 2, Generation counter 2 00:36:43.528 =====Discovery Log Entry 0====== 00:36:43.528 trtype: tcp 00:36:43.528 adrfam: ipv4 00:36:43.528 subtype: current discovery subsystem 00:36:43.528 treq: not specified, sq flow control disable supported 00:36:43.528 portid: 1 00:36:43.528 trsvcid: 4420 00:36:43.528 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:36:43.528 traddr: 10.0.0.1 00:36:43.528 eflags: none 00:36:43.528 sectype: none 00:36:43.528 =====Discovery Log Entry 1====== 00:36:43.528 trtype: tcp 00:36:43.528 adrfam: ipv4 00:36:43.528 subtype: nvme subsystem 00:36:43.528 treq: not specified, sq flow control disable supported 00:36:43.528 portid: 1 00:36:43.528 trsvcid: 4420 00:36:43.528 subnqn: nqn.2016-06.io.spdk:testnqn 00:36:43.528 traddr: 10.0.0.1 00:36:43.528 eflags: none 00:36:43.528 sectype: none 00:36:43.528 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.1 00:36:43.528 trsvcid:4420 subnqn:nqn.2014-08.org.nvmexpress.discovery' 00:36:43.790 ===================================================== 00:36:43.790 NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2014-08.org.nvmexpress.discovery 00:36:43.790 ===================================================== 00:36:43.790 Controller Capabilities/Features 00:36:43.790 ================================ 00:36:43.790 Vendor ID: 0000 00:36:43.790 Subsystem Vendor ID: 0000 00:36:43.790 Serial Number: 2c6be115bf2bea6cb6cc 00:36:43.790 Model Number: Linux 00:36:43.790 Firmware Version: 6.8.9-20 00:36:43.790 Recommended Arb Burst: 0 00:36:43.790 IEEE OUI Identifier: 00 00 00 00:36:43.790 Multi-path I/O 00:36:43.790 May have multiple subsystem ports: No 00:36:43.790 May have multiple controllers: No 00:36:43.790 Associated with SR-IOV VF: No 00:36:43.790 Max Data Transfer Size: Unlimited 00:36:43.790 Max Number of Namespaces: 0 00:36:43.790 Max Number of I/O Queues: 1024 00:36:43.790 NVMe Specification Version (VS): 1.3 00:36:43.790 NVMe Specification Version (Identify): 1.3 00:36:43.790 Maximum Queue Entries: 1024 00:36:43.790 Contiguous Queues Required: No 00:36:43.790 Arbitration Mechanisms Supported 00:36:43.790 Weighted Round Robin: Not Supported 00:36:43.790 Vendor Specific: Not Supported 00:36:43.790 Reset Timeout: 7500 ms 00:36:43.790 Doorbell Stride: 4 bytes 00:36:43.790 NVM Subsystem Reset: Not Supported 00:36:43.790 Command Sets Supported 00:36:43.790 NVM Command Set: Supported 00:36:43.790 Boot Partition: Not Supported 00:36:43.790 Memory Page Size Minimum: 4096 bytes 00:36:43.790 Memory Page Size Maximum: 4096 bytes 00:36:43.790 Persistent Memory Region: Not Supported 00:36:43.790 Optional Asynchronous Events Supported 00:36:43.790 Namespace Attribute Notices: Not Supported 00:36:43.790 Firmware Activation Notices: Not Supported 00:36:43.790 ANA Change Notices: Not Supported 00:36:43.790 PLE Aggregate Log Change Notices: Not Supported 00:36:43.790 LBA Status Info Alert Notices: Not Supported 00:36:43.790 EGE Aggregate Log Change Notices: Not Supported 00:36:43.790 Normal NVM Subsystem Shutdown event: Not Supported 00:36:43.790 Zone Descriptor Change Notices: Not Supported 00:36:43.790 Discovery Log Change Notices: Supported 00:36:43.790 Controller Attributes 00:36:43.790 128-bit Host Identifier: Not Supported 00:36:43.790 Non-Operational Permissive Mode: Not Supported 00:36:43.790 NVM Sets: Not Supported 00:36:43.790 Read Recovery Levels: Not Supported 00:36:43.790 Endurance Groups: Not Supported 00:36:43.790 Predictable Latency Mode: Not Supported 00:36:43.790 Traffic Based Keep ALive: Not Supported 00:36:43.790 Namespace Granularity: Not Supported 00:36:43.790 SQ Associations: Not Supported 00:36:43.790 UUID List: Not Supported 00:36:43.790 Multi-Domain Subsystem: Not Supported 00:36:43.790 Fixed Capacity Management: Not Supported 00:36:43.790 Variable Capacity Management: Not Supported 00:36:43.790 Delete Endurance Group: Not Supported 00:36:43.790 Delete NVM Set: Not Supported 00:36:43.790 Extended LBA Formats Supported: Not Supported 00:36:43.790 Flexible Data Placement Supported: Not Supported 00:36:43.790 00:36:43.790 Controller Memory Buffer Support 00:36:43.790 ================================ 00:36:43.790 Supported: No 00:36:43.790 00:36:43.790 Persistent Memory Region Support 00:36:43.790 ================================ 00:36:43.790 Supported: No 00:36:43.790 00:36:43.790 Admin Command Set Attributes 00:36:43.790 ============================ 00:36:43.790 Security Send/Receive: Not Supported 00:36:43.790 Format NVM: Not Supported 00:36:43.790 Firmware Activate/Download: Not Supported 00:36:43.790 Namespace Management: Not Supported 00:36:43.790 Device Self-Test: Not Supported 00:36:43.790 Directives: Not Supported 00:36:43.790 NVMe-MI: Not Supported 00:36:43.790 Virtualization Management: Not Supported 00:36:43.790 Doorbell Buffer Config: Not Supported 00:36:43.790 Get LBA Status Capability: Not Supported 00:36:43.790 Command & Feature Lockdown Capability: Not Supported 00:36:43.790 Abort Command Limit: 1 00:36:43.790 Async Event Request Limit: 1 00:36:43.790 Number of Firmware Slots: N/A 00:36:43.790 Firmware Slot 1 Read-Only: N/A 00:36:43.790 Firmware Activation Without Reset: N/A 00:36:43.790 Multiple Update Detection Support: N/A 00:36:43.790 Firmware Update Granularity: No Information Provided 00:36:43.790 Per-Namespace SMART Log: No 00:36:43.790 Asymmetric Namespace Access Log Page: Not Supported 00:36:43.790 Subsystem NQN: nqn.2014-08.org.nvmexpress.discovery 00:36:43.790 Command Effects Log Page: Not Supported 00:36:43.790 Get Log Page Extended Data: Supported 00:36:43.790 Telemetry Log Pages: Not Supported 00:36:43.790 Persistent Event Log Pages: Not Supported 00:36:43.790 Supported Log Pages Log Page: May Support 00:36:43.790 Commands Supported & Effects Log Page: Not Supported 00:36:43.790 Feature Identifiers & Effects Log Page:May Support 00:36:43.790 NVMe-MI Commands & Effects Log Page: May Support 00:36:43.790 Data Area 4 for Telemetry Log: Not Supported 00:36:43.790 Error Log Page Entries Supported: 1 00:36:43.790 Keep Alive: Not Supported 00:36:43.790 00:36:43.790 NVM Command Set Attributes 00:36:43.790 ========================== 00:36:43.790 Submission Queue Entry Size 00:36:43.790 Max: 1 00:36:43.790 Min: 1 00:36:43.790 Completion Queue Entry Size 00:36:43.790 Max: 1 00:36:43.790 Min: 1 00:36:43.790 Number of Namespaces: 0 00:36:43.790 Compare Command: Not Supported 00:36:43.790 Write Uncorrectable Command: Not Supported 00:36:43.790 Dataset Management Command: Not Supported 00:36:43.790 Write Zeroes Command: Not Supported 00:36:43.790 Set Features Save Field: Not Supported 00:36:43.790 Reservations: Not Supported 00:36:43.790 Timestamp: Not Supported 00:36:43.790 Copy: Not Supported 00:36:43.790 Volatile Write Cache: Not Present 00:36:43.790 Atomic Write Unit (Normal): 1 00:36:43.790 Atomic Write Unit (PFail): 1 00:36:43.790 Atomic Compare & Write Unit: 1 00:36:43.790 Fused Compare & Write: Not Supported 00:36:43.790 Scatter-Gather List 00:36:43.790 SGL Command Set: Supported 00:36:43.790 SGL Keyed: Not Supported 00:36:43.790 SGL Bit Bucket Descriptor: Not Supported 00:36:43.790 SGL Metadata Pointer: Not Supported 00:36:43.790 Oversized SGL: Not Supported 00:36:43.790 SGL Metadata Address: Not Supported 00:36:43.790 SGL Offset: Supported 00:36:43.790 Transport SGL Data Block: Not Supported 00:36:43.790 Replay Protected Memory Block: Not Supported 00:36:43.790 00:36:43.790 Firmware Slot Information 00:36:43.790 ========================= 00:36:43.790 Active slot: 0 00:36:43.790 00:36:43.790 00:36:43.790 Error Log 00:36:43.790 ========= 00:36:43.790 00:36:43.790 Active Namespaces 00:36:43.790 ================= 00:36:43.790 Discovery Log Page 00:36:43.790 ================== 00:36:43.790 Generation Counter: 2 00:36:43.790 Number of Records: 2 00:36:43.790 Record Format: 0 00:36:43.790 00:36:43.790 Discovery Log Entry 0 00:36:43.790 ---------------------- 00:36:43.790 Transport Type: 3 (TCP) 00:36:43.790 Address Family: 1 (IPv4) 00:36:43.790 Subsystem Type: 3 (Current Discovery Subsystem) 00:36:43.790 Entry Flags: 00:36:43.790 Duplicate Returned Information: 0 00:36:43.791 Explicit Persistent Connection Support for Discovery: 0 00:36:43.791 Transport Requirements: 00:36:43.791 Secure Channel: Not Specified 00:36:43.791 Port ID: 1 (0x0001) 00:36:43.791 Controller ID: 65535 (0xffff) 00:36:43.791 Admin Max SQ Size: 32 00:36:43.791 Transport Service Identifier: 4420 00:36:43.791 NVM Subsystem Qualified Name: nqn.2014-08.org.nvmexpress.discovery 00:36:43.791 Transport Address: 10.0.0.1 00:36:43.791 Discovery Log Entry 1 00:36:43.791 ---------------------- 00:36:43.791 Transport Type: 3 (TCP) 00:36:43.791 Address Family: 1 (IPv4) 00:36:43.791 Subsystem Type: 2 (NVM Subsystem) 00:36:43.791 Entry Flags: 00:36:43.791 Duplicate Returned Information: 0 00:36:43.791 Explicit Persistent Connection Support for Discovery: 0 00:36:43.791 Transport Requirements: 00:36:43.791 Secure Channel: Not Specified 00:36:43.791 Port ID: 1 (0x0001) 00:36:43.791 Controller ID: 65535 (0xffff) 00:36:43.791 Admin Max SQ Size: 32 00:36:43.791 Transport Service Identifier: 4420 00:36:43.791 NVM Subsystem Qualified Name: nqn.2016-06.io.spdk:testnqn 00:36:43.791 Transport Address: 10.0.0.1 00:36:43.791 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:36:43.791 get_feature(0x01) failed 00:36:43.791 get_feature(0x02) failed 00:36:43.791 get_feature(0x04) failed 00:36:43.791 ===================================================== 00:36:43.791 NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2016-06.io.spdk:testnqn 00:36:43.791 ===================================================== 00:36:43.791 Controller Capabilities/Features 00:36:43.791 ================================ 00:36:43.791 Vendor ID: 0000 00:36:43.791 Subsystem Vendor ID: 0000 00:36:43.791 Serial Number: 6555a5d399677b356935 00:36:43.791 Model Number: SPDK-nqn.2016-06.io.spdk:testnqn 00:36:43.791 Firmware Version: 6.8.9-20 00:36:43.791 Recommended Arb Burst: 6 00:36:43.791 IEEE OUI Identifier: 00 00 00 00:36:43.791 Multi-path I/O 00:36:43.791 May have multiple subsystem ports: Yes 00:36:43.791 May have multiple controllers: Yes 00:36:43.791 Associated with SR-IOV VF: No 00:36:43.791 Max Data Transfer Size: Unlimited 00:36:43.791 Max Number of Namespaces: 1024 00:36:43.791 Max Number of I/O Queues: 128 00:36:43.791 NVMe Specification Version (VS): 1.3 00:36:43.791 NVMe Specification Version (Identify): 1.3 00:36:43.791 Maximum Queue Entries: 1024 00:36:43.791 Contiguous Queues Required: No 00:36:43.791 Arbitration Mechanisms Supported 00:36:43.791 Weighted Round Robin: Not Supported 00:36:43.791 Vendor Specific: Not Supported 00:36:43.791 Reset Timeout: 7500 ms 00:36:43.791 Doorbell Stride: 4 bytes 00:36:43.791 NVM Subsystem Reset: Not Supported 00:36:43.791 Command Sets Supported 00:36:43.791 NVM Command Set: Supported 00:36:43.791 Boot Partition: Not Supported 00:36:43.791 Memory Page Size Minimum: 4096 bytes 00:36:43.791 Memory Page Size Maximum: 4096 bytes 00:36:43.791 Persistent Memory Region: Not Supported 00:36:43.791 Optional Asynchronous Events Supported 00:36:43.791 Namespace Attribute Notices: Supported 00:36:43.791 Firmware Activation Notices: Not Supported 00:36:43.791 ANA Change Notices: Supported 00:36:43.791 PLE Aggregate Log Change Notices: Not Supported 00:36:43.791 LBA Status Info Alert Notices: Not Supported 00:36:43.791 EGE Aggregate Log Change Notices: Not Supported 00:36:43.791 Normal NVM Subsystem Shutdown event: Not Supported 00:36:43.791 Zone Descriptor Change Notices: Not Supported 00:36:43.791 Discovery Log Change Notices: Not Supported 00:36:43.791 Controller Attributes 00:36:43.791 128-bit Host Identifier: Supported 00:36:43.791 Non-Operational Permissive Mode: Not Supported 00:36:43.791 NVM Sets: Not Supported 00:36:43.791 Read Recovery Levels: Not Supported 00:36:43.791 Endurance Groups: Not Supported 00:36:43.791 Predictable Latency Mode: Not Supported 00:36:43.791 Traffic Based Keep ALive: Supported 00:36:43.791 Namespace Granularity: Not Supported 00:36:43.791 SQ Associations: Not Supported 00:36:43.791 UUID List: Not Supported 00:36:43.791 Multi-Domain Subsystem: Not Supported 00:36:43.791 Fixed Capacity Management: Not Supported 00:36:43.791 Variable Capacity Management: Not Supported 00:36:43.791 Delete Endurance Group: Not Supported 00:36:43.791 Delete NVM Set: Not Supported 00:36:43.791 Extended LBA Formats Supported: Not Supported 00:36:43.791 Flexible Data Placement Supported: Not Supported 00:36:43.791 00:36:43.791 Controller Memory Buffer Support 00:36:43.791 ================================ 00:36:43.791 Supported: No 00:36:43.791 00:36:43.791 Persistent Memory Region Support 00:36:43.791 ================================ 00:36:43.791 Supported: No 00:36:43.791 00:36:43.791 Admin Command Set Attributes 00:36:43.791 ============================ 00:36:43.791 Security Send/Receive: Not Supported 00:36:43.791 Format NVM: Not Supported 00:36:43.791 Firmware Activate/Download: Not Supported 00:36:43.791 Namespace Management: Not Supported 00:36:43.791 Device Self-Test: Not Supported 00:36:43.791 Directives: Not Supported 00:36:43.791 NVMe-MI: Not Supported 00:36:43.791 Virtualization Management: Not Supported 00:36:43.791 Doorbell Buffer Config: Not Supported 00:36:43.791 Get LBA Status Capability: Not Supported 00:36:43.791 Command & Feature Lockdown Capability: Not Supported 00:36:43.791 Abort Command Limit: 4 00:36:43.791 Async Event Request Limit: 4 00:36:43.791 Number of Firmware Slots: N/A 00:36:43.791 Firmware Slot 1 Read-Only: N/A 00:36:43.791 Firmware Activation Without Reset: N/A 00:36:43.791 Multiple Update Detection Support: N/A 00:36:43.791 Firmware Update Granularity: No Information Provided 00:36:43.791 Per-Namespace SMART Log: Yes 00:36:43.791 Asymmetric Namespace Access Log Page: Supported 00:36:43.791 ANA Transition Time : 10 sec 00:36:43.791 00:36:43.791 Asymmetric Namespace Access Capabilities 00:36:43.791 ANA Optimized State : Supported 00:36:43.791 ANA Non-Optimized State : Supported 00:36:43.791 ANA Inaccessible State : Supported 00:36:43.791 ANA Persistent Loss State : Supported 00:36:43.791 ANA Change State : Supported 00:36:43.791 ANAGRPID is not changed : No 00:36:43.791 Non-Zero ANAGRPID for NS Mgmt Cmd : Not Supported 00:36:43.791 00:36:43.791 ANA Group Identifier Maximum : 128 00:36:43.791 Number of ANA Group Identifiers : 128 00:36:43.791 Max Number of Allowed Namespaces : 1024 00:36:43.791 Subsystem NQN: nqn.2016-06.io.spdk:testnqn 00:36:43.791 Command Effects Log Page: Supported 00:36:43.791 Get Log Page Extended Data: Supported 00:36:43.791 Telemetry Log Pages: Not Supported 00:36:43.791 Persistent Event Log Pages: Not Supported 00:36:43.791 Supported Log Pages Log Page: May Support 00:36:43.791 Commands Supported & Effects Log Page: Not Supported 00:36:43.791 Feature Identifiers & Effects Log Page:May Support 00:36:43.791 NVMe-MI Commands & Effects Log Page: May Support 00:36:43.791 Data Area 4 for Telemetry Log: Not Supported 00:36:43.791 Error Log Page Entries Supported: 128 00:36:43.791 Keep Alive: Supported 00:36:43.791 Keep Alive Granularity: 1000 ms 00:36:43.791 00:36:43.791 NVM Command Set Attributes 00:36:43.791 ========================== 00:36:43.791 Submission Queue Entry Size 00:36:43.791 Max: 64 00:36:43.791 Min: 64 00:36:43.791 Completion Queue Entry Size 00:36:43.791 Max: 16 00:36:43.791 Min: 16 00:36:43.791 Number of Namespaces: 1024 00:36:43.791 Compare Command: Not Supported 00:36:43.791 Write Uncorrectable Command: Not Supported 00:36:43.791 Dataset Management Command: Supported 00:36:43.791 Write Zeroes Command: Supported 00:36:43.791 Set Features Save Field: Not Supported 00:36:43.791 Reservations: Not Supported 00:36:43.791 Timestamp: Not Supported 00:36:43.791 Copy: Not Supported 00:36:43.791 Volatile Write Cache: Present 00:36:43.791 Atomic Write Unit (Normal): 1 00:36:43.791 Atomic Write Unit (PFail): 1 00:36:43.791 Atomic Compare & Write Unit: 1 00:36:43.791 Fused Compare & Write: Not Supported 00:36:43.791 Scatter-Gather List 00:36:43.791 SGL Command Set: Supported 00:36:43.791 SGL Keyed: Not Supported 00:36:43.791 SGL Bit Bucket Descriptor: Not Supported 00:36:43.791 SGL Metadata Pointer: Not Supported 00:36:43.791 Oversized SGL: Not Supported 00:36:43.791 SGL Metadata Address: Not Supported 00:36:43.791 SGL Offset: Supported 00:36:43.791 Transport SGL Data Block: Not Supported 00:36:43.791 Replay Protected Memory Block: Not Supported 00:36:43.791 00:36:43.791 Firmware Slot Information 00:36:43.791 ========================= 00:36:43.791 Active slot: 0 00:36:43.791 00:36:43.791 Asymmetric Namespace Access 00:36:43.791 =========================== 00:36:43.791 Change Count : 0 00:36:43.791 Number of ANA Group Descriptors : 1 00:36:43.791 ANA Group Descriptor : 0 00:36:43.791 ANA Group ID : 1 00:36:43.791 Number of NSID Values : 1 00:36:43.791 Change Count : 0 00:36:43.791 ANA State : 1 00:36:43.791 Namespace Identifier : 1 00:36:43.791 00:36:43.791 Commands Supported and Effects 00:36:43.791 ============================== 00:36:43.791 Admin Commands 00:36:43.791 -------------- 00:36:43.791 Get Log Page (02h): Supported 00:36:43.791 Identify (06h): Supported 00:36:43.791 Abort (08h): Supported 00:36:43.791 Set Features (09h): Supported 00:36:43.791 Get Features (0Ah): Supported 00:36:43.791 Asynchronous Event Request (0Ch): Supported 00:36:43.791 Keep Alive (18h): Supported 00:36:43.792 I/O Commands 00:36:43.792 ------------ 00:36:43.792 Flush (00h): Supported 00:36:43.792 Write (01h): Supported LBA-Change 00:36:43.792 Read (02h): Supported 00:36:43.792 Write Zeroes (08h): Supported LBA-Change 00:36:43.792 Dataset Management (09h): Supported 00:36:43.792 00:36:43.792 Error Log 00:36:43.792 ========= 00:36:43.792 Entry: 0 00:36:43.792 Error Count: 0x3 00:36:43.792 Submission Queue Id: 0x0 00:36:43.792 Command Id: 0x5 00:36:43.792 Phase Bit: 0 00:36:43.792 Status Code: 0x2 00:36:43.792 Status Code Type: 0x0 00:36:43.792 Do Not Retry: 1 00:36:43.792 Error Location: 0x28 00:36:43.792 LBA: 0x0 00:36:43.792 Namespace: 0x0 00:36:43.792 Vendor Log Page: 0x0 00:36:43.792 ----------- 00:36:43.792 Entry: 1 00:36:43.792 Error Count: 0x2 00:36:43.792 Submission Queue Id: 0x0 00:36:43.792 Command Id: 0x5 00:36:43.792 Phase Bit: 0 00:36:43.792 Status Code: 0x2 00:36:43.792 Status Code Type: 0x0 00:36:43.792 Do Not Retry: 1 00:36:43.792 Error Location: 0x28 00:36:43.792 LBA: 0x0 00:36:43.792 Namespace: 0x0 00:36:43.792 Vendor Log Page: 0x0 00:36:43.792 ----------- 00:36:43.792 Entry: 2 00:36:43.792 Error Count: 0x1 00:36:43.792 Submission Queue Id: 0x0 00:36:43.792 Command Id: 0x4 00:36:43.792 Phase Bit: 0 00:36:43.792 Status Code: 0x2 00:36:43.792 Status Code Type: 0x0 00:36:43.792 Do Not Retry: 1 00:36:43.792 Error Location: 0x28 00:36:43.792 LBA: 0x0 00:36:43.792 Namespace: 0x0 00:36:43.792 Vendor Log Page: 0x0 00:36:43.792 00:36:43.792 Number of Queues 00:36:43.792 ================ 00:36:43.792 Number of I/O Submission Queues: 128 00:36:43.792 Number of I/O Completion Queues: 128 00:36:43.792 00:36:43.792 ZNS Specific Controller Data 00:36:43.792 ============================ 00:36:43.792 Zone Append Size Limit: 0 00:36:43.792 00:36:43.792 00:36:43.792 Active Namespaces 00:36:43.792 ================= 00:36:43.792 get_feature(0x05) failed 00:36:43.792 Namespace ID:1 00:36:43.792 Command Set Identifier: NVM (00h) 00:36:43.792 Deallocate: Supported 00:36:43.792 Deallocated/Unwritten Error: Not Supported 00:36:43.792 Deallocated Read Value: Unknown 00:36:43.792 Deallocate in Write Zeroes: Not Supported 00:36:43.792 Deallocated Guard Field: 0xFFFF 00:36:43.792 Flush: Supported 00:36:43.792 Reservation: Not Supported 00:36:43.792 Namespace Sharing Capabilities: Multiple Controllers 00:36:43.792 Size (in LBAs): 1953525168 (931GiB) 00:36:43.792 Capacity (in LBAs): 1953525168 (931GiB) 00:36:43.792 Utilization (in LBAs): 1953525168 (931GiB) 00:36:43.792 UUID: b0b19120-423b-4e7d-998f-b83e8fbc10be 00:36:43.792 Thin Provisioning: Not Supported 00:36:43.792 Per-NS Atomic Units: Yes 00:36:43.792 Atomic Boundary Size (Normal): 0 00:36:43.792 Atomic Boundary Size (PFail): 0 00:36:43.792 Atomic Boundary Offset: 0 00:36:43.792 NGUID/EUI64 Never Reused: No 00:36:43.792 ANA group ID: 1 00:36:43.792 Namespace Write Protected: No 00:36:43.792 Number of LBA Formats: 1 00:36:43.792 Current LBA Format: LBA Format #00 00:36:43.792 LBA Format #00: Data Size: 512 Metadata Size: 0 00:36:43.792 00:36:43.792 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@1 -- # nvmftestfini 00:36:43.792 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@516 -- # nvmfcleanup 00:36:43.792 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@121 -- # sync 00:36:43.792 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:36:43.792 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@124 -- # set +e 00:36:43.792 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@125 -- # for i in {1..20} 00:36:43.792 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:36:43.792 rmmod nvme_tcp 00:36:43.792 rmmod nvme_fabrics 00:36:43.792 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:36:43.792 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@128 -- # set -e 00:36:43.792 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@129 -- # return 0 00:36:43.792 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@517 -- # '[' -n '' ']' 00:36:43.792 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:36:43.792 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:36:43.792 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:36:43.792 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@297 -- # iptr 00:36:43.792 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@791 -- # iptables-save 00:36:43.792 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:36:43.792 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@791 -- # iptables-restore 00:36:43.792 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:36:43.792 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@302 -- # remove_spdk_ns 00:36:43.792 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:36:43.792 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:36:43.792 17:52:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:36:46.330 17:52:27 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:36:46.330 17:52:27 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@1 -- # clean_kernel_target 00:36:46.330 17:52:27 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@712 -- # [[ -e /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn ]] 00:36:46.330 17:52:27 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@714 -- # echo 0 00:36:46.330 17:52:27 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@716 -- # rm -f /sys/kernel/config/nvmet/ports/1/subsystems/nqn.2016-06.io.spdk:testnqn 00:36:46.330 17:52:27 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@717 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:36:46.330 17:52:27 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@718 -- # rmdir /sys/kernel/config/nvmet/ports/1 00:36:46.330 17:52:27 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@719 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:36:46.330 17:52:27 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@721 -- # modules=(/sys/module/nvmet/holders/*) 00:36:46.330 17:52:27 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@723 -- # modprobe -r nvmet_tcp nvmet 00:36:46.330 17:52:27 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@726 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:36:47.264 0000:00:04.7 (8086 0e27): ioatdma -> vfio-pci 00:36:47.264 0000:00:04.6 (8086 0e26): ioatdma -> vfio-pci 00:36:47.264 0000:00:04.5 (8086 0e25): ioatdma -> vfio-pci 00:36:47.264 0000:00:04.4 (8086 0e24): ioatdma -> vfio-pci 00:36:47.264 0000:00:04.3 (8086 0e23): ioatdma -> vfio-pci 00:36:47.264 0000:00:04.2 (8086 0e22): ioatdma -> vfio-pci 00:36:47.264 0000:00:04.1 (8086 0e21): ioatdma -> vfio-pci 00:36:47.264 0000:00:04.0 (8086 0e20): ioatdma -> vfio-pci 00:36:47.264 0000:80:04.7 (8086 0e27): ioatdma -> vfio-pci 00:36:47.264 0000:80:04.6 (8086 0e26): ioatdma -> vfio-pci 00:36:47.522 0000:80:04.5 (8086 0e25): ioatdma -> vfio-pci 00:36:47.522 0000:80:04.4 (8086 0e24): ioatdma -> vfio-pci 00:36:47.522 0000:80:04.3 (8086 0e23): ioatdma -> vfio-pci 00:36:47.522 0000:80:04.2 (8086 0e22): ioatdma -> vfio-pci 00:36:47.522 0000:80:04.1 (8086 0e21): ioatdma -> vfio-pci 00:36:47.522 0000:80:04.0 (8086 0e20): ioatdma -> vfio-pci 00:36:48.457 0000:88:00.0 (8086 0a54): nvme -> vfio-pci 00:36:48.457 00:36:48.457 real 0m10.027s 00:36:48.457 user 0m2.118s 00:36:48.457 sys 0m3.864s 00:36:48.457 17:52:30 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1130 -- # xtrace_disable 00:36:48.457 17:52:30 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@10 -- # set +x 00:36:48.457 ************************************ 00:36:48.457 END TEST nvmf_identify_kernel_target 00:36:48.457 ************************************ 00:36:48.457 17:52:30 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@30 -- # run_test nvmf_auth_host /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/auth.sh --transport=tcp 00:36:48.457 17:52:30 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:36:48.457 17:52:30 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:36:48.457 17:52:30 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:36:48.457 ************************************ 00:36:48.457 START TEST nvmf_auth_host 00:36:48.457 ************************************ 00:36:48.457 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/auth.sh --transport=tcp 00:36:48.715 * Looking for test storage... 00:36:48.715 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1711 -- # lcov --version 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@333 -- # local ver1 ver1_l 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@334 -- # local ver2 ver2_l 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@336 -- # IFS=.-: 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@336 -- # read -ra ver1 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@337 -- # IFS=.-: 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@337 -- # read -ra ver2 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@338 -- # local 'op=<' 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@340 -- # ver1_l=2 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@341 -- # ver2_l=1 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@344 -- # case "$op" in 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@345 -- # : 1 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@364 -- # (( v = 0 )) 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@365 -- # decimal 1 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@353 -- # local d=1 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@355 -- # echo 1 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@365 -- # ver1[v]=1 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@366 -- # decimal 2 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@353 -- # local d=2 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@355 -- # echo 2 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@366 -- # ver2[v]=2 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@368 -- # return 0 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:36:48.715 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:48.715 --rc genhtml_branch_coverage=1 00:36:48.715 --rc genhtml_function_coverage=1 00:36:48.715 --rc genhtml_legend=1 00:36:48.715 --rc geninfo_all_blocks=1 00:36:48.715 --rc geninfo_unexecuted_blocks=1 00:36:48.715 00:36:48.715 ' 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:36:48.715 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:48.715 --rc genhtml_branch_coverage=1 00:36:48.715 --rc genhtml_function_coverage=1 00:36:48.715 --rc genhtml_legend=1 00:36:48.715 --rc geninfo_all_blocks=1 00:36:48.715 --rc geninfo_unexecuted_blocks=1 00:36:48.715 00:36:48.715 ' 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:36:48.715 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:48.715 --rc genhtml_branch_coverage=1 00:36:48.715 --rc genhtml_function_coverage=1 00:36:48.715 --rc genhtml_legend=1 00:36:48.715 --rc geninfo_all_blocks=1 00:36:48.715 --rc geninfo_unexecuted_blocks=1 00:36:48.715 00:36:48.715 ' 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:36:48.715 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:48.715 --rc genhtml_branch_coverage=1 00:36:48.715 --rc genhtml_function_coverage=1 00:36:48.715 --rc genhtml_legend=1 00:36:48.715 --rc geninfo_all_blocks=1 00:36:48.715 --rc geninfo_unexecuted_blocks=1 00:36:48.715 00:36:48.715 ' 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@7 -- # uname -s 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@15 -- # shopt -s extglob 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:48.715 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- paths/export.sh@5 -- # export PATH 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@51 -- # : 0 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:36:48.716 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@55 -- # have_pci_nics=0 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@13 -- # digests=("sha256" "sha384" "sha512") 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@16 -- # dhgroups=("ffdhe2048" "ffdhe3072" "ffdhe4096" "ffdhe6144" "ffdhe8192") 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@17 -- # subnqn=nqn.2024-02.io.spdk:cnode0 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@18 -- # hostnqn=nqn.2024-02.io.spdk:host0 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@19 -- # nvmet_subsys=/sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@20 -- # nvmet_host=/sys/kernel/config/nvmet/hosts/nqn.2024-02.io.spdk:host0 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@21 -- # keys=() 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@21 -- # ckeys=() 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@68 -- # nvmftestinit 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@476 -- # prepare_net_devs 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@438 -- # local -g is_hw=no 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@440 -- # remove_spdk_ns 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@309 -- # xtrace_disable 00:36:48.716 17:52:30 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@315 -- # pci_devs=() 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@315 -- # local -a pci_devs 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@316 -- # pci_net_devs=() 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@317 -- # pci_drivers=() 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@317 -- # local -A pci_drivers 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@319 -- # net_devs=() 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@319 -- # local -ga net_devs 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@320 -- # e810=() 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@320 -- # local -ga e810 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@321 -- # x722=() 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@321 -- # local -ga x722 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@322 -- # mlx=() 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@322 -- # local -ga mlx 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:36:51.295 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:36:51.295 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:36:51.295 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@418 -- # [[ up == up ]] 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:36:51.296 Found net devices under 0000:0a:00.0: cvl_0_0 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@418 -- # [[ up == up ]] 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:36:51.296 Found net devices under 0000:0a:00.1: cvl_0_1 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@442 -- # is_hw=yes 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:36:51.296 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:36:51.296 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.233 ms 00:36:51.296 00:36:51.296 --- 10.0.0.2 ping statistics --- 00:36:51.296 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:36:51.296 rtt min/avg/max/mdev = 0.233/0.233/0.233/0.000 ms 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:36:51.296 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:36:51.296 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.043 ms 00:36:51.296 00:36:51.296 --- 10.0.0.1 ping statistics --- 00:36:51.296 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:36:51.296 rtt min/avg/max/mdev = 0.043/0.043/0.043/0.000 ms 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@450 -- # return 0 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@69 -- # nvmfappstart -L nvme_auth 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@726 -- # xtrace_disable 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@509 -- # nvmfpid=394480 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -L nvme_auth 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@510 -- # waitforlisten 394480 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@835 -- # '[' -z 394480 ']' 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@840 -- # local max_retries=100 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@844 -- # xtrace_disable 00:36:51.296 17:52:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:51.296 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:36:51.296 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@868 -- # return 0 00:36:51.296 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:36:51.296 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@732 -- # xtrace_disable 00:36:51.296 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:51.296 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:36:51.296 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@70 -- # trap 'cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvme-auth.log; cleanup' SIGINT SIGTERM EXIT 00:36:51.296 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@73 -- # gen_dhchap_key null 32 00:36:51.296 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@751 -- # local digest len file key 00:36:51.296 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:36:51.296 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # local -A digests 00:36:51.296 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # digest=null 00:36:51.296 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # len=32 00:36:51.296 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # xxd -p -c0 -l 16 /dev/urandom 00:36:51.296 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # key=eeba7df2f4b10e9f156c2635059dc4e2 00:36:51.296 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # mktemp -t spdk.key-null.XXX 00:36:51.296 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-null.0TL 00:36:51.296 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@757 -- # format_dhchap_key eeba7df2f4b10e9f156c2635059dc4e2 0 00:36:51.296 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@747 -- # format_key DHHC-1 eeba7df2f4b10e9f156c2635059dc4e2 0 00:36:51.296 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # local prefix key digest 00:36:51.296 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:36:51.296 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # key=eeba7df2f4b10e9f156c2635059dc4e2 00:36:51.296 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # digest=0 00:36:51.296 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@733 -- # python - 00:36:51.551 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-null.0TL 00:36:51.551 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-null.0TL 00:36:51.551 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@73 -- # keys[0]=/tmp/spdk.key-null.0TL 00:36:51.551 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@73 -- # gen_dhchap_key sha512 64 00:36:51.551 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@751 -- # local digest len file key 00:36:51.551 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:36:51.551 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # local -A digests 00:36:51.551 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # digest=sha512 00:36:51.551 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # len=64 00:36:51.551 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # xxd -p -c0 -l 32 /dev/urandom 00:36:51.551 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # key=09bbc045524441c0bd642e666df112581c05f2a994977a372de5b8ace45d6ab3 00:36:51.551 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # mktemp -t spdk.key-sha512.XXX 00:36:51.551 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-sha512.wbt 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@757 -- # format_dhchap_key 09bbc045524441c0bd642e666df112581c05f2a994977a372de5b8ace45d6ab3 3 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@747 -- # format_key DHHC-1 09bbc045524441c0bd642e666df112581c05f2a994977a372de5b8ace45d6ab3 3 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # local prefix key digest 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # key=09bbc045524441c0bd642e666df112581c05f2a994977a372de5b8ace45d6ab3 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # digest=3 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@733 -- # python - 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-sha512.wbt 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-sha512.wbt 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@73 -- # ckeys[0]=/tmp/spdk.key-sha512.wbt 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@74 -- # gen_dhchap_key null 48 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@751 -- # local digest len file key 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # local -A digests 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # digest=null 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # len=48 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # xxd -p -c0 -l 24 /dev/urandom 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # key=61a6a6fa45b559825c4e35919d211ca8b40295973b243dd4 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # mktemp -t spdk.key-null.XXX 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-null.Tjn 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@757 -- # format_dhchap_key 61a6a6fa45b559825c4e35919d211ca8b40295973b243dd4 0 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@747 -- # format_key DHHC-1 61a6a6fa45b559825c4e35919d211ca8b40295973b243dd4 0 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # local prefix key digest 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # key=61a6a6fa45b559825c4e35919d211ca8b40295973b243dd4 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # digest=0 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@733 -- # python - 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-null.Tjn 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-null.Tjn 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@74 -- # keys[1]=/tmp/spdk.key-null.Tjn 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@74 -- # gen_dhchap_key sha384 48 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@751 -- # local digest len file key 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # local -A digests 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # digest=sha384 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # len=48 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # xxd -p -c0 -l 24 /dev/urandom 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # key=ffafd86b089a1d9688428d9c647cdaef5074f2a4a4c69fc7 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # mktemp -t spdk.key-sha384.XXX 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-sha384.W34 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@757 -- # format_dhchap_key ffafd86b089a1d9688428d9c647cdaef5074f2a4a4c69fc7 2 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@747 -- # format_key DHHC-1 ffafd86b089a1d9688428d9c647cdaef5074f2a4a4c69fc7 2 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # local prefix key digest 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # key=ffafd86b089a1d9688428d9c647cdaef5074f2a4a4c69fc7 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # digest=2 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@733 -- # python - 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-sha384.W34 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-sha384.W34 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@74 -- # ckeys[1]=/tmp/spdk.key-sha384.W34 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@75 -- # gen_dhchap_key sha256 32 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@751 -- # local digest len file key 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # local -A digests 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # digest=sha256 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # len=32 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # xxd -p -c0 -l 16 /dev/urandom 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # key=5c31f3c03ecd1bcd3051f9358ff44e98 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # mktemp -t spdk.key-sha256.XXX 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-sha256.1bM 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@757 -- # format_dhchap_key 5c31f3c03ecd1bcd3051f9358ff44e98 1 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@747 -- # format_key DHHC-1 5c31f3c03ecd1bcd3051f9358ff44e98 1 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # local prefix key digest 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # key=5c31f3c03ecd1bcd3051f9358ff44e98 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # digest=1 00:36:51.552 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@733 -- # python - 00:36:51.808 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-sha256.1bM 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-sha256.1bM 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@75 -- # keys[2]=/tmp/spdk.key-sha256.1bM 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@75 -- # gen_dhchap_key sha256 32 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@751 -- # local digest len file key 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # local -A digests 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # digest=sha256 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # len=32 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # xxd -p -c0 -l 16 /dev/urandom 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # key=d9bfcc25e4eea92b92182aa103aefd04 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # mktemp -t spdk.key-sha256.XXX 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-sha256.PsQ 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@757 -- # format_dhchap_key d9bfcc25e4eea92b92182aa103aefd04 1 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@747 -- # format_key DHHC-1 d9bfcc25e4eea92b92182aa103aefd04 1 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # local prefix key digest 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # key=d9bfcc25e4eea92b92182aa103aefd04 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # digest=1 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@733 -- # python - 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-sha256.PsQ 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-sha256.PsQ 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@75 -- # ckeys[2]=/tmp/spdk.key-sha256.PsQ 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@76 -- # gen_dhchap_key sha384 48 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@751 -- # local digest len file key 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # local -A digests 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # digest=sha384 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # len=48 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # xxd -p -c0 -l 24 /dev/urandom 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # key=d0e483cb615472890accbd9e981d01b79ce9746f1571681b 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # mktemp -t spdk.key-sha384.XXX 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-sha384.5qL 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@757 -- # format_dhchap_key d0e483cb615472890accbd9e981d01b79ce9746f1571681b 2 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@747 -- # format_key DHHC-1 d0e483cb615472890accbd9e981d01b79ce9746f1571681b 2 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # local prefix key digest 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # key=d0e483cb615472890accbd9e981d01b79ce9746f1571681b 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # digest=2 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@733 -- # python - 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-sha384.5qL 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-sha384.5qL 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@76 -- # keys[3]=/tmp/spdk.key-sha384.5qL 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@76 -- # gen_dhchap_key null 32 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@751 -- # local digest len file key 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # local -A digests 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # digest=null 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # len=32 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # xxd -p -c0 -l 16 /dev/urandom 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # key=2b1e7ae20c8244c1f061de648eb29811 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # mktemp -t spdk.key-null.XXX 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-null.MNx 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@757 -- # format_dhchap_key 2b1e7ae20c8244c1f061de648eb29811 0 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@747 -- # format_key DHHC-1 2b1e7ae20c8244c1f061de648eb29811 0 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # local prefix key digest 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # key=2b1e7ae20c8244c1f061de648eb29811 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # digest=0 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@733 -- # python - 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-null.MNx 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-null.MNx 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@76 -- # ckeys[3]=/tmp/spdk.key-null.MNx 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@77 -- # gen_dhchap_key sha512 64 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@751 -- # local digest len file key 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # local -A digests 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # digest=sha512 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # len=64 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # xxd -p -c0 -l 32 /dev/urandom 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # key=7df9bbc862b7ca1f5337240355320a375593106b22e7cec8d9d6a164700ccd50 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # mktemp -t spdk.key-sha512.XXX 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # file=/tmp/spdk.key-sha512.Wlh 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@757 -- # format_dhchap_key 7df9bbc862b7ca1f5337240355320a375593106b22e7cec8d9d6a164700ccd50 3 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@747 -- # format_key DHHC-1 7df9bbc862b7ca1f5337240355320a375593106b22e7cec8d9d6a164700ccd50 3 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # local prefix key digest 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # prefix=DHHC-1 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # key=7df9bbc862b7ca1f5337240355320a375593106b22e7cec8d9d6a164700ccd50 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@732 -- # digest=3 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@733 -- # python - 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # chmod 0600 /tmp/spdk.key-sha512.Wlh 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@760 -- # echo /tmp/spdk.key-sha512.Wlh 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@77 -- # keys[4]=/tmp/spdk.key-sha512.Wlh 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@77 -- # ckeys[4]= 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@79 -- # waitforlisten 394480 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@835 -- # '[' -z 394480 ']' 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@840 -- # local max_retries=100 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:36:51.809 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@844 -- # xtrace_disable 00:36:51.809 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:52.373 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:36:52.373 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@868 -- # return 0 00:36:52.373 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@80 -- # for i in "${!keys[@]}" 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@81 -- # rpc_cmd keyring_file_add_key key0 /tmp/spdk.key-null.0TL 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # [[ -n /tmp/spdk.key-sha512.wbt ]] 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # rpc_cmd keyring_file_add_key ckey0 /tmp/spdk.key-sha512.wbt 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@80 -- # for i in "${!keys[@]}" 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@81 -- # rpc_cmd keyring_file_add_key key1 /tmp/spdk.key-null.Tjn 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # [[ -n /tmp/spdk.key-sha384.W34 ]] 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # rpc_cmd keyring_file_add_key ckey1 /tmp/spdk.key-sha384.W34 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@80 -- # for i in "${!keys[@]}" 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@81 -- # rpc_cmd keyring_file_add_key key2 /tmp/spdk.key-sha256.1bM 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # [[ -n /tmp/spdk.key-sha256.PsQ ]] 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # rpc_cmd keyring_file_add_key ckey2 /tmp/spdk.key-sha256.PsQ 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@80 -- # for i in "${!keys[@]}" 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@81 -- # rpc_cmd keyring_file_add_key key3 /tmp/spdk.key-sha384.5qL 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # [[ -n /tmp/spdk.key-null.MNx ]] 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # rpc_cmd keyring_file_add_key ckey3 /tmp/spdk.key-null.MNx 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@80 -- # for i in "${!keys[@]}" 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@81 -- # rpc_cmd keyring_file_add_key key4 /tmp/spdk.key-sha512.Wlh 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # [[ -n '' ]] 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@85 -- # nvmet_auth_init 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@35 -- # get_main_ns_ip 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:36:52.374 17:52:33 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:36:52.374 17:52:34 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:36:52.374 17:52:34 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:36:52.374 17:52:34 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:36:52.374 17:52:34 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:36:52.374 17:52:34 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@35 -- # configure_kernel_target nqn.2024-02.io.spdk:cnode0 10.0.0.1 00:36:52.374 17:52:34 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@660 -- # local kernel_name=nqn.2024-02.io.spdk:cnode0 kernel_target_ip=10.0.0.1 00:36:52.374 17:52:34 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@662 -- # nvmet=/sys/kernel/config/nvmet 00:36:52.374 17:52:34 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@663 -- # kernel_subsystem=/sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 00:36:52.374 17:52:34 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@664 -- # kernel_namespace=/sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/namespaces/1 00:36:52.374 17:52:34 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@665 -- # kernel_port=/sys/kernel/config/nvmet/ports/1 00:36:52.374 17:52:34 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@667 -- # local block nvme 00:36:52.374 17:52:34 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@669 -- # [[ ! -e /sys/module/nvmet ]] 00:36:52.374 17:52:34 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@670 -- # modprobe nvmet 00:36:52.374 17:52:34 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@673 -- # [[ -e /sys/kernel/config/nvmet ]] 00:36:52.374 17:52:34 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@675 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:36:53.305 Waiting for block devices as requested 00:36:53.562 0000:88:00.0 (8086 0a54): vfio-pci -> nvme 00:36:53.562 0000:00:04.7 (8086 0e27): vfio-pci -> ioatdma 00:36:53.562 0000:00:04.6 (8086 0e26): vfio-pci -> ioatdma 00:36:53.819 0000:00:04.5 (8086 0e25): vfio-pci -> ioatdma 00:36:53.819 0000:00:04.4 (8086 0e24): vfio-pci -> ioatdma 00:36:53.819 0000:00:04.3 (8086 0e23): vfio-pci -> ioatdma 00:36:54.076 0000:00:04.2 (8086 0e22): vfio-pci -> ioatdma 00:36:54.076 0000:00:04.1 (8086 0e21): vfio-pci -> ioatdma 00:36:54.076 0000:00:04.0 (8086 0e20): vfio-pci -> ioatdma 00:36:54.076 0000:80:04.7 (8086 0e27): vfio-pci -> ioatdma 00:36:54.076 0000:80:04.6 (8086 0e26): vfio-pci -> ioatdma 00:36:54.332 0000:80:04.5 (8086 0e25): vfio-pci -> ioatdma 00:36:54.332 0000:80:04.4 (8086 0e24): vfio-pci -> ioatdma 00:36:54.332 0000:80:04.3 (8086 0e23): vfio-pci -> ioatdma 00:36:54.332 0000:80:04.2 (8086 0e22): vfio-pci -> ioatdma 00:36:54.589 0000:80:04.1 (8086 0e21): vfio-pci -> ioatdma 00:36:54.589 0000:80:04.0 (8086 0e20): vfio-pci -> ioatdma 00:36:55.154 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@678 -- # for block in /sys/block/nvme* 00:36:55.154 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@679 -- # [[ -e /sys/block/nvme0n1 ]] 00:36:55.154 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@680 -- # is_block_zoned nvme0n1 00:36:55.154 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1650 -- # local device=nvme0n1 00:36:55.154 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:36:55.154 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:36:55.154 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@681 -- # block_in_use nvme0n1 00:36:55.154 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@381 -- # local block=nvme0n1 pt 00:36:55.154 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@390 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdk-gpt.py nvme0n1 00:36:55.154 No valid GPT data, bailing 00:36:55.154 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:36:55.154 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@394 -- # pt= 00:36:55.154 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@395 -- # return 1 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@681 -- # nvme=/dev/nvme0n1 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@684 -- # [[ -b /dev/nvme0n1 ]] 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@686 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@687 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/namespaces/1 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@688 -- # mkdir /sys/kernel/config/nvmet/ports/1 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@693 -- # echo SPDK-nqn.2024-02.io.spdk:cnode0 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@695 -- # echo 1 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@696 -- # echo /dev/nvme0n1 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@697 -- # echo 1 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@699 -- # echo 10.0.0.1 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@700 -- # echo tcp 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@701 -- # echo 4420 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@702 -- # echo ipv4 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@705 -- # ln -s /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 /sys/kernel/config/nvmet/ports/1/subsystems/ 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@708 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -a 10.0.0.1 -t tcp -s 4420 00:36:55.155 00:36:55.155 Discovery Log Number of Records 2, Generation counter 2 00:36:55.155 =====Discovery Log Entry 0====== 00:36:55.155 trtype: tcp 00:36:55.155 adrfam: ipv4 00:36:55.155 subtype: current discovery subsystem 00:36:55.155 treq: not specified, sq flow control disable supported 00:36:55.155 portid: 1 00:36:55.155 trsvcid: 4420 00:36:55.155 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:36:55.155 traddr: 10.0.0.1 00:36:55.155 eflags: none 00:36:55.155 sectype: none 00:36:55.155 =====Discovery Log Entry 1====== 00:36:55.155 trtype: tcp 00:36:55.155 adrfam: ipv4 00:36:55.155 subtype: nvme subsystem 00:36:55.155 treq: not specified, sq flow control disable supported 00:36:55.155 portid: 1 00:36:55.155 trsvcid: 4420 00:36:55.155 subnqn: nqn.2024-02.io.spdk:cnode0 00:36:55.155 traddr: 10.0.0.1 00:36:55.155 eflags: none 00:36:55.155 sectype: none 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@36 -- # mkdir /sys/kernel/config/nvmet/hosts/nqn.2024-02.io.spdk:host0 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@37 -- # echo 0 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@38 -- # ln -s /sys/kernel/config/nvmet/hosts/nqn.2024-02.io.spdk:host0 /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/allowed_hosts/nqn.2024-02.io.spdk:host0 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@88 -- # nvmet_auth_set_key sha256 ffdhe2048 1 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: ]] 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@93 -- # IFS=, 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@94 -- # printf %s sha256,sha384,sha512 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@93 -- # IFS=, 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@94 -- # printf %s ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@93 -- # connect_authenticate sha256,sha384,sha512 ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 1 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256,sha384,sha512 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256,sha384,sha512 --dhchap-dhgroups ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:55.155 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:55.438 17:52:36 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:55.438 nvme0n1 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@100 -- # for digest in "${digests[@]}" 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe2048 0 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:36:55.438 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: ]] 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe2048 0 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:55.439 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:55.698 nvme0n1 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe2048 1 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: ]] 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe2048 1 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:55.698 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:55.956 nvme0n1 00:36:55.956 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:55.956 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:36:55.956 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:55.956 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:36:55.956 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:55.956 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:55.956 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:36:55.956 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:36:55.956 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:55.956 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:55.956 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:55.956 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:36:55.956 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe2048 2 00:36:55.956 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:36:55.956 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:36:55.956 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:36:55.956 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:36:55.956 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:36:55.956 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:36:55.956 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:36:55.956 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:36:55.956 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:36:55.956 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: ]] 00:36:55.957 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:36:55.957 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe2048 2 00:36:55.957 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:36:55.957 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:36:55.957 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:36:55.957 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:36:55.957 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:36:55.957 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:36:55.957 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:55.957 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:55.957 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:55.957 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:36:55.957 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:36:55.957 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:36:55.957 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:36:55.957 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:36:55.957 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:36:55.957 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:36:55.957 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:36:55.957 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:36:55.957 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:36:55.957 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:36:55.957 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:36:55.957 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:55.957 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:56.215 nvme0n1 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe2048 3 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: ]] 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe2048 3 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:36:56.215 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:36:56.216 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:36:56.216 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:36:56.216 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:36:56.216 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:36:56.216 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:36:56.216 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:36:56.216 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:36:56.216 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:36:56.216 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:56.216 17:52:37 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:56.474 nvme0n1 00:36:56.474 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:56.474 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:36:56.474 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:56.474 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:36:56.474 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe2048 4 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe2048 4 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:56.475 nvme0n1 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:56.475 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:56.733 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:36:56.733 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:36:56.733 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:56.733 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:56.733 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:56.733 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:36:56.733 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:36:56.733 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe3072 0 00:36:56.733 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:36:56.733 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:36:56.733 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:36:56.733 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:36:56.733 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:36:56.733 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:36:56.733 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:36:56.733 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:36:56.991 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:36:56.991 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: ]] 00:36:56.991 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:36:56.991 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe3072 0 00:36:56.991 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:36:56.991 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:36:56.991 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:36:56.991 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:36:56.991 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:36:56.991 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:36:56.991 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:56.991 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:56.991 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:56.991 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:36:56.991 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:36:56.991 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:36:56.991 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:36:56.992 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:36:56.992 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:36:56.992 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:36:56.992 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:36:56.992 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:36:56.992 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:36:56.992 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:36:56.992 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:36:56.992 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:56.992 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:56.992 nvme0n1 00:36:56.992 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:56.992 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:36:56.992 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:56.992 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:56.992 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:36:56.992 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe3072 1 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: ]] 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe3072 1 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:57.250 17:52:38 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:57.250 nvme0n1 00:36:57.250 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:57.250 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:36:57.250 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:57.250 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:57.250 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:36:57.250 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe3072 2 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: ]] 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe3072 2 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:57.509 nvme0n1 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:36:57.509 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe3072 3 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: ]] 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe3072 3 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:57.767 nvme0n1 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:57.767 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:36:57.768 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:57.768 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:57.768 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:36:57.768 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:36:57.768 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:57.768 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe3072 4 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe3072 4 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:58.027 nvme0n1 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:36:58.027 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe4096 0 00:36:58.028 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:36:58.028 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:36:58.028 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:36:58.028 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:36:58.028 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:36:58.028 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:36:58.028 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:36:58.028 17:52:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:36:58.594 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:36:58.594 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: ]] 00:36:58.594 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:36:58.594 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe4096 0 00:36:58.594 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:36:58.594 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:36:58.594 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:36:58.594 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:36:58.594 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:36:58.594 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:36:58.594 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:58.594 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:58.594 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:58.594 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:36:58.594 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:36:58.594 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:36:58.594 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:36:58.594 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:36:58.594 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:36:58.594 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:36:58.594 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:36:58.594 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:36:58.594 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:36:58.594 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:36:58.594 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:36:58.594 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:58.594 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:59.161 nvme0n1 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe4096 1 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: ]] 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe4096 1 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:59.161 17:52:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:59.419 nvme0n1 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe4096 2 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: ]] 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe4096 2 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:59.419 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:59.677 nvme0n1 00:36:59.677 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:59.677 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:36:59.677 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:59.677 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:59.677 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:36:59.677 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:59.677 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:36:59.677 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:36:59.677 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:59.677 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:59.677 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:59.677 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:36:59.677 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe4096 3 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: ]] 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe4096 3 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:59.678 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:59.937 nvme0n1 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe4096 4 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe4096 4 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:36:59.937 17:52:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:00.196 nvme0n1 00:37:00.196 17:52:42 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:00.196 17:52:42 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:00.196 17:52:42 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:00.196 17:52:42 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:00.196 17:52:42 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:00.455 17:52:42 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:00.455 17:52:42 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:00.455 17:52:42 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:00.455 17:52:42 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:00.455 17:52:42 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:00.455 17:52:42 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:00.455 17:52:42 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:37:00.455 17:52:42 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:00.455 17:52:42 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe6144 0 00:37:00.455 17:52:42 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:00.455 17:52:42 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:37:00.455 17:52:42 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:37:00.455 17:52:42 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:37:00.455 17:52:42 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:37:00.455 17:52:42 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:37:00.455 17:52:42 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:37:00.455 17:52:42 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:37:02.356 17:52:43 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:37:02.356 17:52:43 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: ]] 00:37:02.356 17:52:43 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:37:02.356 17:52:43 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe6144 0 00:37:02.356 17:52:43 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:02.356 17:52:43 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:37:02.356 17:52:43 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:37:02.356 17:52:43 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:37:02.356 17:52:43 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:02.356 17:52:43 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:37:02.356 17:52:43 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:02.356 17:52:43 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:02.356 17:52:43 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:02.356 17:52:43 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:02.356 17:52:43 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:02.356 17:52:43 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:02.356 17:52:43 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:02.356 17:52:43 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:02.356 17:52:43 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:02.356 17:52:43 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:02.356 17:52:43 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:02.356 17:52:43 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:02.356 17:52:43 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:02.356 17:52:43 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:02.356 17:52:43 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:37:02.356 17:52:43 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:02.356 17:52:43 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:02.614 nvme0n1 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe6144 1 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: ]] 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe6144 1 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:02.615 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:03.183 nvme0n1 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe6144 2 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: ]] 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe6144 2 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:03.183 17:52:44 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:03.746 nvme0n1 00:37:03.746 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:03.746 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:03.746 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:03.746 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:03.746 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:03.746 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:03.746 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:03.746 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:03.746 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe6144 3 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: ]] 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe6144 3 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:03.747 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:04.312 nvme0n1 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe6144 4 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe6144 4 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:04.312 17:52:45 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:04.876 nvme0n1 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe8192 0 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: ]] 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe8192 0 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:04.876 17:52:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:05.808 nvme0n1 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe8192 1 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: ]] 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe8192 1 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:37:05.808 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:05.809 17:52:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:06.743 nvme0n1 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe8192 2 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: ]] 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe8192 2 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:06.743 17:52:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:07.673 nvme0n1 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe8192 3 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: ]] 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe8192 3 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:07.673 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:07.674 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:37:07.674 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:07.674 17:52:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:08.604 nvme0n1 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe8192 4 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe8192 4 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:08.604 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:08.605 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:08.605 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:08.605 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:08.605 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:08.605 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:08.605 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:08.605 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:08.605 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:08.605 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:08.605 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:08.605 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:08.605 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:08.605 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:37:08.605 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:08.605 17:52:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:09.537 nvme0n1 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@100 -- # for digest in "${digests[@]}" 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe2048 0 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: ]] 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe2048 0 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:09.537 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:37:09.538 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:09.538 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:09.538 nvme0n1 00:37:09.538 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:09.538 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:09.538 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:09.538 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:09.538 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:09.538 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe2048 1 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: ]] 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe2048 1 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:09.795 nvme0n1 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:09.795 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe2048 2 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: ]] 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe2048 2 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:10.053 nvme0n1 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe2048 3 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: ]] 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe2048 3 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:10.053 17:52:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:10.311 nvme0n1 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe2048 4 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe2048 4 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:10.311 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:10.312 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:10.312 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:10.312 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:10.312 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:10.312 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:10.312 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:10.312 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:10.312 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:10.312 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:10.312 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:37:10.312 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:10.312 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:10.569 nvme0n1 00:37:10.569 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:10.569 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:10.569 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:10.569 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe3072 0 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: ]] 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe3072 0 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:10.570 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:10.828 nvme0n1 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe3072 1 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: ]] 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe3072 1 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:10.828 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:10.829 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:10.829 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:10.829 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:10.829 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:10.829 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:10.829 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:10.829 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:10.829 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:10.829 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:10.829 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:10.829 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:10.829 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:10.829 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:37:10.829 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:10.829 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:11.087 nvme0n1 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe3072 2 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: ]] 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe3072 2 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:11.087 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:11.088 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:11.088 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:11.088 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:11.088 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:11.088 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:11.088 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:11.088 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:11.088 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:11.088 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:37:11.088 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:11.088 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:11.346 nvme0n1 00:37:11.346 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:11.346 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:11.346 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:11.346 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:11.346 17:52:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:11.346 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:11.346 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:11.346 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:11.346 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:11.346 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:11.346 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:11.346 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:11.346 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe3072 3 00:37:11.346 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:11.346 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:37:11.346 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:37:11.346 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:37:11.346 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:37:11.346 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:37:11.346 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:37:11.346 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:37:11.346 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:37:11.346 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: ]] 00:37:11.346 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:37:11.346 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe3072 3 00:37:11.346 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:11.346 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:37:11.347 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:37:11.347 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:37:11.347 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:11.347 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:37:11.347 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:11.347 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:11.347 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:11.347 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:11.347 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:11.347 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:11.347 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:11.347 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:11.347 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:11.347 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:11.347 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:11.347 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:11.347 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:11.347 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:11.347 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:37:11.347 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:11.347 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:11.650 nvme0n1 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe3072 4 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe3072 4 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:11.650 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:12.007 nvme0n1 00:37:12.007 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe4096 0 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: ]] 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe4096 0 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:12.008 nvme0n1 00:37:12.008 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:12.296 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:12.296 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:12.296 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:12.296 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:12.296 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:12.296 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:12.296 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:12.296 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:12.296 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:12.296 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:12.296 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:12.296 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe4096 1 00:37:12.296 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:12.296 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:37:12.296 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:37:12.296 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:37:12.296 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:37:12.296 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:37:12.296 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:37:12.296 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:37:12.296 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:37:12.296 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: ]] 00:37:12.296 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:37:12.297 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe4096 1 00:37:12.297 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:12.297 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:37:12.297 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:37:12.297 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:37:12.297 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:12.297 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:37:12.297 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:12.297 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:12.297 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:12.297 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:12.297 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:12.297 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:12.297 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:12.297 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:12.297 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:12.297 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:12.297 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:12.297 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:12.297 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:12.297 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:12.297 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:37:12.297 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:12.297 17:52:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:12.594 nvme0n1 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe4096 2 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: ]] 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe4096 2 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:12.594 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:12.903 nvme0n1 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe4096 3 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: ]] 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe4096 3 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:12.903 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:13.170 nvme0n1 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe4096 4 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe4096 4 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:13.170 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:13.171 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:13.171 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:13.171 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:13.171 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:13.171 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:13.171 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:13.171 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:37:13.171 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:13.171 17:52:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:13.448 nvme0n1 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe6144 0 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: ]] 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe6144 0 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:13.448 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:14.025 nvme0n1 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe6144 1 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: ]] 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe6144 1 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:14.025 17:52:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:14.596 nvme0n1 00:37:14.596 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:14.596 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:14.596 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:14.596 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:14.596 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:14.596 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:14.596 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:14.596 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:14.596 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:14.596 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:14.596 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:14.596 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:14.596 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe6144 2 00:37:14.596 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:14.596 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:37:14.596 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:37:14.596 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:37:14.596 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:37:14.596 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:37:14.596 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:37:14.596 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:37:14.597 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:37:14.597 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: ]] 00:37:14.597 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:37:14.597 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe6144 2 00:37:14.597 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:14.597 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:37:14.597 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:37:14.597 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:37:14.597 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:14.597 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:37:14.597 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:14.597 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:14.597 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:14.597 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:14.597 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:14.597 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:14.597 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:14.597 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:14.597 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:14.597 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:14.597 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:14.597 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:14.597 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:14.597 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:14.597 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:37:14.597 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:14.597 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:15.163 nvme0n1 00:37:15.163 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:15.163 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:15.163 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:15.163 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:15.163 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:15.163 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:15.163 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:15.163 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:15.163 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:15.163 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:15.163 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:15.163 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:15.163 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe6144 3 00:37:15.163 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:15.163 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:37:15.163 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:37:15.163 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:37:15.163 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:37:15.163 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: ]] 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe6144 3 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:15.164 17:52:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:15.728 nvme0n1 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe6144 4 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe6144 4 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:37:15.728 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:15.729 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:15.729 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:15.729 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:15.729 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:15.729 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:15.729 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:15.729 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:15.729 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:15.729 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:15.729 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:15.729 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:15.729 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:15.729 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:15.729 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:37:15.729 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:15.729 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:16.294 nvme0n1 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe8192 0 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: ]] 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe8192 0 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:16.294 17:52:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:16.294 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:16.294 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:16.294 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:16.294 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:16.294 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:16.294 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:16.294 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:16.294 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:16.294 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:16.294 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:16.294 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:16.294 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:16.294 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:37:16.294 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:16.294 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:17.229 nvme0n1 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe8192 1 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: ]] 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe8192 1 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:17.229 17:52:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:18.164 nvme0n1 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe8192 2 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: ]] 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe8192 2 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:18.164 17:52:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:19.099 nvme0n1 00:37:19.099 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:19.099 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:19.099 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:19.099 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:19.099 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:19.099 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:19.099 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:19.099 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:19.099 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:19.099 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:19.099 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:19.099 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:19.099 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe8192 3 00:37:19.099 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: ]] 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe8192 3 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:19.100 17:53:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:20.035 nvme0n1 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe8192 4 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe8192 4 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:20.035 17:53:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:20.972 nvme0n1 00:37:20.972 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:20.972 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:20.972 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:20.972 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:20.972 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:20.972 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:20.972 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:20.972 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:20.972 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:20.972 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:20.972 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:20.972 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@100 -- # for digest in "${digests[@]}" 00:37:20.972 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:37:20.972 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:20.972 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe2048 0 00:37:20.972 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:20.972 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:37:20.972 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:37:20.972 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: ]] 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe2048 0 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:20.973 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:21.234 nvme0n1 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe2048 1 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: ]] 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe2048 1 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:21.234 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:21.235 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:21.235 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:21.235 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:21.235 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:37:21.235 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:21.235 17:53:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:21.235 nvme0n1 00:37:21.235 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:21.235 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:21.235 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:21.235 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:21.235 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:21.235 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:21.493 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:21.493 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:21.493 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:21.493 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:21.493 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:21.493 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:21.493 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe2048 2 00:37:21.493 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:21.493 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:37:21.493 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:37:21.493 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:37:21.493 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:37:21.493 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:37:21.493 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:37:21.493 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:37:21.493 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:37:21.493 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: ]] 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe2048 2 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:21.494 nvme0n1 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe2048 3 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:37:21.494 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: ]] 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe2048 3 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:21.753 nvme0n1 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe2048 4 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe2048 4 00:37:21.753 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:21.754 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:37:21.754 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:37:21.754 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:37:21.754 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:21.754 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:37:21.754 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:21.754 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:21.754 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:21.754 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:21.754 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:21.754 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:21.754 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:21.754 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:21.754 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:21.754 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:21.754 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:21.754 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:21.754 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:21.754 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:21.754 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:37:21.754 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:21.754 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:22.013 nvme0n1 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe3072 0 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: ]] 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe3072 0 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:22.013 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:37:22.014 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:22.014 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:22.014 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:22.014 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:22.014 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:22.014 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:22.014 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:22.014 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:22.014 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:22.014 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:22.014 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:22.014 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:22.014 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:22.014 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:22.014 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:37:22.014 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:22.014 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:22.272 nvme0n1 00:37:22.272 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:22.272 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:22.272 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:22.272 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:22.272 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:22.272 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:22.272 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:22.272 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:22.272 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:22.272 17:53:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:22.272 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:22.272 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:22.272 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe3072 1 00:37:22.272 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:22.272 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:37:22.272 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:37:22.272 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:37:22.272 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:37:22.272 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:37:22.272 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:37:22.272 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:37:22.272 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:37:22.272 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: ]] 00:37:22.272 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:37:22.272 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe3072 1 00:37:22.272 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:22.272 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:37:22.272 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:37:22.272 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:37:22.272 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:22.272 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:37:22.272 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:22.272 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:22.272 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:22.272 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:22.272 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:22.272 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:22.272 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:22.273 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:22.273 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:22.273 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:22.273 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:22.273 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:22.273 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:22.273 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:22.273 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:37:22.273 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:22.273 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:22.531 nvme0n1 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe3072 2 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: ]] 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe3072 2 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:22.531 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:22.532 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:22.532 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:22.532 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:22.532 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:37:22.532 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:22.532 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:22.791 nvme0n1 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe3072 3 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: ]] 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe3072 3 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:22.791 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:23.049 nvme0n1 00:37:23.049 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:23.049 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:23.049 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:23.049 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:23.049 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:23.049 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:23.049 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:23.049 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:23.049 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:23.049 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:23.049 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:23.049 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:23.049 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe3072 4 00:37:23.049 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:23.049 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:37:23.049 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:37:23.049 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:37:23.049 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:37:23.049 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:37:23.049 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:37:23.049 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:37:23.050 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:37:23.050 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:37:23.050 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe3072 4 00:37:23.050 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:23.050 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:37:23.050 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:37:23.050 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:37:23.050 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:23.050 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:37:23.050 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:23.050 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:23.050 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:23.050 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:23.050 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:23.050 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:23.050 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:23.050 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:23.050 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:23.050 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:23.050 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:23.050 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:23.050 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:23.050 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:23.050 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:37:23.050 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:23.050 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:23.308 nvme0n1 00:37:23.308 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:23.308 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:23.308 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:23.308 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:23.308 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:23.308 17:53:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe4096 0 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: ]] 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe4096 0 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:23.308 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:23.309 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:23.309 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:37:23.309 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:23.309 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:23.566 nvme0n1 00:37:23.566 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:23.566 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:23.566 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:23.566 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:23.566 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:23.566 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:23.566 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:23.566 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:23.566 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:23.566 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:23.566 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:23.566 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:23.566 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe4096 1 00:37:23.566 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:23.566 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: ]] 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe4096 1 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:23.567 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:23.825 nvme0n1 00:37:23.825 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:23.825 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:23.825 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:23.825 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:23.825 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe4096 2 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: ]] 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe4096 2 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:24.083 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:24.342 nvme0n1 00:37:24.342 17:53:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe4096 3 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: ]] 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe4096 3 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:24.342 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:24.601 nvme0n1 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe4096 4 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe4096 4 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:24.601 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:24.860 nvme0n1 00:37:24.860 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:24.860 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:24.860 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:24.860 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:24.860 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:25.118 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:25.118 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:25.118 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:25.118 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:25.118 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:25.118 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:25.118 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:37:25.118 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:25.118 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe6144 0 00:37:25.118 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:25.118 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:37:25.118 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:37:25.118 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:37:25.118 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:37:25.118 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:37:25.118 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:37:25.118 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:37:25.118 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:37:25.118 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: ]] 00:37:25.119 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:37:25.119 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe6144 0 00:37:25.119 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:25.119 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:37:25.119 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:37:25.119 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:37:25.119 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:25.119 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:37:25.119 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:25.119 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:25.119 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:25.119 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:25.119 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:25.119 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:25.119 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:25.119 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:25.119 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:25.119 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:25.119 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:25.119 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:25.119 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:25.119 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:25.119 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:37:25.119 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:25.119 17:53:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:25.683 nvme0n1 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe6144 1 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: ]] 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe6144 1 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:25.683 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:26.248 nvme0n1 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe6144 2 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: ]] 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe6144 2 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:26.248 17:53:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:26.811 nvme0n1 00:37:26.811 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:26.811 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:26.811 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe6144 3 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: ]] 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe6144 3 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:26.812 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:27.374 nvme0n1 00:37:27.374 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:27.374 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:27.374 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:27.374 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:27.374 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:27.374 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:27.375 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:27.375 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:27.375 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:27.375 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:27.375 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:27.375 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:27.375 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe6144 4 00:37:27.375 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:27.375 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:37:27.375 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:37:27.375 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:37:27.375 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:37:27.375 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:37:27.375 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:37:27.375 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:37:27.375 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:37:27.375 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:37:27.375 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe6144 4 00:37:27.375 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:27.375 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:37:27.375 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:37:27.375 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:37:27.375 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:27.375 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:37:27.375 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:27.375 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:27.375 17:53:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:27.375 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:27.375 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:27.375 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:27.375 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:27.375 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:27.375 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:27.375 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:27.375 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:27.375 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:27.375 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:27.375 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:27.375 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:37:27.375 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:27.375 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:27.940 nvme0n1 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe8192 0 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZWViYTdkZjJmNGIxMGU5ZjE1NmMyNjM1MDU5ZGM0ZTJ7gSwy: 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: ]] 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:MDliYmMwNDU1MjQ0NDFjMGJkNjQyZTY2NmRmMTEyNTgxYzA1ZjJhOTk0OTc3YTM3MmRlNWI4YWNlNDVkNmFiM7AirjE=: 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe8192 0 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:27.940 17:53:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:28.873 nvme0n1 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe8192 1 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: ]] 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe8192 1 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:28.873 17:53:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:29.807 nvme0n1 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe8192 2 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: ]] 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe8192 2 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:29.807 17:53:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:30.742 nvme0n1 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe8192 3 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDBlNDgzY2I2MTU0NzI4OTBhY2NiZDllOTgxZDAxYjc5Y2U5NzQ2ZjE1NzE2ODFiiXc/ew==: 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: ]] 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MmIxZTdhZTIwYzgyNDRjMWYwNjFkZTY0OGViMjk4MTFotfkU: 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe8192 3 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:30.742 17:53:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:31.675 nvme0n1 00:37:31.675 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:31.675 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:31.675 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:31.675 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:31.675 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:31.675 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:31.675 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:31.675 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:31.675 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:31.675 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:31.675 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:31.675 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:37:31.675 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe8192 4 00:37:31.675 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:31.675 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:N2RmOWJiYzg2MmI3Y2ExZjUzMzcyNDAzNTUzMjBhMzc1NTkzMTA2YjIyZTdjZWM4ZDlkNmExNjQ3MDBjY2Q1ML5aeAs=: 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe8192 4 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:31.676 17:53:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:32.240 nvme0n1 00:37:32.240 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:32.240 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:37:32.240 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:32.240 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:32.240 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:37:32.240 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe2048 1 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: ]] 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@111 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@112 -- # get_main_ns_ip 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@112 -- # NOT rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@652 -- # local es=0 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@655 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:32.498 request: 00:37:32.498 { 00:37:32.498 "name": "nvme0", 00:37:32.498 "trtype": "tcp", 00:37:32.498 "traddr": "10.0.0.1", 00:37:32.498 "adrfam": "ipv4", 00:37:32.498 "trsvcid": "4420", 00:37:32.498 "subnqn": "nqn.2024-02.io.spdk:cnode0", 00:37:32.498 "hostnqn": "nqn.2024-02.io.spdk:host0", 00:37:32.498 "prchk_reftag": false, 00:37:32.498 "prchk_guard": false, 00:37:32.498 "hdgst": false, 00:37:32.498 "ddgst": false, 00:37:32.498 "allow_unrecognized_csi": false, 00:37:32.498 "method": "bdev_nvme_attach_controller", 00:37:32.498 "req_id": 1 00:37:32.498 } 00:37:32.498 Got JSON-RPC error response 00:37:32.498 response: 00:37:32.498 { 00:37:32.498 "code": -5, 00:37:32.498 "message": "Input/output error" 00:37:32.498 } 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@655 -- # es=1 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@114 -- # rpc_cmd bdev_nvme_get_controllers 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@114 -- # jq length 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@114 -- # (( 0 == 0 )) 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@117 -- # get_main_ns_ip 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@117 -- # NOT rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@652 -- # local es=0 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:37:32.498 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:37:32.499 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:37:32.499 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:37:32.499 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@655 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:37:32.499 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:32.499 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:32.499 request: 00:37:32.499 { 00:37:32.499 "name": "nvme0", 00:37:32.499 "trtype": "tcp", 00:37:32.499 "traddr": "10.0.0.1", 00:37:32.499 "adrfam": "ipv4", 00:37:32.499 "trsvcid": "4420", 00:37:32.499 "subnqn": "nqn.2024-02.io.spdk:cnode0", 00:37:32.499 "hostnqn": "nqn.2024-02.io.spdk:host0", 00:37:32.499 "prchk_reftag": false, 00:37:32.499 "prchk_guard": false, 00:37:32.499 "hdgst": false, 00:37:32.499 "ddgst": false, 00:37:32.499 "dhchap_key": "key2", 00:37:32.499 "allow_unrecognized_csi": false, 00:37:32.499 "method": "bdev_nvme_attach_controller", 00:37:32.499 "req_id": 1 00:37:32.499 } 00:37:32.499 Got JSON-RPC error response 00:37:32.499 response: 00:37:32.499 { 00:37:32.499 "code": -5, 00:37:32.499 "message": "Input/output error" 00:37:32.499 } 00:37:32.499 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:37:32.499 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@655 -- # es=1 00:37:32.499 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:37:32.499 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:37:32.499 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:37:32.499 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@120 -- # rpc_cmd bdev_nvme_get_controllers 00:37:32.499 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@120 -- # jq length 00:37:32.499 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:32.499 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:32.756 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:32.756 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@120 -- # (( 0 == 0 )) 00:37:32.756 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@123 -- # get_main_ns_ip 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@123 -- # NOT rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@652 -- # local es=0 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@655 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:32.757 request: 00:37:32.757 { 00:37:32.757 "name": "nvme0", 00:37:32.757 "trtype": "tcp", 00:37:32.757 "traddr": "10.0.0.1", 00:37:32.757 "adrfam": "ipv4", 00:37:32.757 "trsvcid": "4420", 00:37:32.757 "subnqn": "nqn.2024-02.io.spdk:cnode0", 00:37:32.757 "hostnqn": "nqn.2024-02.io.spdk:host0", 00:37:32.757 "prchk_reftag": false, 00:37:32.757 "prchk_guard": false, 00:37:32.757 "hdgst": false, 00:37:32.757 "ddgst": false, 00:37:32.757 "dhchap_key": "key1", 00:37:32.757 "dhchap_ctrlr_key": "ckey2", 00:37:32.757 "allow_unrecognized_csi": false, 00:37:32.757 "method": "bdev_nvme_attach_controller", 00:37:32.757 "req_id": 1 00:37:32.757 } 00:37:32.757 Got JSON-RPC error response 00:37:32.757 response: 00:37:32.757 { 00:37:32.757 "code": -5, 00:37:32.757 "message": "Input/output error" 00:37:32.757 } 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@655 -- # es=1 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@128 -- # get_main_ns_ip 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@128 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:32.757 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:33.015 nvme0n1 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@132 -- # nvmet_auth_set_key sha256 ffdhe2048 2 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: ]] 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@133 -- # rpc_cmd bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@134 -- # rpc_cmd bdev_nvme_get_controllers 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@134 -- # jq -r '.[].name' 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@134 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@136 -- # NOT rpc_cmd bdev_nvme_set_keys nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@652 -- # local es=0 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_nvme_set_keys nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@655 -- # rpc_cmd bdev_nvme_set_keys nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:33.015 request: 00:37:33.015 { 00:37:33.015 "name": "nvme0", 00:37:33.015 "dhchap_key": "key1", 00:37:33.015 "dhchap_ctrlr_key": "ckey2", 00:37:33.015 "method": "bdev_nvme_set_keys", 00:37:33.015 "req_id": 1 00:37:33.015 } 00:37:33.015 Got JSON-RPC error response 00:37:33.015 response: 00:37:33.015 { 00:37:33.015 "code": -13, 00:37:33.015 "message": "Permission denied" 00:37:33.015 } 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@655 -- # es=1 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # rpc_cmd bdev_nvme_get_controllers 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # jq length 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # (( 1 != 0 )) 00:37:33.015 17:53:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@138 -- # sleep 1s 00:37:34.387 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # rpc_cmd bdev_nvme_get_controllers 00:37:34.387 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # jq length 00:37:34.387 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:34.387 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:34.387 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:34.387 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # (( 0 != 0 )) 00:37:34.387 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@141 -- # nvmet_auth_set_key sha256 ffdhe2048 1 00:37:34.387 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:34.387 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:37:34.387 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:37:34.387 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:37:34.387 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:37:34.387 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:37:34.387 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:37:34.387 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:37:34.388 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:NjFhNmE2ZmE0NWI1NTk4MjVjNGUzNTkxOWQyMTFjYThiNDAyOTU5NzNiMjQzZGQ0QpHMcQ==: 00:37:34.388 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: ]] 00:37:34.388 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:ZmZhZmQ4NmIwODlhMWQ5Njg4NDI4ZDljNjQ3Y2RhZWY1MDc0ZjJhNGE0YzY5ZmM3Zk3CaQ==: 00:37:34.388 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@142 -- # get_main_ns_ip 00:37:34.388 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@769 -- # local ip 00:37:34.388 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates=() 00:37:34.388 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # local -A ip_candidates 00:37:34.388 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:37:34.388 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:37:34.388 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:37:34.388 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:37:34.388 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:37:34.388 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:37:34.388 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:37:34.388 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@142 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:37:34.388 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:34.388 17:53:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:34.388 nvme0n1 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@146 -- # nvmet_auth_set_key sha256 ffdhe2048 2 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NWMzMWYzYzAzZWNkMWJjZDMwNTFmOTM1OGZmNDRlOTiTRepu: 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: ]] 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:ZDliZmNjMjVlNGVlYTkyYjkyMTgyYWExMDNhZWZkMDRN+bK9: 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@147 -- # NOT rpc_cmd bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey1 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@652 -- # local es=0 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey1 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@655 -- # rpc_cmd bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey1 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:34.388 request: 00:37:34.388 { 00:37:34.388 "name": "nvme0", 00:37:34.388 "dhchap_key": "key2", 00:37:34.388 "dhchap_ctrlr_key": "ckey1", 00:37:34.388 "method": "bdev_nvme_set_keys", 00:37:34.388 "req_id": 1 00:37:34.388 } 00:37:34.388 Got JSON-RPC error response 00:37:34.388 response: 00:37:34.388 { 00:37:34.388 "code": -13, 00:37:34.388 "message": "Permission denied" 00:37:34.388 } 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@655 -- # es=1 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@148 -- # rpc_cmd bdev_nvme_get_controllers 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@148 -- # jq length 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@148 -- # (( 1 != 0 )) 00:37:34.388 17:53:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@149 -- # sleep 1s 00:37:35.322 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@148 -- # rpc_cmd bdev_nvme_get_controllers 00:37:35.322 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@148 -- # jq length 00:37:35.322 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:35.322 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:35.322 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:35.322 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@148 -- # (( 0 != 0 )) 00:37:35.322 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@152 -- # trap - SIGINT SIGTERM EXIT 00:37:35.322 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@153 -- # cleanup 00:37:35.322 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@24 -- # nvmftestfini 00:37:35.322 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@516 -- # nvmfcleanup 00:37:35.322 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@121 -- # sync 00:37:35.322 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:37:35.322 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@124 -- # set +e 00:37:35.322 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@125 -- # for i in {1..20} 00:37:35.322 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:37:35.322 rmmod nvme_tcp 00:37:35.581 rmmod nvme_fabrics 00:37:35.581 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:37:35.581 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@128 -- # set -e 00:37:35.581 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@129 -- # return 0 00:37:35.581 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@517 -- # '[' -n 394480 ']' 00:37:35.581 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@518 -- # killprocess 394480 00:37:35.581 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@954 -- # '[' -z 394480 ']' 00:37:35.581 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@958 -- # kill -0 394480 00:37:35.581 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@959 -- # uname 00:37:35.581 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:37:35.581 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 394480 00:37:35.581 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:37:35.581 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:37:35.581 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@972 -- # echo 'killing process with pid 394480' 00:37:35.581 killing process with pid 394480 00:37:35.581 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@973 -- # kill 394480 00:37:35.581 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@978 -- # wait 394480 00:37:35.581 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:37:35.581 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:37:35.581 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:37:35.581 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@297 -- # iptr 00:37:35.581 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@791 -- # iptables-save 00:37:35.581 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:37:35.581 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@791 -- # iptables-restore 00:37:35.581 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:37:35.581 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@302 -- # remove_spdk_ns 00:37:35.581 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:37:35.581 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:37:35.581 17:53:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:37:38.121 17:53:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:37:38.121 17:53:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@25 -- # rm /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/allowed_hosts/nqn.2024-02.io.spdk:host0 00:37:38.121 17:53:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@26 -- # rmdir /sys/kernel/config/nvmet/hosts/nqn.2024-02.io.spdk:host0 00:37:38.121 17:53:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@27 -- # clean_kernel_target 00:37:38.121 17:53:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@712 -- # [[ -e /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 ]] 00:37:38.121 17:53:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@714 -- # echo 0 00:37:38.121 17:53:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@716 -- # rm -f /sys/kernel/config/nvmet/ports/1/subsystems/nqn.2024-02.io.spdk:cnode0 00:37:38.121 17:53:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@717 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/namespaces/1 00:37:38.121 17:53:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@718 -- # rmdir /sys/kernel/config/nvmet/ports/1 00:37:38.121 17:53:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@719 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 00:37:38.121 17:53:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@721 -- # modules=(/sys/module/nvmet/holders/*) 00:37:38.121 17:53:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@723 -- # modprobe -r nvmet_tcp nvmet 00:37:38.121 17:53:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@726 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:37:39.059 0000:00:04.7 (8086 0e27): ioatdma -> vfio-pci 00:37:39.059 0000:00:04.6 (8086 0e26): ioatdma -> vfio-pci 00:37:39.059 0000:00:04.5 (8086 0e25): ioatdma -> vfio-pci 00:37:39.059 0000:00:04.4 (8086 0e24): ioatdma -> vfio-pci 00:37:39.059 0000:00:04.3 (8086 0e23): ioatdma -> vfio-pci 00:37:39.059 0000:00:04.2 (8086 0e22): ioatdma -> vfio-pci 00:37:39.059 0000:00:04.1 (8086 0e21): ioatdma -> vfio-pci 00:37:39.059 0000:00:04.0 (8086 0e20): ioatdma -> vfio-pci 00:37:39.059 0000:80:04.7 (8086 0e27): ioatdma -> vfio-pci 00:37:39.059 0000:80:04.6 (8086 0e26): ioatdma -> vfio-pci 00:37:39.059 0000:80:04.5 (8086 0e25): ioatdma -> vfio-pci 00:37:39.059 0000:80:04.4 (8086 0e24): ioatdma -> vfio-pci 00:37:39.059 0000:80:04.3 (8086 0e23): ioatdma -> vfio-pci 00:37:39.059 0000:80:04.2 (8086 0e22): ioatdma -> vfio-pci 00:37:39.059 0000:80:04.1 (8086 0e21): ioatdma -> vfio-pci 00:37:39.059 0000:80:04.0 (8086 0e20): ioatdma -> vfio-pci 00:37:39.996 0000:88:00.0 (8086 0a54): nvme -> vfio-pci 00:37:40.255 17:53:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@28 -- # rm -f /tmp/spdk.key-null.0TL /tmp/spdk.key-null.Tjn /tmp/spdk.key-sha256.1bM /tmp/spdk.key-sha384.5qL /tmp/spdk.key-sha512.Wlh /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvme-auth.log 00:37:40.256 17:53:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:37:41.633 0000:00:04.7 (8086 0e27): Already using the vfio-pci driver 00:37:41.633 0000:88:00.0 (8086 0a54): Already using the vfio-pci driver 00:37:41.633 0000:00:04.6 (8086 0e26): Already using the vfio-pci driver 00:37:41.633 0000:00:04.5 (8086 0e25): Already using the vfio-pci driver 00:37:41.633 0000:00:04.4 (8086 0e24): Already using the vfio-pci driver 00:37:41.633 0000:00:04.3 (8086 0e23): Already using the vfio-pci driver 00:37:41.633 0000:00:04.2 (8086 0e22): Already using the vfio-pci driver 00:37:41.633 0000:00:04.1 (8086 0e21): Already using the vfio-pci driver 00:37:41.633 0000:00:04.0 (8086 0e20): Already using the vfio-pci driver 00:37:41.633 0000:80:04.7 (8086 0e27): Already using the vfio-pci driver 00:37:41.633 0000:80:04.6 (8086 0e26): Already using the vfio-pci driver 00:37:41.633 0000:80:04.5 (8086 0e25): Already using the vfio-pci driver 00:37:41.633 0000:80:04.4 (8086 0e24): Already using the vfio-pci driver 00:37:41.633 0000:80:04.3 (8086 0e23): Already using the vfio-pci driver 00:37:41.633 0000:80:04.2 (8086 0e22): Already using the vfio-pci driver 00:37:41.633 0000:80:04.1 (8086 0e21): Already using the vfio-pci driver 00:37:41.633 0000:80:04.0 (8086 0e20): Already using the vfio-pci driver 00:37:41.633 00:37:41.633 real 0m53.032s 00:37:41.633 user 0m50.697s 00:37:41.633 sys 0m6.271s 00:37:41.633 17:53:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1130 -- # xtrace_disable 00:37:41.633 17:53:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:37:41.633 ************************************ 00:37:41.633 END TEST nvmf_auth_host 00:37:41.633 ************************************ 00:37:41.633 17:53:23 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@32 -- # [[ tcp == \t\c\p ]] 00:37:41.633 17:53:23 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@33 -- # run_test nvmf_digest /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/digest.sh --transport=tcp 00:37:41.633 17:53:23 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:37:41.633 17:53:23 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:37:41.633 17:53:23 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:37:41.633 ************************************ 00:37:41.633 START TEST nvmf_digest 00:37:41.633 ************************************ 00:37:41.633 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/digest.sh --transport=tcp 00:37:41.633 * Looking for test storage... 00:37:41.633 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:37:41.633 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:37:41.633 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1711 -- # lcov --version 00:37:41.633 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:37:41.892 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:37:41.892 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:37:41.892 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@333 -- # local ver1 ver1_l 00:37:41.892 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@334 -- # local ver2 ver2_l 00:37:41.892 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@336 -- # IFS=.-: 00:37:41.892 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@336 -- # read -ra ver1 00:37:41.892 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@337 -- # IFS=.-: 00:37:41.892 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@337 -- # read -ra ver2 00:37:41.892 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@338 -- # local 'op=<' 00:37:41.892 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@340 -- # ver1_l=2 00:37:41.892 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@341 -- # ver2_l=1 00:37:41.892 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:37:41.892 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@344 -- # case "$op" in 00:37:41.892 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@345 -- # : 1 00:37:41.892 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@364 -- # (( v = 0 )) 00:37:41.892 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@365 -- # decimal 1 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@353 -- # local d=1 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@355 -- # echo 1 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@365 -- # ver1[v]=1 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@366 -- # decimal 2 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@353 -- # local d=2 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@355 -- # echo 2 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@366 -- # ver2[v]=2 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@368 -- # return 0 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:37:41.893 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:37:41.893 --rc genhtml_branch_coverage=1 00:37:41.893 --rc genhtml_function_coverage=1 00:37:41.893 --rc genhtml_legend=1 00:37:41.893 --rc geninfo_all_blocks=1 00:37:41.893 --rc geninfo_unexecuted_blocks=1 00:37:41.893 00:37:41.893 ' 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:37:41.893 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:37:41.893 --rc genhtml_branch_coverage=1 00:37:41.893 --rc genhtml_function_coverage=1 00:37:41.893 --rc genhtml_legend=1 00:37:41.893 --rc geninfo_all_blocks=1 00:37:41.893 --rc geninfo_unexecuted_blocks=1 00:37:41.893 00:37:41.893 ' 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:37:41.893 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:37:41.893 --rc genhtml_branch_coverage=1 00:37:41.893 --rc genhtml_function_coverage=1 00:37:41.893 --rc genhtml_legend=1 00:37:41.893 --rc geninfo_all_blocks=1 00:37:41.893 --rc geninfo_unexecuted_blocks=1 00:37:41.893 00:37:41.893 ' 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:37:41.893 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:37:41.893 --rc genhtml_branch_coverage=1 00:37:41.893 --rc genhtml_function_coverage=1 00:37:41.893 --rc genhtml_legend=1 00:37:41.893 --rc geninfo_all_blocks=1 00:37:41.893 --rc geninfo_unexecuted_blocks=1 00:37:41.893 00:37:41.893 ' 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@12 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@7 -- # uname -s 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@15 -- # shopt -s extglob 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- paths/export.sh@5 -- # export PATH 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@51 -- # : 0 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:37:41.893 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@55 -- # have_pci_nics=0 00:37:41.893 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@14 -- # nqn=nqn.2016-06.io.spdk:cnode1 00:37:41.894 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@15 -- # bperfsock=/var/tmp/bperf.sock 00:37:41.894 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@16 -- # runtime=2 00:37:41.894 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@136 -- # [[ tcp != \t\c\p ]] 00:37:41.894 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@138 -- # nvmftestinit 00:37:41.894 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:37:41.894 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:37:41.894 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@476 -- # prepare_net_devs 00:37:41.894 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@438 -- # local -g is_hw=no 00:37:41.894 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@440 -- # remove_spdk_ns 00:37:41.894 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:37:41.894 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:37:41.894 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:37:41.894 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:37:41.894 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:37:41.894 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@309 -- # xtrace_disable 00:37:41.894 17:53:23 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@10 -- # set +x 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@315 -- # pci_devs=() 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@315 -- # local -a pci_devs 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@316 -- # pci_net_devs=() 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@317 -- # pci_drivers=() 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@317 -- # local -A pci_drivers 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@319 -- # net_devs=() 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@319 -- # local -ga net_devs 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@320 -- # e810=() 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@320 -- # local -ga e810 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@321 -- # x722=() 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@321 -- # local -ga x722 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@322 -- # mlx=() 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@322 -- # local -ga mlx 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:37:44.427 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:37:44.427 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@418 -- # [[ up == up ]] 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:37:44.427 Found net devices under 0000:0a:00.0: cvl_0_0 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@418 -- # [[ up == up ]] 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:37:44.427 Found net devices under 0000:0a:00.1: cvl_0_1 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@442 -- # is_hw=yes 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:37:44.427 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:37:44.427 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.161 ms 00:37:44.427 00:37:44.427 --- 10.0.0.2 ping statistics --- 00:37:44.427 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:37:44.427 rtt min/avg/max/mdev = 0.161/0.161/0.161/0.000 ms 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:37:44.427 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:37:44.427 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.051 ms 00:37:44.427 00:37:44.427 --- 10.0.0.1 ping statistics --- 00:37:44.427 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:37:44.427 rtt min/avg/max/mdev = 0.051/0.051/0.051/0.000 ms 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@450 -- # return 0 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@140 -- # trap cleanup SIGINT SIGTERM EXIT 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@141 -- # [[ 0 -eq 1 ]] 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@145 -- # run_test nvmf_digest_clean run_digest 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1111 -- # xtrace_disable 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@10 -- # set +x 00:37:44.427 ************************************ 00:37:44.427 START TEST nvmf_digest_clean 00:37:44.427 ************************************ 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@1129 -- # run_digest 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@120 -- # local dsa_initiator 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@121 -- # [[ '' == \d\s\a\_\i\n\i\t\i\a\t\o\r ]] 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@121 -- # dsa_initiator=false 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@123 -- # tgt_params=("--wait-for-rpc") 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@124 -- # nvmfappstart --wait-for-rpc 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@726 -- # xtrace_disable 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@509 -- # nvmfpid=404358 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --wait-for-rpc 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@510 -- # waitforlisten 404358 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@835 -- # '[' -z 404358 ']' 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@840 -- # local max_retries=100 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:37:44.427 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@844 -- # xtrace_disable 00:37:44.427 17:53:25 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:37:44.427 [2024-12-06 17:53:25.883166] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:37:44.427 [2024-12-06 17:53:25.883245] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:37:44.427 [2024-12-06 17:53:25.956044] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:37:44.427 [2024-12-06 17:53:26.000681] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:37:44.427 [2024-12-06 17:53:26.000735] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:37:44.427 [2024-12-06 17:53:26.000759] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:37:44.427 [2024-12-06 17:53:26.000769] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:37:44.427 [2024-12-06 17:53:26.000780] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:37:44.427 [2024-12-06 17:53:26.001325] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:37:44.427 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:37:44.427 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@868 -- # return 0 00:37:44.427 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:37:44.427 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@732 -- # xtrace_disable 00:37:44.427 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:37:44.427 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:37:44.427 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@125 -- # [[ '' == \d\s\a\_\t\a\r\g\e\t ]] 00:37:44.427 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@126 -- # common_target_config 00:37:44.427 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@43 -- # rpc_cmd 00:37:44.427 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:44.427 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:37:44.427 null0 00:37:44.427 [2024-12-06 17:53:26.249830] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:37:44.685 [2024-12-06 17:53:26.274099] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:37:44.685 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:44.685 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@128 -- # run_bperf randread 4096 128 false 00:37:44.685 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@77 -- # local rw bs qd scan_dsa 00:37:44.685 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@78 -- # local acc_module acc_executed exp_module 00:37:44.685 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # rw=randread 00:37:44.685 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # bs=4096 00:37:44.685 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # qd=128 00:37:44.685 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # scan_dsa=false 00:37:44.685 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@83 -- # bperfpid=404378 00:37:44.685 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randread -o 4096 -t 2 -q 128 -z --wait-for-rpc 00:37:44.685 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@84 -- # waitforlisten 404378 /var/tmp/bperf.sock 00:37:44.685 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@835 -- # '[' -z 404378 ']' 00:37:44.685 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bperf.sock 00:37:44.685 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@840 -- # local max_retries=100 00:37:44.685 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:37:44.685 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:37:44.685 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@844 -- # xtrace_disable 00:37:44.685 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:37:44.685 [2024-12-06 17:53:26.321065] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:37:44.685 [2024-12-06 17:53:26.321127] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid404378 ] 00:37:44.685 [2024-12-06 17:53:26.389906] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:37:44.685 [2024-12-06 17:53:26.440199] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:37:44.942 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:37:44.942 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@868 -- # return 0 00:37:44.942 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@86 -- # false 00:37:44.942 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@87 -- # bperf_rpc framework_start_init 00:37:44.942 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:37:45.200 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@89 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:37:45.200 17:53:26 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:37:45.457 nvme0n1 00:37:45.457 17:53:27 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@92 -- # bperf_py perform_tests 00:37:45.457 17:53:27 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:37:45.714 Running I/O for 2 seconds... 00:37:47.616 18649.00 IOPS, 72.85 MiB/s [2024-12-06T16:53:29.455Z] 18670.00 IOPS, 72.93 MiB/s 00:37:47.616 Latency(us) 00:37:47.616 [2024-12-06T16:53:29.455Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:37:47.616 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 128, IO size: 4096) 00:37:47.616 nvme0n1 : 2.01 18670.47 72.93 0.00 0.00 6846.33 3422.44 15340.28 00:37:47.616 [2024-12-06T16:53:29.455Z] =================================================================================================================== 00:37:47.616 [2024-12-06T16:53:29.455Z] Total : 18670.47 72.93 0.00 0.00 6846.33 3422.44 15340.28 00:37:47.616 { 00:37:47.616 "results": [ 00:37:47.616 { 00:37:47.616 "job": "nvme0n1", 00:37:47.616 "core_mask": "0x2", 00:37:47.616 "workload": "randread", 00:37:47.616 "status": "finished", 00:37:47.616 "queue_depth": 128, 00:37:47.616 "io_size": 4096, 00:37:47.616 "runtime": 2.01018, 00:37:47.616 "iops": 18670.467321334407, 00:37:47.616 "mibps": 72.93151297396253, 00:37:47.616 "io_failed": 0, 00:37:47.616 "io_timeout": 0, 00:37:47.616 "avg_latency_us": 6846.327880576748, 00:37:47.616 "min_latency_us": 3422.4355555555558, 00:37:47.616 "max_latency_us": 15340.278518518518 00:37:47.616 } 00:37:47.616 ], 00:37:47.616 "core_count": 1 00:37:47.616 } 00:37:47.616 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # read -r acc_module acc_executed 00:37:47.616 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # get_accel_stats 00:37:47.616 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@36 -- # bperf_rpc accel_get_stats 00:37:47.616 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@37 -- # jq -rc '.operations[] 00:37:47.616 | select(.opcode=="crc32c") 00:37:47.616 | "\(.module_name) \(.executed)"' 00:37:47.616 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:37:47.873 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # false 00:37:47.873 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # exp_module=software 00:37:47.873 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@95 -- # (( acc_executed > 0 )) 00:37:47.873 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@96 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:37:47.873 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@98 -- # killprocess 404378 00:37:47.873 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # '[' -z 404378 ']' 00:37:47.873 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@958 -- # kill -0 404378 00:37:47.873 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@959 -- # uname 00:37:47.873 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:37:47.873 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 404378 00:37:48.132 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:37:48.132 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:37:48.132 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@972 -- # echo 'killing process with pid 404378' 00:37:48.132 killing process with pid 404378 00:37:48.132 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@973 -- # kill 404378 00:37:48.132 Received shutdown signal, test time was about 2.000000 seconds 00:37:48.132 00:37:48.132 Latency(us) 00:37:48.132 [2024-12-06T16:53:29.971Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:37:48.132 [2024-12-06T16:53:29.971Z] =================================================================================================================== 00:37:48.132 [2024-12-06T16:53:29.971Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:37:48.132 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@978 -- # wait 404378 00:37:48.132 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@129 -- # run_bperf randread 131072 16 false 00:37:48.132 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@77 -- # local rw bs qd scan_dsa 00:37:48.132 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@78 -- # local acc_module acc_executed exp_module 00:37:48.132 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # rw=randread 00:37:48.132 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # bs=131072 00:37:48.132 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # qd=16 00:37:48.132 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # scan_dsa=false 00:37:48.132 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@83 -- # bperfpid=404906 00:37:48.132 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randread -o 131072 -t 2 -q 16 -z --wait-for-rpc 00:37:48.132 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@84 -- # waitforlisten 404906 /var/tmp/bperf.sock 00:37:48.132 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@835 -- # '[' -z 404906 ']' 00:37:48.132 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bperf.sock 00:37:48.132 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@840 -- # local max_retries=100 00:37:48.132 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:37:48.132 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:37:48.132 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@844 -- # xtrace_disable 00:37:48.132 17:53:29 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:37:48.132 [2024-12-06 17:53:29.969375] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:37:48.132 [2024-12-06 17:53:29.969451] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid404906 ] 00:37:48.132 I/O size of 131072 is greater than zero copy threshold (65536). 00:37:48.132 Zero copy mechanism will not be used. 00:37:48.389 [2024-12-06 17:53:30.039176] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:37:48.389 [2024-12-06 17:53:30.088363] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:37:48.389 17:53:30 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:37:48.389 17:53:30 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@868 -- # return 0 00:37:48.389 17:53:30 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@86 -- # false 00:37:48.389 17:53:30 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@87 -- # bperf_rpc framework_start_init 00:37:48.389 17:53:30 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:37:48.954 17:53:30 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@89 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:37:48.954 17:53:30 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:37:49.211 nvme0n1 00:37:49.211 17:53:30 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@92 -- # bperf_py perform_tests 00:37:49.211 17:53:30 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:37:49.211 I/O size of 131072 is greater than zero copy threshold (65536). 00:37:49.211 Zero copy mechanism will not be used. 00:37:49.211 Running I/O for 2 seconds... 00:37:51.513 5736.00 IOPS, 717.00 MiB/s [2024-12-06T16:53:33.352Z] 5962.50 IOPS, 745.31 MiB/s 00:37:51.513 Latency(us) 00:37:51.513 [2024-12-06T16:53:33.352Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:37:51.513 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 16, IO size: 131072) 00:37:51.513 nvme0n1 : 2.00 5960.50 745.06 0.00 0.00 2680.30 691.77 10631.40 00:37:51.513 [2024-12-06T16:53:33.352Z] =================================================================================================================== 00:37:51.513 [2024-12-06T16:53:33.352Z] Total : 5960.50 745.06 0.00 0.00 2680.30 691.77 10631.40 00:37:51.513 { 00:37:51.513 "results": [ 00:37:51.513 { 00:37:51.513 "job": "nvme0n1", 00:37:51.513 "core_mask": "0x2", 00:37:51.513 "workload": "randread", 00:37:51.513 "status": "finished", 00:37:51.513 "queue_depth": 16, 00:37:51.513 "io_size": 131072, 00:37:51.513 "runtime": 2.003355, 00:37:51.513 "iops": 5960.501259137796, 00:37:51.513 "mibps": 745.0626573922245, 00:37:51.513 "io_failed": 0, 00:37:51.513 "io_timeout": 0, 00:37:51.513 "avg_latency_us": 2680.2978863362146, 00:37:51.513 "min_latency_us": 691.7688888888889, 00:37:51.513 "max_latency_us": 10631.395555555555 00:37:51.513 } 00:37:51.513 ], 00:37:51.513 "core_count": 1 00:37:51.513 } 00:37:51.513 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # read -r acc_module acc_executed 00:37:51.513 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # get_accel_stats 00:37:51.513 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@36 -- # bperf_rpc accel_get_stats 00:37:51.513 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@37 -- # jq -rc '.operations[] 00:37:51.513 | select(.opcode=="crc32c") 00:37:51.513 | "\(.module_name) \(.executed)"' 00:37:51.513 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:37:51.513 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # false 00:37:51.513 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # exp_module=software 00:37:51.513 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@95 -- # (( acc_executed > 0 )) 00:37:51.513 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@96 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:37:51.513 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@98 -- # killprocess 404906 00:37:51.513 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # '[' -z 404906 ']' 00:37:51.513 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@958 -- # kill -0 404906 00:37:51.513 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@959 -- # uname 00:37:51.770 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:37:51.770 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 404906 00:37:51.770 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:37:51.770 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:37:51.770 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@972 -- # echo 'killing process with pid 404906' 00:37:51.770 killing process with pid 404906 00:37:51.770 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@973 -- # kill 404906 00:37:51.770 Received shutdown signal, test time was about 2.000000 seconds 00:37:51.770 00:37:51.770 Latency(us) 00:37:51.770 [2024-12-06T16:53:33.609Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:37:51.770 [2024-12-06T16:53:33.609Z] =================================================================================================================== 00:37:51.770 [2024-12-06T16:53:33.609Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:37:51.770 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@978 -- # wait 404906 00:37:51.770 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@130 -- # run_bperf randwrite 4096 128 false 00:37:51.770 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@77 -- # local rw bs qd scan_dsa 00:37:51.770 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@78 -- # local acc_module acc_executed exp_module 00:37:51.770 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # rw=randwrite 00:37:51.770 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # bs=4096 00:37:51.770 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # qd=128 00:37:51.770 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # scan_dsa=false 00:37:51.770 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@83 -- # bperfpid=405313 00:37:51.770 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randwrite -o 4096 -t 2 -q 128 -z --wait-for-rpc 00:37:51.770 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@84 -- # waitforlisten 405313 /var/tmp/bperf.sock 00:37:51.770 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@835 -- # '[' -z 405313 ']' 00:37:51.770 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bperf.sock 00:37:51.770 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@840 -- # local max_retries=100 00:37:51.770 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:37:51.770 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:37:51.770 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@844 -- # xtrace_disable 00:37:51.770 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:37:52.029 [2024-12-06 17:53:33.636910] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:37:52.029 [2024-12-06 17:53:33.637042] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid405313 ] 00:37:52.029 [2024-12-06 17:53:33.707231] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:37:52.029 [2024-12-06 17:53:33.751824] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:37:52.287 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:37:52.287 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@868 -- # return 0 00:37:52.287 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@86 -- # false 00:37:52.287 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@87 -- # bperf_rpc framework_start_init 00:37:52.287 17:53:33 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:37:52.544 17:53:34 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@89 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:37:52.544 17:53:34 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:37:52.802 nvme0n1 00:37:52.802 17:53:34 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@92 -- # bperf_py perform_tests 00:37:52.802 17:53:34 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:37:53.059 Running I/O for 2 seconds... 00:37:54.924 20606.00 IOPS, 80.49 MiB/s [2024-12-06T16:53:36.763Z] 20095.00 IOPS, 78.50 MiB/s 00:37:54.924 Latency(us) 00:37:54.924 [2024-12-06T16:53:36.763Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:37:54.924 Job: nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:37:54.924 nvme0n1 : 2.01 20098.15 78.51 0.00 0.00 6354.99 2706.39 11650.84 00:37:54.924 [2024-12-06T16:53:36.763Z] =================================================================================================================== 00:37:54.924 [2024-12-06T16:53:36.763Z] Total : 20098.15 78.51 0.00 0.00 6354.99 2706.39 11650.84 00:37:54.924 { 00:37:54.924 "results": [ 00:37:54.924 { 00:37:54.924 "job": "nvme0n1", 00:37:54.924 "core_mask": "0x2", 00:37:54.924 "workload": "randwrite", 00:37:54.924 "status": "finished", 00:37:54.924 "queue_depth": 128, 00:37:54.924 "io_size": 4096, 00:37:54.924 "runtime": 2.007647, 00:37:54.924 "iops": 20098.154705483583, 00:37:54.924 "mibps": 78.50841681829525, 00:37:54.924 "io_failed": 0, 00:37:54.924 "io_timeout": 0, 00:37:54.924 "avg_latency_us": 6354.992594061223, 00:37:54.924 "min_latency_us": 2706.394074074074, 00:37:54.924 "max_latency_us": 11650.844444444445 00:37:54.924 } 00:37:54.924 ], 00:37:54.924 "core_count": 1 00:37:54.924 } 00:37:54.924 17:53:36 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # read -r acc_module acc_executed 00:37:54.924 17:53:36 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # get_accel_stats 00:37:54.924 17:53:36 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@36 -- # bperf_rpc accel_get_stats 00:37:54.924 17:53:36 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@37 -- # jq -rc '.operations[] 00:37:54.924 | select(.opcode=="crc32c") 00:37:54.924 | "\(.module_name) \(.executed)"' 00:37:54.924 17:53:36 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:37:55.182 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # false 00:37:55.182 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # exp_module=software 00:37:55.182 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@95 -- # (( acc_executed > 0 )) 00:37:55.182 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@96 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:37:55.182 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@98 -- # killprocess 405313 00:37:55.182 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # '[' -z 405313 ']' 00:37:55.182 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@958 -- # kill -0 405313 00:37:55.182 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@959 -- # uname 00:37:55.182 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:37:55.182 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 405313 00:37:55.440 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:37:55.440 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:37:55.440 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@972 -- # echo 'killing process with pid 405313' 00:37:55.440 killing process with pid 405313 00:37:55.440 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@973 -- # kill 405313 00:37:55.440 Received shutdown signal, test time was about 2.000000 seconds 00:37:55.440 00:37:55.440 Latency(us) 00:37:55.440 [2024-12-06T16:53:37.279Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:37:55.440 [2024-12-06T16:53:37.279Z] =================================================================================================================== 00:37:55.440 [2024-12-06T16:53:37.279Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:37:55.440 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@978 -- # wait 405313 00:37:55.440 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@131 -- # run_bperf randwrite 131072 16 false 00:37:55.440 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@77 -- # local rw bs qd scan_dsa 00:37:55.440 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@78 -- # local acc_module acc_executed exp_module 00:37:55.440 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # rw=randwrite 00:37:55.440 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # bs=131072 00:37:55.440 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # qd=16 00:37:55.440 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # scan_dsa=false 00:37:55.440 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@83 -- # bperfpid=405717 00:37:55.440 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randwrite -o 131072 -t 2 -q 16 -z --wait-for-rpc 00:37:55.440 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@84 -- # waitforlisten 405717 /var/tmp/bperf.sock 00:37:55.440 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@835 -- # '[' -z 405717 ']' 00:37:55.440 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bperf.sock 00:37:55.440 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@840 -- # local max_retries=100 00:37:55.440 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:37:55.440 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:37:55.440 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@844 -- # xtrace_disable 00:37:55.440 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:37:55.698 [2024-12-06 17:53:37.303318] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:37:55.698 [2024-12-06 17:53:37.303410] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid405717 ] 00:37:55.698 I/O size of 131072 is greater than zero copy threshold (65536). 00:37:55.698 Zero copy mechanism will not be used. 00:37:55.698 [2024-12-06 17:53:37.368773] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:37:55.698 [2024-12-06 17:53:37.411528] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:37:55.698 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:37:55.698 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@868 -- # return 0 00:37:55.698 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@86 -- # false 00:37:55.698 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@87 -- # bperf_rpc framework_start_init 00:37:55.698 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:37:56.264 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@89 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:37:56.264 17:53:37 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:37:56.522 nvme0n1 00:37:56.522 17:53:38 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@92 -- # bperf_py perform_tests 00:37:56.522 17:53:38 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:37:56.780 I/O size of 131072 is greater than zero copy threshold (65536). 00:37:56.780 Zero copy mechanism will not be used. 00:37:56.780 Running I/O for 2 seconds... 00:37:58.648 5487.00 IOPS, 685.88 MiB/s [2024-12-06T16:53:40.487Z] 5712.00 IOPS, 714.00 MiB/s 00:37:58.648 Latency(us) 00:37:58.648 [2024-12-06T16:53:40.487Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:37:58.648 Job: nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 16, IO size: 131072) 00:37:58.648 nvme0n1 : 2.00 5707.81 713.48 0.00 0.00 2795.64 1674.81 5218.61 00:37:58.648 [2024-12-06T16:53:40.487Z] =================================================================================================================== 00:37:58.648 [2024-12-06T16:53:40.487Z] Total : 5707.81 713.48 0.00 0.00 2795.64 1674.81 5218.61 00:37:58.648 { 00:37:58.648 "results": [ 00:37:58.648 { 00:37:58.648 "job": "nvme0n1", 00:37:58.648 "core_mask": "0x2", 00:37:58.648 "workload": "randwrite", 00:37:58.648 "status": "finished", 00:37:58.648 "queue_depth": 16, 00:37:58.648 "io_size": 131072, 00:37:58.648 "runtime": 2.004622, 00:37:58.648 "iops": 5707.809252816741, 00:37:58.648 "mibps": 713.4761566020926, 00:37:58.648 "io_failed": 0, 00:37:58.648 "io_timeout": 0, 00:37:58.648 "avg_latency_us": 2795.6354152019526, 00:37:58.648 "min_latency_us": 1674.808888888889, 00:37:58.648 "max_latency_us": 5218.607407407408 00:37:58.648 } 00:37:58.648 ], 00:37:58.648 "core_count": 1 00:37:58.648 } 00:37:58.648 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # read -r acc_module acc_executed 00:37:58.648 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # get_accel_stats 00:37:58.648 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@36 -- # bperf_rpc accel_get_stats 00:37:58.648 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:37:58.648 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@37 -- # jq -rc '.operations[] 00:37:58.648 | select(.opcode=="crc32c") 00:37:58.648 | "\(.module_name) \(.executed)"' 00:37:58.906 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # false 00:37:58.906 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # exp_module=software 00:37:58.906 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@95 -- # (( acc_executed > 0 )) 00:37:58.906 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@96 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:37:58.906 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@98 -- # killprocess 405717 00:37:58.906 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # '[' -z 405717 ']' 00:37:58.906 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@958 -- # kill -0 405717 00:37:58.906 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@959 -- # uname 00:37:58.906 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:37:58.906 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 405717 00:37:58.906 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:37:58.906 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:37:58.906 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@972 -- # echo 'killing process with pid 405717' 00:37:58.906 killing process with pid 405717 00:37:58.906 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@973 -- # kill 405717 00:37:58.906 Received shutdown signal, test time was about 2.000000 seconds 00:37:58.906 00:37:58.906 Latency(us) 00:37:58.906 [2024-12-06T16:53:40.745Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:37:58.906 [2024-12-06T16:53:40.745Z] =================================================================================================================== 00:37:58.906 [2024-12-06T16:53:40.745Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:37:58.906 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@978 -- # wait 405717 00:37:59.164 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@132 -- # killprocess 404358 00:37:59.164 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # '[' -z 404358 ']' 00:37:59.164 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@958 -- # kill -0 404358 00:37:59.164 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@959 -- # uname 00:37:59.164 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:37:59.164 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 404358 00:37:59.164 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:37:59.164 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:37:59.164 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@972 -- # echo 'killing process with pid 404358' 00:37:59.164 killing process with pid 404358 00:37:59.164 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@973 -- # kill 404358 00:37:59.164 17:53:40 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@978 -- # wait 404358 00:37:59.423 00:37:59.423 real 0m15.316s 00:37:59.423 user 0m30.746s 00:37:59.423 sys 0m4.247s 00:37:59.423 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@1130 -- # xtrace_disable 00:37:59.423 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:37:59.423 ************************************ 00:37:59.423 END TEST nvmf_digest_clean 00:37:59.423 ************************************ 00:37:59.423 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@147 -- # run_test nvmf_digest_error run_digest_error 00:37:59.423 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:37:59.423 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1111 -- # xtrace_disable 00:37:59.423 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@10 -- # set +x 00:37:59.423 ************************************ 00:37:59.423 START TEST nvmf_digest_error 00:37:59.423 ************************************ 00:37:59.423 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@1129 -- # run_digest_error 00:37:59.423 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@102 -- # nvmfappstart --wait-for-rpc 00:37:59.423 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:37:59.423 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@726 -- # xtrace_disable 00:37:59.423 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:37:59.423 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@509 -- # nvmfpid=406247 00:37:59.423 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --wait-for-rpc 00:37:59.423 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@510 -- # waitforlisten 406247 00:37:59.423 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@835 -- # '[' -z 406247 ']' 00:37:59.423 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:37:59.423 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@840 -- # local max_retries=100 00:37:59.423 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:37:59.423 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:37:59.423 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@844 -- # xtrace_disable 00:37:59.423 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:37:59.423 [2024-12-06 17:53:41.254994] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:37:59.423 [2024-12-06 17:53:41.255086] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:37:59.681 [2024-12-06 17:53:41.330469] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:37:59.681 [2024-12-06 17:53:41.374422] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:37:59.681 [2024-12-06 17:53:41.374476] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:37:59.681 [2024-12-06 17:53:41.374500] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:37:59.681 [2024-12-06 17:53:41.374511] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:37:59.681 [2024-12-06 17:53:41.374521] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:37:59.681 [2024-12-06 17:53:41.375106] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:37:59.681 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:37:59.681 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@868 -- # return 0 00:37:59.681 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:37:59.681 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@732 -- # xtrace_disable 00:37:59.681 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:37:59.681 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:37:59.681 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@104 -- # rpc_cmd accel_assign_opc -o crc32c -m error 00:37:59.681 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:59.681 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:37:59.681 [2024-12-06 17:53:41.503808] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation crc32c will be assigned to module error 00:37:59.681 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:59.681 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@105 -- # common_target_config 00:37:59.681 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@43 -- # rpc_cmd 00:37:59.681 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@563 -- # xtrace_disable 00:37:59.681 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:37:59.939 null0 00:37:59.939 [2024-12-06 17:53:41.619748] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:37:59.939 [2024-12-06 17:53:41.643968] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:37:59.939 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:37:59.939 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@108 -- # run_bperf_err randread 4096 128 00:37:59.939 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@54 -- # local rw bs qd 00:37:59.939 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # rw=randread 00:37:59.939 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # bs=4096 00:37:59.939 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # qd=128 00:37:59.939 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@58 -- # bperfpid=406299 00:37:59.939 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@60 -- # waitforlisten 406299 /var/tmp/bperf.sock 00:37:59.939 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randread -o 4096 -t 2 -q 128 -z 00:37:59.939 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@835 -- # '[' -z 406299 ']' 00:37:59.939 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bperf.sock 00:37:59.939 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@840 -- # local max_retries=100 00:37:59.939 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:37:59.939 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:37:59.939 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@844 -- # xtrace_disable 00:37:59.939 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:37:59.939 [2024-12-06 17:53:41.691376] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:37:59.939 [2024-12-06 17:53:41.691441] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid406299 ] 00:37:59.939 [2024-12-06 17:53:41.756532] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:38:00.197 [2024-12-06 17:53:41.803131] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:38:00.197 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:38:00.197 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@868 -- # return 0 00:38:00.197 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@61 -- # bperf_rpc bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:38:00.197 17:53:41 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:38:00.762 17:53:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@63 -- # rpc_cmd accel_error_inject_error -o crc32c -t disable 00:38:00.762 17:53:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:00.762 17:53:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:38:00.762 17:53:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:00.762 17:53:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@64 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:38:00.763 17:53:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:38:01.021 nvme0n1 00:38:01.021 17:53:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@67 -- # rpc_cmd accel_error_inject_error -o crc32c -t corrupt -i 256 00:38:01.021 17:53:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:01.021 17:53:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:38:01.021 17:53:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:01.021 17:53:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@69 -- # bperf_py perform_tests 00:38:01.021 17:53:42 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:38:01.021 Running I/O for 2 seconds... 00:38:01.021 [2024-12-06 17:53:42.830702] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.021 [2024-12-06 17:53:42.830761] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:23598 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.021 [2024-12-06 17:53:42.830794] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.021 [2024-12-06 17:53:42.844955] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.021 [2024-12-06 17:53:42.845001] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:14191 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.021 [2024-12-06 17:53:42.845018] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:91 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.312 [2024-12-06 17:53:42.862350] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.312 [2024-12-06 17:53:42.862384] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:20688 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.312 [2024-12-06 17:53:42.862403] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:24 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.312 [2024-12-06 17:53:42.878975] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.312 [2024-12-06 17:53:42.879008] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:3141 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.312 [2024-12-06 17:53:42.879030] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:26 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.312 [2024-12-06 17:53:42.892056] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.312 [2024-12-06 17:53:42.892085] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:12807 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.312 [2024-12-06 17:53:42.892105] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:24 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.312 [2024-12-06 17:53:42.907753] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.312 [2024-12-06 17:53:42.907784] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:96 nsid:1 lba:15210 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.312 [2024-12-06 17:53:42.907801] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:96 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.312 [2024-12-06 17:53:42.921346] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.312 [2024-12-06 17:53:42.921375] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:20304 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.312 [2024-12-06 17:53:42.921399] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:67 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.312 [2024-12-06 17:53:42.933568] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.312 [2024-12-06 17:53:42.933597] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:79 nsid:1 lba:10774 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.312 [2024-12-06 17:53:42.933614] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:79 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.312 [2024-12-06 17:53:42.948775] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.312 [2024-12-06 17:53:42.948803] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:4072 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.312 [2024-12-06 17:53:42.948824] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:109 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.312 [2024-12-06 17:53:42.960033] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.312 [2024-12-06 17:53:42.960067] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:104 nsid:1 lba:14144 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.312 [2024-12-06 17:53:42.960087] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:104 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.312 [2024-12-06 17:53:42.974092] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.312 [2024-12-06 17:53:42.974119] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:78 nsid:1 lba:11403 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.312 [2024-12-06 17:53:42.974139] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:78 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.312 [2024-12-06 17:53:42.990116] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.312 [2024-12-06 17:53:42.990145] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:20173 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.312 [2024-12-06 17:53:42.990164] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:113 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.312 [2024-12-06 17:53:43.003330] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.312 [2024-12-06 17:53:43.003359] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:23574 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.312 [2024-12-06 17:53:43.003379] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:67 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.312 [2024-12-06 17:53:43.015564] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.312 [2024-12-06 17:53:43.015592] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:13249 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.312 [2024-12-06 17:53:43.015610] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:126 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.312 [2024-12-06 17:53:43.030020] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.312 [2024-12-06 17:53:43.030050] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:106 nsid:1 lba:24384 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.313 [2024-12-06 17:53:43.030070] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:106 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.313 [2024-12-06 17:53:43.044367] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.313 [2024-12-06 17:53:43.044413] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:17777 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.313 [2024-12-06 17:53:43.044436] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:108 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.313 [2024-12-06 17:53:43.055774] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.313 [2024-12-06 17:53:43.055804] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:96 nsid:1 lba:25415 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.313 [2024-12-06 17:53:43.055824] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:96 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.313 [2024-12-06 17:53:43.070281] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.313 [2024-12-06 17:53:43.070309] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:7088 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.313 [2024-12-06 17:53:43.070351] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:39 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.313 [2024-12-06 17:53:43.083896] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.313 [2024-12-06 17:53:43.083927] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:106 nsid:1 lba:18333 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.313 [2024-12-06 17:53:43.083946] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:106 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.313 [2024-12-06 17:53:43.096939] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.313 [2024-12-06 17:53:43.096984] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:106 nsid:1 lba:2558 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.313 [2024-12-06 17:53:43.097000] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:106 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.313 [2024-12-06 17:53:43.112718] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.313 [2024-12-06 17:53:43.112764] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:7839 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.313 [2024-12-06 17:53:43.112782] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:20 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.600 [2024-12-06 17:53:43.126689] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.600 [2024-12-06 17:53:43.126721] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:106 nsid:1 lba:21949 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.600 [2024-12-06 17:53:43.126740] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:106 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.600 [2024-12-06 17:53:43.140761] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.600 [2024-12-06 17:53:43.140789] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:10559 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.600 [2024-12-06 17:53:43.140811] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:44 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.600 [2024-12-06 17:53:43.156787] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.600 [2024-12-06 17:53:43.156818] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:20317 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.600 [2024-12-06 17:53:43.156837] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:55 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.600 [2024-12-06 17:53:43.174956] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.600 [2024-12-06 17:53:43.174999] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:20917 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.600 [2024-12-06 17:53:43.175019] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:49 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.600 [2024-12-06 17:53:43.187888] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.600 [2024-12-06 17:53:43.187918] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:64 nsid:1 lba:23180 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.600 [2024-12-06 17:53:43.187941] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:64 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.600 [2024-12-06 17:53:43.202554] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.600 [2024-12-06 17:53:43.202603] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:23859 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.600 [2024-12-06 17:53:43.202623] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:42 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.600 [2024-12-06 17:53:43.216861] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.600 [2024-12-06 17:53:43.216893] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:124 nsid:1 lba:23911 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.600 [2024-12-06 17:53:43.216913] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:124 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.600 [2024-12-06 17:53:43.229355] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.600 [2024-12-06 17:53:43.229384] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:23694 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.600 [2024-12-06 17:53:43.229402] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:32 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.600 [2024-12-06 17:53:43.243608] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.600 [2024-12-06 17:53:43.243638] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:683 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.600 [2024-12-06 17:53:43.243680] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:34 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.600 [2024-12-06 17:53:43.256973] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.600 [2024-12-06 17:53:43.257018] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:5110 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.600 [2024-12-06 17:53:43.257036] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.600 [2024-12-06 17:53:43.269604] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.600 [2024-12-06 17:53:43.269634] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:24293 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.600 [2024-12-06 17:53:43.269654] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:29 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.600 [2024-12-06 17:53:43.283695] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.600 [2024-12-06 17:53:43.283740] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:6559 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.600 [2024-12-06 17:53:43.283756] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:99 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.600 [2024-12-06 17:53:43.296472] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.600 [2024-12-06 17:53:43.296500] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:14495 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.600 [2024-12-06 17:53:43.296537] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.600 [2024-12-06 17:53:43.308782] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.601 [2024-12-06 17:53:43.308813] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:104 nsid:1 lba:10103 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.601 [2024-12-06 17:53:43.308832] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:104 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.601 [2024-12-06 17:53:43.323585] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.601 [2024-12-06 17:53:43.323615] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:2961 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.601 [2024-12-06 17:53:43.323632] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:80 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.601 [2024-12-06 17:53:43.334374] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.601 [2024-12-06 17:53:43.334404] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:1999 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.601 [2024-12-06 17:53:43.334421] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:25 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.601 [2024-12-06 17:53:43.350084] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.601 [2024-12-06 17:53:43.350116] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:64 nsid:1 lba:25302 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.601 [2024-12-06 17:53:43.350136] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:64 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.601 [2024-12-06 17:53:43.364356] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.601 [2024-12-06 17:53:43.364386] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:96 nsid:1 lba:5394 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.601 [2024-12-06 17:53:43.364406] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:96 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.601 [2024-12-06 17:53:43.378002] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.601 [2024-12-06 17:53:43.378034] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:17942 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.601 [2024-12-06 17:53:43.378052] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:99 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.601 [2024-12-06 17:53:43.390661] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.601 [2024-12-06 17:53:43.390717] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:6257 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.601 [2024-12-06 17:53:43.390735] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:80 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.601 [2024-12-06 17:53:43.402920] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.601 [2024-12-06 17:53:43.402977] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:18869 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.601 [2024-12-06 17:53:43.402994] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.859 [2024-12-06 17:53:43.417390] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.859 [2024-12-06 17:53:43.417419] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:112 nsid:1 lba:6038 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.859 [2024-12-06 17:53:43.417438] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:112 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.859 [2024-12-06 17:53:43.430034] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.859 [2024-12-06 17:53:43.430063] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:3595 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.859 [2024-12-06 17:53:43.430083] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:58 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.859 [2024-12-06 17:53:43.445543] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.859 [2024-12-06 17:53:43.445572] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:18765 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.859 [2024-12-06 17:53:43.445593] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.859 [2024-12-06 17:53:43.458409] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.859 [2024-12-06 17:53:43.458439] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:396 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.859 [2024-12-06 17:53:43.458470] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.859 [2024-12-06 17:53:43.470898] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.859 [2024-12-06 17:53:43.470931] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:24895 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.859 [2024-12-06 17:53:43.470964] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:65 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.859 [2024-12-06 17:53:43.486084] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.859 [2024-12-06 17:53:43.486112] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:25083 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.859 [2024-12-06 17:53:43.486128] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:14 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.859 [2024-12-06 17:53:43.498519] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.859 [2024-12-06 17:53:43.498547] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:10904 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.859 [2024-12-06 17:53:43.498562] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:52 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.859 [2024-12-06 17:53:43.510083] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.859 [2024-12-06 17:53:43.510111] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:16539 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.859 [2024-12-06 17:53:43.510126] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:54 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.859 [2024-12-06 17:53:43.524094] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.859 [2024-12-06 17:53:43.524123] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:6588 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.859 [2024-12-06 17:53:43.524140] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:36 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.859 [2024-12-06 17:53:43.537245] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.859 [2024-12-06 17:53:43.537289] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:68 nsid:1 lba:2399 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.859 [2024-12-06 17:53:43.537305] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:68 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.859 [2024-12-06 17:53:43.549299] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.859 [2024-12-06 17:53:43.549334] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:15606 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.859 [2024-12-06 17:53:43.549351] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:17 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.859 [2024-12-06 17:53:43.564803] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.859 [2024-12-06 17:53:43.564834] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:71 nsid:1 lba:13589 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.859 [2024-12-06 17:53:43.564852] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:71 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.859 [2024-12-06 17:53:43.578437] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.859 [2024-12-06 17:53:43.578466] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:13353 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.859 [2024-12-06 17:53:43.578482] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:51 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.859 [2024-12-06 17:53:43.589993] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.859 [2024-12-06 17:53:43.590021] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:19105 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.859 [2024-12-06 17:53:43.590037] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.859 [2024-12-06 17:53:43.605404] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.859 [2024-12-06 17:53:43.605435] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:73 nsid:1 lba:11134 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.859 [2024-12-06 17:53:43.605452] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:73 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.859 [2024-12-06 17:53:43.622428] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.859 [2024-12-06 17:53:43.622458] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:117 nsid:1 lba:13085 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.859 [2024-12-06 17:53:43.622474] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:117 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.859 [2024-12-06 17:53:43.632517] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.859 [2024-12-06 17:53:43.632545] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:11540 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.859 [2024-12-06 17:53:43.632561] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:113 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.859 [2024-12-06 17:53:43.647270] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.859 [2024-12-06 17:53:43.647314] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:5401 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.859 [2024-12-06 17:53:43.647331] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:26 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.859 [2024-12-06 17:53:43.660233] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.859 [2024-12-06 17:53:43.660262] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:18869 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.859 [2024-12-06 17:53:43.660284] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:31 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.859 [2024-12-06 17:53:43.674700] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.859 [2024-12-06 17:53:43.674733] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:14017 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.859 [2024-12-06 17:53:43.674751] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:60 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:01.859 [2024-12-06 17:53:43.687070] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:01.859 [2024-12-06 17:53:43.687101] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:117 nsid:1 lba:19580 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:01.859 [2024-12-06 17:53:43.687118] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:117 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.117 [2024-12-06 17:53:43.701312] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.117 [2024-12-06 17:53:43.701340] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:13074 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.117 [2024-12-06 17:53:43.701356] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.117 [2024-12-06 17:53:43.715543] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.117 [2024-12-06 17:53:43.715573] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:10639 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.117 [2024-12-06 17:53:43.715590] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:44 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.117 [2024-12-06 17:53:43.726672] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.117 [2024-12-06 17:53:43.726716] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:77 nsid:1 lba:24069 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.117 [2024-12-06 17:53:43.726731] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:77 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.117 [2024-12-06 17:53:43.740527] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.117 [2024-12-06 17:53:43.740557] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:13260 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.117 [2024-12-06 17:53:43.740574] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.117 [2024-12-06 17:53:43.754166] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.117 [2024-12-06 17:53:43.754195] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:966 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.117 [2024-12-06 17:53:43.754211] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:19 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.117 [2024-12-06 17:53:43.768442] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.117 [2024-12-06 17:53:43.768471] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:73 nsid:1 lba:5493 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.118 [2024-12-06 17:53:43.768503] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:73 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.118 [2024-12-06 17:53:43.779613] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.118 [2024-12-06 17:53:43.779646] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:20186 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.118 [2024-12-06 17:53:43.779662] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:42 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.118 [2024-12-06 17:53:43.793180] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.118 [2024-12-06 17:53:43.793210] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:996 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.118 [2024-12-06 17:53:43.793226] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.118 [2024-12-06 17:53:43.806181] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.118 [2024-12-06 17:53:43.806210] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:19804 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.118 [2024-12-06 17:53:43.806225] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:48 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.118 18553.00 IOPS, 72.47 MiB/s [2024-12-06T16:53:43.957Z] [2024-12-06 17:53:43.819171] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.118 [2024-12-06 17:53:43.819200] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:11834 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.118 [2024-12-06 17:53:43.819216] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:103 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.118 [2024-12-06 17:53:43.833361] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.118 [2024-12-06 17:53:43.833391] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:17698 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.118 [2024-12-06 17:53:43.833408] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:109 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.118 [2024-12-06 17:53:43.846376] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.118 [2024-12-06 17:53:43.846407] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:15119 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.118 [2024-12-06 17:53:43.846423] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:58 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.118 [2024-12-06 17:53:43.859498] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.118 [2024-12-06 17:53:43.859527] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:24304 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.118 [2024-12-06 17:53:43.859543] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:55 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.118 [2024-12-06 17:53:43.873130] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.118 [2024-12-06 17:53:43.873160] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:117 nsid:1 lba:19748 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.118 [2024-12-06 17:53:43.873177] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:117 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.118 [2024-12-06 17:53:43.884941] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.118 [2024-12-06 17:53:43.884971] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:19629 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.118 [2024-12-06 17:53:43.885001] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.118 [2024-12-06 17:53:43.897570] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.118 [2024-12-06 17:53:43.897599] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:83 nsid:1 lba:14092 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.118 [2024-12-06 17:53:43.897615] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:83 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.118 [2024-12-06 17:53:43.910557] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.118 [2024-12-06 17:53:43.910587] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:11915 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.118 [2024-12-06 17:53:43.910604] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:46 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.118 [2024-12-06 17:53:43.924818] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.118 [2024-12-06 17:53:43.924863] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:112 nsid:1 lba:4584 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.118 [2024-12-06 17:53:43.924880] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:112 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.118 [2024-12-06 17:53:43.935933] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.118 [2024-12-06 17:53:43.935962] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:117 nsid:1 lba:11069 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.118 [2024-12-06 17:53:43.935992] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:117 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.118 [2024-12-06 17:53:43.949960] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.118 [2024-12-06 17:53:43.950006] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:112 nsid:1 lba:25354 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.118 [2024-12-06 17:53:43.950023] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:112 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.376 [2024-12-06 17:53:43.963478] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.376 [2024-12-06 17:53:43.963508] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:6950 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.376 [2024-12-06 17:53:43.963525] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:22 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.376 [2024-12-06 17:53:43.976893] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.376 [2024-12-06 17:53:43.976922] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:3244 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.376 [2024-12-06 17:53:43.976938] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:91 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.376 [2024-12-06 17:53:43.989401] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.376 [2024-12-06 17:53:43.989444] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:14981 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.376 [2024-12-06 17:53:43.989459] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:109 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.376 [2024-12-06 17:53:44.002748] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.376 [2024-12-06 17:53:44.002777] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:117 nsid:1 lba:23971 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.376 [2024-12-06 17:53:44.002799] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:117 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.376 [2024-12-06 17:53:44.017278] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.376 [2024-12-06 17:53:44.017324] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:82 nsid:1 lba:7851 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.376 [2024-12-06 17:53:44.017341] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:82 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.376 [2024-12-06 17:53:44.028863] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.376 [2024-12-06 17:53:44.028895] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:10246 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.376 [2024-12-06 17:53:44.028913] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:113 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.376 [2024-12-06 17:53:44.043615] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.376 [2024-12-06 17:53:44.043643] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:11812 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.377 [2024-12-06 17:53:44.043659] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:32 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.377 [2024-12-06 17:53:44.058933] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.377 [2024-12-06 17:53:44.058977] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:15078 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.377 [2024-12-06 17:53:44.058993] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:38 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.377 [2024-12-06 17:53:44.069164] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.377 [2024-12-06 17:53:44.069192] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:8178 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.377 [2024-12-06 17:53:44.069207] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:40 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.377 [2024-12-06 17:53:44.084035] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.377 [2024-12-06 17:53:44.084063] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:25338 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.377 [2024-12-06 17:53:44.084078] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:122 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.377 [2024-12-06 17:53:44.100186] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.377 [2024-12-06 17:53:44.100217] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:17838 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.377 [2024-12-06 17:53:44.100233] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:50 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.377 [2024-12-06 17:53:44.116645] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.377 [2024-12-06 17:53:44.116696] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:14928 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.377 [2024-12-06 17:53:44.116715] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:90 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.377 [2024-12-06 17:53:44.131727] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.377 [2024-12-06 17:53:44.131759] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:17606 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.377 [2024-12-06 17:53:44.131778] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:39 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.377 [2024-12-06 17:53:44.143929] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.377 [2024-12-06 17:53:44.143961] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:14079 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.377 [2024-12-06 17:53:44.143992] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:122 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.377 [2024-12-06 17:53:44.155201] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.377 [2024-12-06 17:53:44.155230] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:7843 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.377 [2024-12-06 17:53:44.155247] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:47 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.377 [2024-12-06 17:53:44.168325] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.377 [2024-12-06 17:53:44.168355] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:1233 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.377 [2024-12-06 17:53:44.168371] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:19 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.377 [2024-12-06 17:53:44.181930] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.377 [2024-12-06 17:53:44.181960] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:4064 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.377 [2024-12-06 17:53:44.181978] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:26 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.377 [2024-12-06 17:53:44.194916] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.377 [2024-12-06 17:53:44.194946] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:12795 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.377 [2024-12-06 17:53:44.194963] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.377 [2024-12-06 17:53:44.208701] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.377 [2024-12-06 17:53:44.208729] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:102 nsid:1 lba:11893 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.377 [2024-12-06 17:53:44.208746] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:102 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.635 [2024-12-06 17:53:44.223060] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.635 [2024-12-06 17:53:44.223089] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:9925 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.635 [2024-12-06 17:53:44.223105] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:70 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.635 [2024-12-06 17:53:44.237083] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.635 [2024-12-06 17:53:44.237113] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:85 nsid:1 lba:2794 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.635 [2024-12-06 17:53:44.237135] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:85 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.635 [2024-12-06 17:53:44.248403] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.635 [2024-12-06 17:53:44.248447] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:7233 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.635 [2024-12-06 17:53:44.248463] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:28 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.635 [2024-12-06 17:53:44.261735] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.635 [2024-12-06 17:53:44.261778] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:21638 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.635 [2024-12-06 17:53:44.261795] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.635 [2024-12-06 17:53:44.275671] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.635 [2024-12-06 17:53:44.275700] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:9720 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.636 [2024-12-06 17:53:44.275733] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:56 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.636 [2024-12-06 17:53:44.289287] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.636 [2024-12-06 17:53:44.289318] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:16509 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.636 [2024-12-06 17:53:44.289335] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:111 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.636 [2024-12-06 17:53:44.302086] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.636 [2024-12-06 17:53:44.302117] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:3736 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.636 [2024-12-06 17:53:44.302134] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:88 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.636 [2024-12-06 17:53:44.314859] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.636 [2024-12-06 17:53:44.314892] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:9254 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.636 [2024-12-06 17:53:44.314909] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.636 [2024-12-06 17:53:44.328996] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.636 [2024-12-06 17:53:44.329025] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:4009 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.636 [2024-12-06 17:53:44.329041] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.636 [2024-12-06 17:53:44.342485] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.636 [2024-12-06 17:53:44.342514] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:86 nsid:1 lba:4642 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.636 [2024-12-06 17:53:44.342530] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:86 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.636 [2024-12-06 17:53:44.356846] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.636 [2024-12-06 17:53:44.356884] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:24631 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.636 [2024-12-06 17:53:44.356902] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:35 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.636 [2024-12-06 17:53:44.368898] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.636 [2024-12-06 17:53:44.368930] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:19095 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.636 [2024-12-06 17:53:44.368947] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:45 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.636 [2024-12-06 17:53:44.383249] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.636 [2024-12-06 17:53:44.383278] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:7814 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.636 [2024-12-06 17:53:44.383294] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:17 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.636 [2024-12-06 17:53:44.396324] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.636 [2024-12-06 17:53:44.396367] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:2644 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.636 [2024-12-06 17:53:44.396382] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:45 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.636 [2024-12-06 17:53:44.408355] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.636 [2024-12-06 17:53:44.408383] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:110 nsid:1 lba:23744 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.636 [2024-12-06 17:53:44.408398] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.636 [2024-12-06 17:53:44.422232] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.636 [2024-12-06 17:53:44.422260] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:12555 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.636 [2024-12-06 17:53:44.422276] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:100 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.636 [2024-12-06 17:53:44.434788] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.636 [2024-12-06 17:53:44.434819] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:6315 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.636 [2024-12-06 17:53:44.434836] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:61 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.636 [2024-12-06 17:53:44.449468] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.636 [2024-12-06 17:53:44.449496] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:6779 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.636 [2024-12-06 17:53:44.449512] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:52 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.636 [2024-12-06 17:53:44.460746] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.636 [2024-12-06 17:53:44.460775] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:125 nsid:1 lba:17996 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.636 [2024-12-06 17:53:44.460791] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:125 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.895 [2024-12-06 17:53:44.476442] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.895 [2024-12-06 17:53:44.476475] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:8868 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.895 [2024-12-06 17:53:44.476493] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:121 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.895 [2024-12-06 17:53:44.491681] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.895 [2024-12-06 17:53:44.491738] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:3494 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.895 [2024-12-06 17:53:44.491756] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.895 [2024-12-06 17:53:44.504385] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.895 [2024-12-06 17:53:44.504415] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:22770 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.895 [2024-12-06 17:53:44.504432] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:84 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.895 [2024-12-06 17:53:44.517678] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.895 [2024-12-06 17:53:44.517710] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:5983 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.895 [2024-12-06 17:53:44.517727] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:37 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.895 [2024-12-06 17:53:44.530756] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.895 [2024-12-06 17:53:44.530788] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:11890 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.895 [2024-12-06 17:53:44.530805] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:49 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.895 [2024-12-06 17:53:44.543603] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.895 [2024-12-06 17:53:44.543634] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:10357 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.895 [2024-12-06 17:53:44.543675] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:18 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.895 [2024-12-06 17:53:44.556725] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.896 [2024-12-06 17:53:44.556772] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:73 nsid:1 lba:4865 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.896 [2024-12-06 17:53:44.556790] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:73 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.896 [2024-12-06 17:53:44.571607] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.896 [2024-12-06 17:53:44.571640] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:24814 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.896 [2024-12-06 17:53:44.571658] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:44 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.896 [2024-12-06 17:53:44.584609] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.896 [2024-12-06 17:53:44.584655] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:22905 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.896 [2024-12-06 17:53:44.584692] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.896 [2024-12-06 17:53:44.596952] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.896 [2024-12-06 17:53:44.596997] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:7557 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.896 [2024-12-06 17:53:44.597014] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:17 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.896 [2024-12-06 17:53:44.609840] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.896 [2024-12-06 17:53:44.609872] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:125 nsid:1 lba:22241 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.896 [2024-12-06 17:53:44.609890] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:125 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.896 [2024-12-06 17:53:44.622821] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.896 [2024-12-06 17:53:44.622854] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:98 nsid:1 lba:11358 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.896 [2024-12-06 17:53:44.622873] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:98 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.896 [2024-12-06 17:53:44.635330] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.896 [2024-12-06 17:53:44.635361] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:8359 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.896 [2024-12-06 17:53:44.635377] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:105 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.896 [2024-12-06 17:53:44.648332] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.896 [2024-12-06 17:53:44.648361] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:2228 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.896 [2024-12-06 17:53:44.648377] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:103 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.896 [2024-12-06 17:53:44.661272] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.896 [2024-12-06 17:53:44.661303] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:11711 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.896 [2024-12-06 17:53:44.661320] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:105 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.896 [2024-12-06 17:53:44.674398] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.896 [2024-12-06 17:53:44.674427] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:23944 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.896 [2024-12-06 17:53:44.674444] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:34 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.896 [2024-12-06 17:53:44.687030] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.896 [2024-12-06 17:53:44.687075] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:125 nsid:1 lba:23264 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.896 [2024-12-06 17:53:44.687092] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:125 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.896 [2024-12-06 17:53:44.700296] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.896 [2024-12-06 17:53:44.700333] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:5967 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.896 [2024-12-06 17:53:44.700350] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:95 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.896 [2024-12-06 17:53:44.714138] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.896 [2024-12-06 17:53:44.714166] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:125 nsid:1 lba:8625 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.896 [2024-12-06 17:53:44.714183] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:125 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:02.896 [2024-12-06 17:53:44.728035] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:02.896 [2024-12-06 17:53:44.728064] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:172 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:02.896 [2024-12-06 17:53:44.728080] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:80 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:03.154 [2024-12-06 17:53:44.741992] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:03.154 [2024-12-06 17:53:44.742024] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:124 nsid:1 lba:17141 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:03.154 [2024-12-06 17:53:44.742042] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:124 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:03.154 [2024-12-06 17:53:44.754337] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:03.154 [2024-12-06 17:53:44.754365] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:17940 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:03.154 [2024-12-06 17:53:44.754381] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:90 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:03.154 [2024-12-06 17:53:44.767500] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:03.154 [2024-12-06 17:53:44.767531] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:22465 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:03.154 [2024-12-06 17:53:44.767547] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:94 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:03.154 [2024-12-06 17:53:44.780409] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:03.154 [2024-12-06 17:53:44.780438] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:104 nsid:1 lba:16517 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:03.155 [2024-12-06 17:53:44.780453] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:104 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:03.155 [2024-12-06 17:53:44.798102] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:03.155 [2024-12-06 17:53:44.798147] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:2471 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:03.155 [2024-12-06 17:53:44.798164] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:03.155 [2024-12-06 17:53:44.809819] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:03.155 [2024-12-06 17:53:44.809849] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:15153 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:03.155 [2024-12-06 17:53:44.809872] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:120 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:03.155 18825.00 IOPS, 73.54 MiB/s [2024-12-06T16:53:44.994Z] [2024-12-06 17:53:44.822634] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1081a90) 00:38:03.155 [2024-12-06 17:53:44.822669] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:23018 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:03.155 [2024-12-06 17:53:44.822703] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:14 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:38:03.155 00:38:03.155 Latency(us) 00:38:03.155 [2024-12-06T16:53:44.994Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:38:03.155 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 128, IO size: 4096) 00:38:03.155 nvme0n1 : 2.05 18460.91 72.11 0.00 0.00 6789.93 3470.98 47380.10 00:38:03.155 [2024-12-06T16:53:44.994Z] =================================================================================================================== 00:38:03.155 [2024-12-06T16:53:44.994Z] Total : 18460.91 72.11 0.00 0.00 6789.93 3470.98 47380.10 00:38:03.155 { 00:38:03.155 "results": [ 00:38:03.155 { 00:38:03.155 "job": "nvme0n1", 00:38:03.155 "core_mask": "0x2", 00:38:03.155 "workload": "randread", 00:38:03.155 "status": "finished", 00:38:03.155 "queue_depth": 128, 00:38:03.155 "io_size": 4096, 00:38:03.155 "runtime": 2.046378, 00:38:03.155 "iops": 18460.909958961638, 00:38:03.155 "mibps": 72.1129295271939, 00:38:03.155 "io_failed": 0, 00:38:03.155 "io_timeout": 0, 00:38:03.155 "avg_latency_us": 6789.928814693247, 00:38:03.155 "min_latency_us": 3470.9807407407407, 00:38:03.155 "max_latency_us": 47380.10074074074 00:38:03.155 } 00:38:03.155 ], 00:38:03.155 "core_count": 1 00:38:03.155 } 00:38:03.155 17:53:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # get_transient_errcount nvme0n1 00:38:03.155 17:53:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@27 -- # bperf_rpc bdev_get_iostat -b nvme0n1 00:38:03.155 17:53:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_get_iostat -b nvme0n1 00:38:03.155 17:53:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@28 -- # jq -r '.bdevs[0] 00:38:03.155 | .driver_specific 00:38:03.155 | .nvme_error 00:38:03.155 | .status_code 00:38:03.155 | .command_transient_transport_error' 00:38:03.416 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # (( 148 > 0 )) 00:38:03.416 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@73 -- # killprocess 406299 00:38:03.416 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # '[' -z 406299 ']' 00:38:03.416 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@958 -- # kill -0 406299 00:38:03.416 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@959 -- # uname 00:38:03.416 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:38:03.416 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 406299 00:38:03.416 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:38:03.416 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:38:03.416 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@972 -- # echo 'killing process with pid 406299' 00:38:03.416 killing process with pid 406299 00:38:03.416 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@973 -- # kill 406299 00:38:03.416 Received shutdown signal, test time was about 2.000000 seconds 00:38:03.416 00:38:03.416 Latency(us) 00:38:03.416 [2024-12-06T16:53:45.255Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:38:03.416 [2024-12-06T16:53:45.255Z] =================================================================================================================== 00:38:03.416 [2024-12-06T16:53:45.255Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:38:03.416 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@978 -- # wait 406299 00:38:03.673 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@109 -- # run_bperf_err randread 131072 16 00:38:03.673 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@54 -- # local rw bs qd 00:38:03.673 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # rw=randread 00:38:03.673 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # bs=131072 00:38:03.673 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # qd=16 00:38:03.673 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@58 -- # bperfpid=406706 00:38:03.673 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randread -o 131072 -t 2 -q 16 -z 00:38:03.673 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@60 -- # waitforlisten 406706 /var/tmp/bperf.sock 00:38:03.673 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@835 -- # '[' -z 406706 ']' 00:38:03.673 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bperf.sock 00:38:03.673 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@840 -- # local max_retries=100 00:38:03.673 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:38:03.673 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:38:03.673 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@844 -- # xtrace_disable 00:38:03.673 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:38:03.673 [2024-12-06 17:53:45.438760] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:38:03.673 [2024-12-06 17:53:45.438845] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid406706 ] 00:38:03.673 I/O size of 131072 is greater than zero copy threshold (65536). 00:38:03.673 Zero copy mechanism will not be used. 00:38:03.673 [2024-12-06 17:53:45.504922] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:38:03.930 [2024-12-06 17:53:45.550496] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:38:03.930 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:38:03.930 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@868 -- # return 0 00:38:03.930 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@61 -- # bperf_rpc bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:38:03.930 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:38:04.187 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@63 -- # rpc_cmd accel_error_inject_error -o crc32c -t disable 00:38:04.187 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:04.187 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:38:04.187 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:04.187 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@64 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:38:04.187 17:53:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:38:04.752 nvme0n1 00:38:04.752 17:53:46 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@67 -- # rpc_cmd accel_error_inject_error -o crc32c -t corrupt -i 32 00:38:04.752 17:53:46 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:04.752 17:53:46 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:38:04.752 17:53:46 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:04.752 17:53:46 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@69 -- # bperf_py perform_tests 00:38:04.752 17:53:46 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:38:04.752 I/O size of 131072 is greater than zero copy threshold (65536). 00:38:04.752 Zero copy mechanism will not be used. 00:38:04.752 Running I/O for 2 seconds... 00:38:04.752 [2024-12-06 17:53:46.540553] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:04.752 [2024-12-06 17:53:46.540626] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:25280 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:04.752 [2024-12-06 17:53:46.540647] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:04.752 [2024-12-06 17:53:46.546652] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:04.752 [2024-12-06 17:53:46.546699] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:20480 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:04.752 [2024-12-06 17:53:46.546726] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:04.752 [2024-12-06 17:53:46.552635] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:04.752 [2024-12-06 17:53:46.552678] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:04.752 [2024-12-06 17:53:46.552698] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:04.752 [2024-12-06 17:53:46.558427] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:04.752 [2024-12-06 17:53:46.558462] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:11648 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:04.752 [2024-12-06 17:53:46.558482] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:04.752 [2024-12-06 17:53:46.563759] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:04.752 [2024-12-06 17:53:46.563793] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16160 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:04.752 [2024-12-06 17:53:46.563812] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:04.752 [2024-12-06 17:53:46.568939] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:04.752 [2024-12-06 17:53:46.568990] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:04.752 [2024-12-06 17:53:46.569008] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:04.752 [2024-12-06 17:53:46.574163] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:04.752 [2024-12-06 17:53:46.574208] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:3040 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:04.752 [2024-12-06 17:53:46.574242] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:04.752 [2024-12-06 17:53:46.580508] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:04.752 [2024-12-06 17:53:46.580543] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:15488 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:04.752 [2024-12-06 17:53:46.580562] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:04.752 [2024-12-06 17:53:46.588151] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:04.752 [2024-12-06 17:53:46.588186] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:8192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:04.752 [2024-12-06 17:53:46.588204] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.010 [2024-12-06 17:53:46.595965] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.010 [2024-12-06 17:53:46.596015] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25504 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.010 [2024-12-06 17:53:46.596033] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.010 [2024-12-06 17:53:46.603672] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.010 [2024-12-06 17:53:46.603706] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:14624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.010 [2024-12-06 17:53:46.603725] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.010 [2024-12-06 17:53:46.611332] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.010 [2024-12-06 17:53:46.611365] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:17856 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.010 [2024-12-06 17:53:46.611383] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.010 [2024-12-06 17:53:46.619310] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.010 [2024-12-06 17:53:46.619344] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:18624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.010 [2024-12-06 17:53:46.619377] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.010 [2024-12-06 17:53:46.627027] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.010 [2024-12-06 17:53:46.627062] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:24608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.010 [2024-12-06 17:53:46.627081] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.010 [2024-12-06 17:53:46.634817] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.010 [2024-12-06 17:53:46.634851] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:11008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.010 [2024-12-06 17:53:46.634870] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:14 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.010 [2024-12-06 17:53:46.642626] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.010 [2024-12-06 17:53:46.642660] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24864 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.010 [2024-12-06 17:53:46.642687] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.010 [2024-12-06 17:53:46.650281] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.010 [2024-12-06 17:53:46.650331] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.010 [2024-12-06 17:53:46.650348] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.010 [2024-12-06 17:53:46.657988] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.010 [2024-12-06 17:53:46.658024] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:18144 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.010 [2024-12-06 17:53:46.658061] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.010 [2024-12-06 17:53:46.665786] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.011 [2024-12-06 17:53:46.665821] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:21440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.011 [2024-12-06 17:53:46.665840] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.011 [2024-12-06 17:53:46.673460] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.011 [2024-12-06 17:53:46.673494] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:4224 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.011 [2024-12-06 17:53:46.673513] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.011 [2024-12-06 17:53:46.681120] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.011 [2024-12-06 17:53:46.681153] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:1120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.011 [2024-12-06 17:53:46.681185] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.011 [2024-12-06 17:53:46.688853] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.011 [2024-12-06 17:53:46.688888] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:12544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.011 [2024-12-06 17:53:46.688907] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.011 [2024-12-06 17:53:46.696692] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.011 [2024-12-06 17:53:46.696730] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:20512 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.011 [2024-12-06 17:53:46.696763] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.011 [2024-12-06 17:53:46.703242] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.011 [2024-12-06 17:53:46.703276] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:20512 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.011 [2024-12-06 17:53:46.703300] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.011 [2024-12-06 17:53:46.709292] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.011 [2024-12-06 17:53:46.709327] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:4256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.011 [2024-12-06 17:53:46.709346] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.011 [2024-12-06 17:53:46.715727] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.011 [2024-12-06 17:53:46.715761] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:18784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.011 [2024-12-06 17:53:46.715780] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.011 [2024-12-06 17:53:46.721585] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.011 [2024-12-06 17:53:46.721619] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:9792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.011 [2024-12-06 17:53:46.721638] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.011 [2024-12-06 17:53:46.727777] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.011 [2024-12-06 17:53:46.727811] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:3456 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.011 [2024-12-06 17:53:46.727844] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.011 [2024-12-06 17:53:46.734587] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.011 [2024-12-06 17:53:46.734622] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:16896 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.011 [2024-12-06 17:53:46.734641] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.011 [2024-12-06 17:53:46.741016] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.011 [2024-12-06 17:53:46.741050] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:23744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.011 [2024-12-06 17:53:46.741084] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:14 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.011 [2024-12-06 17:53:46.746774] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.011 [2024-12-06 17:53:46.746809] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:12992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.011 [2024-12-06 17:53:46.746828] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.011 [2024-12-06 17:53:46.752038] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.011 [2024-12-06 17:53:46.752071] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:8384 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.011 [2024-12-06 17:53:46.752090] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.011 [2024-12-06 17:53:46.758086] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.011 [2024-12-06 17:53:46.758121] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:1088 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.011 [2024-12-06 17:53:46.758154] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.011 [2024-12-06 17:53:46.764252] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.011 [2024-12-06 17:53:46.764286] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:21088 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.011 [2024-12-06 17:53:46.764306] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.011 [2024-12-06 17:53:46.768720] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.011 [2024-12-06 17:53:46.768751] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:7200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.011 [2024-12-06 17:53:46.768768] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.011 [2024-12-06 17:53:46.776370] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.011 [2024-12-06 17:53:46.776403] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:22496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.011 [2024-12-06 17:53:46.776422] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.011 [2024-12-06 17:53:46.782530] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.011 [2024-12-06 17:53:46.782564] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:9376 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.011 [2024-12-06 17:53:46.782582] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.011 [2024-12-06 17:53:46.788846] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.011 [2024-12-06 17:53:46.788881] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:19200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.011 [2024-12-06 17:53:46.788899] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.011 [2024-12-06 17:53:46.794949] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.011 [2024-12-06 17:53:46.795015] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:2720 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.011 [2024-12-06 17:53:46.795046] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.011 [2024-12-06 17:53:46.800357] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.011 [2024-12-06 17:53:46.800393] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:21184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.011 [2024-12-06 17:53:46.800412] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.012 [2024-12-06 17:53:46.806093] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.012 [2024-12-06 17:53:46.806128] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25280 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.012 [2024-12-06 17:53:46.806154] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.012 [2024-12-06 17:53:46.813058] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.012 [2024-12-06 17:53:46.813106] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:15744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.012 [2024-12-06 17:53:46.813123] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.012 [2024-12-06 17:53:46.820425] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.012 [2024-12-06 17:53:46.820461] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.012 [2024-12-06 17:53:46.820479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.012 [2024-12-06 17:53:46.826792] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.012 [2024-12-06 17:53:46.826826] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.012 [2024-12-06 17:53:46.826844] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.012 [2024-12-06 17:53:46.832995] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.012 [2024-12-06 17:53:46.833029] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:14368 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.012 [2024-12-06 17:53:46.833046] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.012 [2024-12-06 17:53:46.839388] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.012 [2024-12-06 17:53:46.839433] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:2816 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.012 [2024-12-06 17:53:46.839450] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.012 [2024-12-06 17:53:46.845616] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.012 [2024-12-06 17:53:46.845651] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:18176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.012 [2024-12-06 17:53:46.845680] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.270 [2024-12-06 17:53:46.851552] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.270 [2024-12-06 17:53:46.851587] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:3424 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.270 [2024-12-06 17:53:46.851606] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.270 [2024-12-06 17:53:46.857492] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.270 [2024-12-06 17:53:46.857526] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:22496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.270 [2024-12-06 17:53:46.857547] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.270 [2024-12-06 17:53:46.864404] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.270 [2024-12-06 17:53:46.864444] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24160 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.270 [2024-12-06 17:53:46.864478] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.270 [2024-12-06 17:53:46.871570] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.270 [2024-12-06 17:53:46.871605] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:17344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.270 [2024-12-06 17:53:46.871624] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.270 [2024-12-06 17:53:46.878725] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.270 [2024-12-06 17:53:46.878759] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:13120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.270 [2024-12-06 17:53:46.878778] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.270 [2024-12-06 17:53:46.885225] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.270 [2024-12-06 17:53:46.885259] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:13248 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.270 [2024-12-06 17:53:46.885278] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.270 [2024-12-06 17:53:46.889422] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.270 [2024-12-06 17:53:46.889483] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:5280 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.270 [2024-12-06 17:53:46.889519] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.270 [2024-12-06 17:53:46.894345] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.270 [2024-12-06 17:53:46.894391] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:21344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.270 [2024-12-06 17:53:46.894409] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.270 [2024-12-06 17:53:46.899705] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.270 [2024-12-06 17:53:46.899749] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:14432 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.270 [2024-12-06 17:53:46.899772] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.270 [2024-12-06 17:53:46.905312] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.270 [2024-12-06 17:53:46.905346] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:11008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:46.905372] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:46.910611] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:46.910647] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:6272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:46.910686] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:46.915832] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:46.915867] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:10368 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:46.915887] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:46.921071] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:46.921119] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:23360 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:46.921137] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:46.927074] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:46.927107] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:46.927126] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:46.932354] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:46.932388] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:7904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:46.932422] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:46.937499] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:46.937534] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:22432 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:46.937552] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:46.942563] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:46.942597] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:4928 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:46.942631] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:46.947733] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:46.947767] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:21376 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:46.947785] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:46.952910] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:46.952944] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:8992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:46.952962] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:46.958078] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:46.958112] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:6848 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:46.958154] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:46.963255] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:46.963287] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:17088 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:46.963312] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:46.968447] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:46.968480] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:4512 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:46.968515] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:46.973761] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:46.973795] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:46.973825] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:46.979148] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:46.979196] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:17088 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:46.979238] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:46.984463] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:46.984496] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:6496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:46.984524] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:46.989812] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:46.989861] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:23776 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:46.989890] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:46.995027] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:46.995060] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:480 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:46.995078] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:47.000278] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:47.000309] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:8640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:47.000329] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:47.005560] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:47.005612] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:3360 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:47.005633] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:47.010855] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:47.010903] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:15072 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:47.010922] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:47.015979] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:47.016012] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:23648 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:47.016046] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:47.021090] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:47.021124] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:9312 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:47.021167] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:47.026319] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:47.026366] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:10784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:47.026385] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:47.031528] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:47.031575] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:1760 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:47.031594] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:47.036873] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:47.036908] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:5216 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:47.036927] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:47.042141] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:47.042174] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:23840 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:47.042192] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:47.047573] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:47.047620] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:4672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:47.047637] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:47.053493] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:47.053528] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:13920 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:47.053546] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:47.060880] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:47.060914] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:21120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:47.060933] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:47.068607] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:47.068674] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:3136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:47.068694] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:47.076413] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:47.076461] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:18752 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:47.076479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:47.084033] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:47.084066] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:12480 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:47.084099] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:47.091807] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:47.091842] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:7648 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:47.091860] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:47.099485] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:47.099518] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:3136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:47.099550] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.271 [2024-12-06 17:53:47.107273] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.271 [2024-12-06 17:53:47.107307] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:10592 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.271 [2024-12-06 17:53:47.107325] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.530 [2024-12-06 17:53:47.115130] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.530 [2024-12-06 17:53:47.115162] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:24448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.530 [2024-12-06 17:53:47.115200] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.530 [2024-12-06 17:53:47.122901] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.530 [2024-12-06 17:53:47.122953] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:9088 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.530 [2024-12-06 17:53:47.122972] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.530 [2024-12-06 17:53:47.130701] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.530 [2024-12-06 17:53:47.130732] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:16800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.530 [2024-12-06 17:53:47.130749] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.530 [2024-12-06 17:53:47.138288] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.530 [2024-12-06 17:53:47.138320] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:5664 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.530 [2024-12-06 17:53:47.138359] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.530 [2024-12-06 17:53:47.145956] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.530 [2024-12-06 17:53:47.146003] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:3616 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.530 [2024-12-06 17:53:47.146020] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.530 [2024-12-06 17:53:47.153693] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.530 [2024-12-06 17:53:47.153726] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.530 [2024-12-06 17:53:47.153744] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.530 [2024-12-06 17:53:47.161318] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.530 [2024-12-06 17:53:47.161353] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:12992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.530 [2024-12-06 17:53:47.161377] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.530 [2024-12-06 17:53:47.169072] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.530 [2024-12-06 17:53:47.169118] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:20960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.530 [2024-12-06 17:53:47.169138] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.530 [2024-12-06 17:53:47.175980] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.530 [2024-12-06 17:53:47.176011] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:11744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.530 [2024-12-06 17:53:47.176031] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.530 [2024-12-06 17:53:47.181785] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.530 [2024-12-06 17:53:47.181819] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:21216 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.530 [2024-12-06 17:53:47.181852] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.530 [2024-12-06 17:53:47.187961] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.530 [2024-12-06 17:53:47.188009] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:12320 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.530 [2024-12-06 17:53:47.188028] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.530 [2024-12-06 17:53:47.194332] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.530 [2024-12-06 17:53:47.194381] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:5216 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.530 [2024-12-06 17:53:47.194418] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.531 [2024-12-06 17:53:47.200707] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.531 [2024-12-06 17:53:47.200740] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:6784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.531 [2024-12-06 17:53:47.200759] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.531 [2024-12-06 17:53:47.206315] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.531 [2024-12-06 17:53:47.206364] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:22688 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.531 [2024-12-06 17:53:47.206388] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.531 [2024-12-06 17:53:47.212107] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.531 [2024-12-06 17:53:47.212139] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:1056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.531 [2024-12-06 17:53:47.212157] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.531 [2024-12-06 17:53:47.217294] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.531 [2024-12-06 17:53:47.217328] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:6016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.531 [2024-12-06 17:53:47.217362] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.531 [2024-12-06 17:53:47.222451] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.531 [2024-12-06 17:53:47.222500] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.531 [2024-12-06 17:53:47.222519] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:14 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.531 [2024-12-06 17:53:47.227658] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.531 [2024-12-06 17:53:47.227700] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:3104 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.531 [2024-12-06 17:53:47.227741] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.531 [2024-12-06 17:53:47.232792] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.531 [2024-12-06 17:53:47.232826] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:11072 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.531 [2024-12-06 17:53:47.232844] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.531 [2024-12-06 17:53:47.238050] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.531 [2024-12-06 17:53:47.238083] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:20288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.531 [2024-12-06 17:53:47.238101] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.531 [2024-12-06 17:53:47.243129] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.531 [2024-12-06 17:53:47.243177] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:20800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.531 [2024-12-06 17:53:47.243194] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.531 [2024-12-06 17:53:47.248366] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.531 [2024-12-06 17:53:47.248399] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:5280 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.531 [2024-12-06 17:53:47.248427] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.531 [2024-12-06 17:53:47.253457] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.531 [2024-12-06 17:53:47.253505] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:1984 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.531 [2024-12-06 17:53:47.253522] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.531 [2024-12-06 17:53:47.258656] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.531 [2024-12-06 17:53:47.258711] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:7584 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.531 [2024-12-06 17:53:47.258729] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.531 [2024-12-06 17:53:47.263982] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.531 [2024-12-06 17:53:47.264029] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:13408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.531 [2024-12-06 17:53:47.264047] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:14 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.531 [2024-12-06 17:53:47.269143] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.531 [2024-12-06 17:53:47.269176] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:16064 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.531 [2024-12-06 17:53:47.269200] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.531 [2024-12-06 17:53:47.274309] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.531 [2024-12-06 17:53:47.274349] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:4736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.531 [2024-12-06 17:53:47.274367] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.531 [2024-12-06 17:53:47.279672] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.531 [2024-12-06 17:53:47.279720] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:14944 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.531 [2024-12-06 17:53:47.279744] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.531 [2024-12-06 17:53:47.284961] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.531 [2024-12-06 17:53:47.285011] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:18912 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.531 [2024-12-06 17:53:47.285030] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.531 [2024-12-06 17:53:47.290423] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.531 [2024-12-06 17:53:47.290455] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:19904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.531 [2024-12-06 17:53:47.290474] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.531 [2024-12-06 17:53:47.296456] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.531 [2024-12-06 17:53:47.296495] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:10592 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.531 [2024-12-06 17:53:47.296514] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.531 [2024-12-06 17:53:47.302139] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.531 [2024-12-06 17:53:47.302173] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:24512 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.531 [2024-12-06 17:53:47.302192] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.531 [2024-12-06 17:53:47.308180] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.531 [2024-12-06 17:53:47.308214] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:18272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.531 [2024-12-06 17:53:47.308239] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.531 [2024-12-06 17:53:47.314297] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.531 [2024-12-06 17:53:47.314331] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:3680 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.531 [2024-12-06 17:53:47.314350] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:14 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.531 [2024-12-06 17:53:47.320341] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.531 [2024-12-06 17:53:47.320374] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:16672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.531 [2024-12-06 17:53:47.320392] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.531 [2024-12-06 17:53:47.326604] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.531 [2024-12-06 17:53:47.326638] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:2400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.531 [2024-12-06 17:53:47.326656] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.531 [2024-12-06 17:53:47.332710] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.531 [2024-12-06 17:53:47.332744] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:16416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.531 [2024-12-06 17:53:47.332763] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.531 [2024-12-06 17:53:47.338764] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.531 [2024-12-06 17:53:47.338798] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:18336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.531 [2024-12-06 17:53:47.338822] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.531 [2024-12-06 17:53:47.342303] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.531 [2024-12-06 17:53:47.342350] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:6176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.532 [2024-12-06 17:53:47.342367] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.532 [2024-12-06 17:53:47.348591] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.532 [2024-12-06 17:53:47.348622] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:15712 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.532 [2024-12-06 17:53:47.348639] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.532 [2024-12-06 17:53:47.354829] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.532 [2024-12-06 17:53:47.354861] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:12704 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.532 [2024-12-06 17:53:47.354879] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.532 [2024-12-06 17:53:47.360746] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.532 [2024-12-06 17:53:47.360777] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:18016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.532 [2024-12-06 17:53:47.360796] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.791 [2024-12-06 17:53:47.367082] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.791 [2024-12-06 17:53:47.367116] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:22272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.791 [2024-12-06 17:53:47.367134] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.791 [2024-12-06 17:53:47.372442] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.791 [2024-12-06 17:53:47.372475] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:15264 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.791 [2024-12-06 17:53:47.372506] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.791 [2024-12-06 17:53:47.377678] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.791 [2024-12-06 17:53:47.377712] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:16192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.791 [2024-12-06 17:53:47.377731] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.791 [2024-12-06 17:53:47.382918] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.791 [2024-12-06 17:53:47.382952] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:4512 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.791 [2024-12-06 17:53:47.382985] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.791 [2024-12-06 17:53:47.388098] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.791 [2024-12-06 17:53:47.388144] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:7136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.791 [2024-12-06 17:53:47.388161] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.791 [2024-12-06 17:53:47.393203] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.791 [2024-12-06 17:53:47.393250] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:5664 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.791 [2024-12-06 17:53:47.393267] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.791 [2024-12-06 17:53:47.398371] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.791 [2024-12-06 17:53:47.398403] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:6816 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.791 [2024-12-06 17:53:47.398419] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.791 [2024-12-06 17:53:47.403480] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.791 [2024-12-06 17:53:47.403514] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:7744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.791 [2024-12-06 17:53:47.403532] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.791 [2024-12-06 17:53:47.408603] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.791 [2024-12-06 17:53:47.408636] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:10112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.791 [2024-12-06 17:53:47.408662] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.791 [2024-12-06 17:53:47.413813] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.791 [2024-12-06 17:53:47.413862] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:23744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.791 [2024-12-06 17:53:47.413881] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.791 [2024-12-06 17:53:47.418942] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.791 [2024-12-06 17:53:47.418981] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:3808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.791 [2024-12-06 17:53:47.419001] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.791 [2024-12-06 17:53:47.424003] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.792 [2024-12-06 17:53:47.424037] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:3904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.792 [2024-12-06 17:53:47.424075] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.792 [2024-12-06 17:53:47.429257] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.792 [2024-12-06 17:53:47.429289] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:8160 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.792 [2024-12-06 17:53:47.429308] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.792 [2024-12-06 17:53:47.434321] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.792 [2024-12-06 17:53:47.434353] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:1376 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.792 [2024-12-06 17:53:47.434371] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.792 [2024-12-06 17:53:47.439470] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.792 [2024-12-06 17:53:47.439502] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:13760 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.792 [2024-12-06 17:53:47.439519] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.792 [2024-12-06 17:53:47.444343] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.792 [2024-12-06 17:53:47.444377] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:21920 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.792 [2024-12-06 17:53:47.444395] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.792 [2024-12-06 17:53:47.449386] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.792 [2024-12-06 17:53:47.449419] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:18912 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.792 [2024-12-06 17:53:47.449436] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.792 [2024-12-06 17:53:47.454608] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.792 [2024-12-06 17:53:47.454641] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:3072 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.792 [2024-12-06 17:53:47.454684] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.792 [2024-12-06 17:53:47.459792] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.792 [2024-12-06 17:53:47.459826] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:24192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.792 [2024-12-06 17:53:47.459852] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.792 [2024-12-06 17:53:47.464888] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.792 [2024-12-06 17:53:47.464921] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:1024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.792 [2024-12-06 17:53:47.464941] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.792 [2024-12-06 17:53:47.470098] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.792 [2024-12-06 17:53:47.470132] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:14272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.792 [2024-12-06 17:53:47.470157] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.792 [2024-12-06 17:53:47.475391] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.792 [2024-12-06 17:53:47.475425] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:4768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.792 [2024-12-06 17:53:47.475444] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.792 [2024-12-06 17:53:47.480585] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.792 [2024-12-06 17:53:47.480617] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:7840 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.792 [2024-12-06 17:53:47.480636] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.792 [2024-12-06 17:53:47.485841] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.792 [2024-12-06 17:53:47.485875] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:18336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.792 [2024-12-06 17:53:47.485893] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.792 [2024-12-06 17:53:47.491112] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.792 [2024-12-06 17:53:47.491143] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:4608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.792 [2024-12-06 17:53:47.491163] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.792 [2024-12-06 17:53:47.496493] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.792 [2024-12-06 17:53:47.496526] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:2336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.792 [2024-12-06 17:53:47.496544] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.792 [2024-12-06 17:53:47.501510] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.792 [2024-12-06 17:53:47.501542] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.792 [2024-12-06 17:53:47.501560] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.792 [2024-12-06 17:53:47.506543] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.792 [2024-12-06 17:53:47.506581] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:22400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.792 [2024-12-06 17:53:47.506600] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.792 [2024-12-06 17:53:47.511827] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.792 [2024-12-06 17:53:47.511861] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:19200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.792 [2024-12-06 17:53:47.511880] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.792 [2024-12-06 17:53:47.516945] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.792 [2024-12-06 17:53:47.516993] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:10592 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.792 [2024-12-06 17:53:47.517011] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.792 [2024-12-06 17:53:47.522132] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.792 [2024-12-06 17:53:47.522164] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:21664 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.792 [2024-12-06 17:53:47.522182] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.792 [2024-12-06 17:53:47.527332] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.792 [2024-12-06 17:53:47.527364] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:23648 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.792 [2024-12-06 17:53:47.527382] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.792 [2024-12-06 17:53:47.532412] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.792 [2024-12-06 17:53:47.532444] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:4928 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.792 [2024-12-06 17:53:47.532475] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.792 5164.00 IOPS, 645.50 MiB/s [2024-12-06T16:53:47.631Z] [2024-12-06 17:53:47.539318] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.792 [2024-12-06 17:53:47.539352] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:7040 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.792 [2024-12-06 17:53:47.539384] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.793 [2024-12-06 17:53:47.544478] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.793 [2024-12-06 17:53:47.544512] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:13184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.793 [2024-12-06 17:53:47.544530] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.793 [2024-12-06 17:53:47.549811] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.793 [2024-12-06 17:53:47.549843] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:17696 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.793 [2024-12-06 17:53:47.549862] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.793 [2024-12-06 17:53:47.555440] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.793 [2024-12-06 17:53:47.555471] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:13600 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.793 [2024-12-06 17:53:47.555488] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.793 [2024-12-06 17:53:47.560924] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.793 [2024-12-06 17:53:47.560973] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:4384 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.793 [2024-12-06 17:53:47.560990] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.793 [2024-12-06 17:53:47.566940] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.793 [2024-12-06 17:53:47.566987] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:7040 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.793 [2024-12-06 17:53:47.567005] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.793 [2024-12-06 17:53:47.572707] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.793 [2024-12-06 17:53:47.572742] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:6880 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.793 [2024-12-06 17:53:47.572782] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.793 [2024-12-06 17:53:47.578498] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.793 [2024-12-06 17:53:47.578531] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.793 [2024-12-06 17:53:47.578548] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.793 [2024-12-06 17:53:47.584862] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.793 [2024-12-06 17:53:47.584894] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:13952 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.793 [2024-12-06 17:53:47.584912] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.793 [2024-12-06 17:53:47.590433] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.793 [2024-12-06 17:53:47.590464] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:0 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.793 [2024-12-06 17:53:47.590482] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.793 [2024-12-06 17:53:47.595511] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.793 [2024-12-06 17:53:47.595544] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:24672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.793 [2024-12-06 17:53:47.595563] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.793 [2024-12-06 17:53:47.601005] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.793 [2024-12-06 17:53:47.601038] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:15712 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.793 [2024-12-06 17:53:47.601065] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.793 [2024-12-06 17:53:47.606115] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.793 [2024-12-06 17:53:47.606148] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:19840 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.793 [2024-12-06 17:53:47.606168] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:05.793 [2024-12-06 17:53:47.611325] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.793 [2024-12-06 17:53:47.611358] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:10880 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.793 [2024-12-06 17:53:47.611383] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:05.793 [2024-12-06 17:53:47.616417] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.793 [2024-12-06 17:53:47.616465] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:8864 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.793 [2024-12-06 17:53:47.616488] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:05.793 [2024-12-06 17:53:47.621504] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.793 [2024-12-06 17:53:47.621536] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:19168 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.793 [2024-12-06 17:53:47.621553] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:05.793 [2024-12-06 17:53:47.626541] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:05.793 [2024-12-06 17:53:47.626572] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:6656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:05.793 [2024-12-06 17:53:47.626591] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.052 [2024-12-06 17:53:47.631797] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.052 [2024-12-06 17:53:47.631831] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:8800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.052 [2024-12-06 17:53:47.631849] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.052 [2024-12-06 17:53:47.637268] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.052 [2024-12-06 17:53:47.637301] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:10400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.052 [2024-12-06 17:53:47.637319] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.052 [2024-12-06 17:53:47.643046] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.052 [2024-12-06 17:53:47.643078] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:2976 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.052 [2024-12-06 17:53:47.643097] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.052 [2024-12-06 17:53:47.648202] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.052 [2024-12-06 17:53:47.648237] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:3552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.052 [2024-12-06 17:53:47.648254] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.052 [2024-12-06 17:53:47.653596] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.052 [2024-12-06 17:53:47.653626] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:9536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.052 [2024-12-06 17:53:47.653644] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.052 [2024-12-06 17:53:47.658808] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.052 [2024-12-06 17:53:47.658839] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:21344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.052 [2024-12-06 17:53:47.658857] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.052 [2024-12-06 17:53:47.664087] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.052 [2024-12-06 17:53:47.664120] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:2112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.052 [2024-12-06 17:53:47.664137] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.052 [2024-12-06 17:53:47.669280] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.052 [2024-12-06 17:53:47.669312] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:9504 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.052 [2024-12-06 17:53:47.669338] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.052 [2024-12-06 17:53:47.674555] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.052 [2024-12-06 17:53:47.674590] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:8768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.052 [2024-12-06 17:53:47.674616] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.052 [2024-12-06 17:53:47.679765] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.053 [2024-12-06 17:53:47.679799] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:20928 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.053 [2024-12-06 17:53:47.679818] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.053 [2024-12-06 17:53:47.684903] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.053 [2024-12-06 17:53:47.684937] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:6944 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.053 [2024-12-06 17:53:47.684975] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.053 [2024-12-06 17:53:47.690147] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.053 [2024-12-06 17:53:47.690179] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:5536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.053 [2024-12-06 17:53:47.690201] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.053 [2024-12-06 17:53:47.695259] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.053 [2024-12-06 17:53:47.695291] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:3872 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.053 [2024-12-06 17:53:47.695308] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.053 [2024-12-06 17:53:47.700310] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.053 [2024-12-06 17:53:47.700343] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:2176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.053 [2024-12-06 17:53:47.700366] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.053 [2024-12-06 17:53:47.705456] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.053 [2024-12-06 17:53:47.705486] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:24768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.053 [2024-12-06 17:53:47.705503] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.053 [2024-12-06 17:53:47.710624] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.053 [2024-12-06 17:53:47.710689] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:4992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.053 [2024-12-06 17:53:47.710710] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.053 [2024-12-06 17:53:47.715976] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.053 [2024-12-06 17:53:47.716023] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:14112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.053 [2024-12-06 17:53:47.716040] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.053 [2024-12-06 17:53:47.721384] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.053 [2024-12-06 17:53:47.721414] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:3808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.053 [2024-12-06 17:53:47.721436] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.053 [2024-12-06 17:53:47.726767] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.053 [2024-12-06 17:53:47.726800] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:11136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.053 [2024-12-06 17:53:47.726819] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.053 [2024-12-06 17:53:47.731945] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.053 [2024-12-06 17:53:47.731979] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:14496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.053 [2024-12-06 17:53:47.732012] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.053 [2024-12-06 17:53:47.737087] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.053 [2024-12-06 17:53:47.737123] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:13248 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.053 [2024-12-06 17:53:47.737142] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.053 [2024-12-06 17:53:47.742222] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.053 [2024-12-06 17:53:47.742253] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:19872 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.053 [2024-12-06 17:53:47.742271] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.053 [2024-12-06 17:53:47.747388] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.053 [2024-12-06 17:53:47.747434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:7136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.053 [2024-12-06 17:53:47.747454] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.053 [2024-12-06 17:53:47.752784] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.053 [2024-12-06 17:53:47.752818] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:15552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.053 [2024-12-06 17:53:47.752837] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.053 [2024-12-06 17:53:47.758064] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.053 [2024-12-06 17:53:47.758095] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:10336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.053 [2024-12-06 17:53:47.758113] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.053 [2024-12-06 17:53:47.763259] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.053 [2024-12-06 17:53:47.763289] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:6080 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.053 [2024-12-06 17:53:47.763306] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.053 [2024-12-06 17:53:47.768576] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.053 [2024-12-06 17:53:47.768607] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:11328 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.053 [2024-12-06 17:53:47.768624] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.053 [2024-12-06 17:53:47.773970] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.053 [2024-12-06 17:53:47.774002] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:512 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.053 [2024-12-06 17:53:47.774035] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.053 [2024-12-06 17:53:47.779034] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.053 [2024-12-06 17:53:47.779081] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:22176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.053 [2024-12-06 17:53:47.779098] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.053 [2024-12-06 17:53:47.784206] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.053 [2024-12-06 17:53:47.784254] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:2592 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.053 [2024-12-06 17:53:47.784273] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.053 [2024-12-06 17:53:47.789337] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.053 [2024-12-06 17:53:47.789370] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:20768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.053 [2024-12-06 17:53:47.789402] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.053 [2024-12-06 17:53:47.794262] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.053 [2024-12-06 17:53:47.794294] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:1056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.053 [2024-12-06 17:53:47.794325] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.053 [2024-12-06 17:53:47.799811] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.053 [2024-12-06 17:53:47.799843] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:19904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.053 [2024-12-06 17:53:47.799866] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.053 [2024-12-06 17:53:47.805749] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.053 [2024-12-06 17:53:47.805783] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:21440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.053 [2024-12-06 17:53:47.805801] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.053 [2024-12-06 17:53:47.811714] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.053 [2024-12-06 17:53:47.811762] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:4288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.053 [2024-12-06 17:53:47.811781] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.053 [2024-12-06 17:53:47.818956] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.053 [2024-12-06 17:53:47.818987] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:19808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.053 [2024-12-06 17:53:47.819020] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.054 [2024-12-06 17:53:47.826400] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.054 [2024-12-06 17:53:47.826434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:1824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.054 [2024-12-06 17:53:47.826466] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.054 [2024-12-06 17:53:47.834121] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.054 [2024-12-06 17:53:47.834169] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:3456 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.054 [2024-12-06 17:53:47.834193] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.054 [2024-12-06 17:53:47.841698] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.054 [2024-12-06 17:53:47.841742] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:13280 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.054 [2024-12-06 17:53:47.841760] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.054 [2024-12-06 17:53:47.849340] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.054 [2024-12-06 17:53:47.849371] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:10624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.054 [2024-12-06 17:53:47.849388] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.054 [2024-12-06 17:53:47.856999] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.054 [2024-12-06 17:53:47.857030] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:14976 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.054 [2024-12-06 17:53:47.857047] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.054 [2024-12-06 17:53:47.864724] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.054 [2024-12-06 17:53:47.864769] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:15968 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.054 [2024-12-06 17:53:47.864787] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.054 [2024-12-06 17:53:47.872477] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.054 [2024-12-06 17:53:47.872525] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:17600 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.054 [2024-12-06 17:53:47.872543] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.054 [2024-12-06 17:53:47.880139] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.054 [2024-12-06 17:53:47.880172] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:14944 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.054 [2024-12-06 17:53:47.880189] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.054 [2024-12-06 17:53:47.887928] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.054 [2024-12-06 17:53:47.887961] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:24800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.054 [2024-12-06 17:53:47.887997] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.312 [2024-12-06 17:53:47.895728] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.312 [2024-12-06 17:53:47.895762] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:20288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.312 [2024-12-06 17:53:47.895780] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.312 [2024-12-06 17:53:47.903524] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.312 [2024-12-06 17:53:47.903562] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:1952 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.312 [2024-12-06 17:53:47.903581] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.312 [2024-12-06 17:53:47.911257] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.312 [2024-12-06 17:53:47.911305] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:1024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.312 [2024-12-06 17:53:47.911323] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.312 [2024-12-06 17:53:47.919038] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.312 [2024-12-06 17:53:47.919086] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:5632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.312 [2024-12-06 17:53:47.919103] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.312 [2024-12-06 17:53:47.926731] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.312 [2024-12-06 17:53:47.926766] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:23328 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.312 [2024-12-06 17:53:47.926784] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.312 [2024-12-06 17:53:47.933474] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.312 [2024-12-06 17:53:47.933527] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.312 [2024-12-06 17:53:47.933544] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.312 [2024-12-06 17:53:47.939121] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.313 [2024-12-06 17:53:47.939154] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:4192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.313 [2024-12-06 17:53:47.939187] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.313 [2024-12-06 17:53:47.945371] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.313 [2024-12-06 17:53:47.945406] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:11104 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.313 [2024-12-06 17:53:47.945439] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.313 [2024-12-06 17:53:47.951837] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.313 [2024-12-06 17:53:47.951872] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:9952 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.313 [2024-12-06 17:53:47.951891] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.313 [2024-12-06 17:53:47.958236] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.313 [2024-12-06 17:53:47.958270] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:16448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.313 [2024-12-06 17:53:47.958289] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.313 [2024-12-06 17:53:47.965044] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.313 [2024-12-06 17:53:47.965078] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:6976 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.313 [2024-12-06 17:53:47.965111] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.313 [2024-12-06 17:53:47.970502] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.313 [2024-12-06 17:53:47.970549] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:8256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.313 [2024-12-06 17:53:47.970574] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.313 [2024-12-06 17:53:47.976469] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.313 [2024-12-06 17:53:47.976503] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:13696 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.313 [2024-12-06 17:53:47.976521] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.313 [2024-12-06 17:53:47.982311] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.313 [2024-12-06 17:53:47.982344] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:5248 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.313 [2024-12-06 17:53:47.982377] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.313 [2024-12-06 17:53:47.988814] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.313 [2024-12-06 17:53:47.988849] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:15072 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.313 [2024-12-06 17:53:47.988867] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.313 [2024-12-06 17:53:47.994700] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.313 [2024-12-06 17:53:47.994745] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:18944 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.313 [2024-12-06 17:53:47.994763] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.313 [2024-12-06 17:53:48.000805] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.313 [2024-12-06 17:53:48.000839] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:2048 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.313 [2024-12-06 17:53:48.000857] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.313 [2024-12-06 17:53:48.006956] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.313 [2024-12-06 17:53:48.006990] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:22720 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.313 [2024-12-06 17:53:48.007010] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.313 [2024-12-06 17:53:48.013391] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.313 [2024-12-06 17:53:48.013425] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.313 [2024-12-06 17:53:48.013463] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.313 [2024-12-06 17:53:48.020828] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.313 [2024-12-06 17:53:48.020863] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.313 [2024-12-06 17:53:48.020882] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.313 [2024-12-06 17:53:48.027122] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.313 [2024-12-06 17:53:48.027162] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:10528 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.313 [2024-12-06 17:53:48.027182] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.313 [2024-12-06 17:53:48.034392] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.313 [2024-12-06 17:53:48.034427] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:3936 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.313 [2024-12-06 17:53:48.034460] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.313 [2024-12-06 17:53:48.042299] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.313 [2024-12-06 17:53:48.042347] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:8544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.313 [2024-12-06 17:53:48.042365] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.313 [2024-12-06 17:53:48.050254] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.313 [2024-12-06 17:53:48.050288] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:6272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.313 [2024-12-06 17:53:48.050321] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.313 [2024-12-06 17:53:48.058083] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.313 [2024-12-06 17:53:48.058132] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:2528 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.313 [2024-12-06 17:53:48.058150] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.313 [2024-12-06 17:53:48.066271] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.313 [2024-12-06 17:53:48.066306] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:1216 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.313 [2024-12-06 17:53:48.066324] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.313 [2024-12-06 17:53:48.074460] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.313 [2024-12-06 17:53:48.074509] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:2624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.313 [2024-12-06 17:53:48.074527] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.313 [2024-12-06 17:53:48.082282] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.313 [2024-12-06 17:53:48.082337] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:20992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.314 [2024-12-06 17:53:48.082355] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.314 [2024-12-06 17:53:48.090018] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.314 [2024-12-06 17:53:48.090053] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:352 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.314 [2024-12-06 17:53:48.090087] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.314 [2024-12-06 17:53:48.097748] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.314 [2024-12-06 17:53:48.097782] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:14240 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.314 [2024-12-06 17:53:48.097800] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.314 [2024-12-06 17:53:48.105442] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.314 [2024-12-06 17:53:48.105476] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:2912 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.314 [2024-12-06 17:53:48.105510] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.314 [2024-12-06 17:53:48.113140] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.314 [2024-12-06 17:53:48.113176] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:20000 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.314 [2024-12-06 17:53:48.113194] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.314 [2024-12-06 17:53:48.120880] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.314 [2024-12-06 17:53:48.120927] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:10304 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.314 [2024-12-06 17:53:48.120944] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.314 [2024-12-06 17:53:48.128634] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.314 [2024-12-06 17:53:48.128674] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:15392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.314 [2024-12-06 17:53:48.128694] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.314 [2024-12-06 17:53:48.136311] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.314 [2024-12-06 17:53:48.136345] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:12512 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.314 [2024-12-06 17:53:48.136378] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.314 [2024-12-06 17:53:48.144076] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.314 [2024-12-06 17:53:48.144124] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:21632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.314 [2024-12-06 17:53:48.144150] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.572 [2024-12-06 17:53:48.150916] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.572 [2024-12-06 17:53:48.150965] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:13760 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.572 [2024-12-06 17:53:48.150983] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.572 [2024-12-06 17:53:48.156811] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.572 [2024-12-06 17:53:48.156845] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:19616 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.572 [2024-12-06 17:53:48.156880] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.572 [2024-12-06 17:53:48.163206] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.572 [2024-12-06 17:53:48.163239] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:4224 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.572 [2024-12-06 17:53:48.163272] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.572 [2024-12-06 17:53:48.168908] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.572 [2024-12-06 17:53:48.168942] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16000 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.572 [2024-12-06 17:53:48.168977] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.572 [2024-12-06 17:53:48.174236] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.572 [2024-12-06 17:53:48.174268] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:4768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.572 [2024-12-06 17:53:48.174286] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.572 [2024-12-06 17:53:48.179441] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.573 [2024-12-06 17:53:48.179475] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:8672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.573 [2024-12-06 17:53:48.179493] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.573 [2024-12-06 17:53:48.184603] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.573 [2024-12-06 17:53:48.184637] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:15936 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.573 [2024-12-06 17:53:48.184655] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.573 [2024-12-06 17:53:48.189787] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.573 [2024-12-06 17:53:48.189836] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:4992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.573 [2024-12-06 17:53:48.189855] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.573 [2024-12-06 17:53:48.195128] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.573 [2024-12-06 17:53:48.195182] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.573 [2024-12-06 17:53:48.195202] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.573 [2024-12-06 17:53:48.201030] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.573 [2024-12-06 17:53:48.201065] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:24096 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.573 [2024-12-06 17:53:48.201084] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.573 [2024-12-06 17:53:48.206481] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.573 [2024-12-06 17:53:48.206514] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:5152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.573 [2024-12-06 17:53:48.206531] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.573 [2024-12-06 17:53:48.211816] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.573 [2024-12-06 17:53:48.211850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:6400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.573 [2024-12-06 17:53:48.211883] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.573 [2024-12-06 17:53:48.217215] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.573 [2024-12-06 17:53:48.217263] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:6176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.573 [2024-12-06 17:53:48.217280] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.573 [2024-12-06 17:53:48.222687] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.573 [2024-12-06 17:53:48.222727] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:1472 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.573 [2024-12-06 17:53:48.222745] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.573 [2024-12-06 17:53:48.228804] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.573 [2024-12-06 17:53:48.228852] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:11776 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.573 [2024-12-06 17:53:48.228869] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.573 [2024-12-06 17:53:48.233970] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.573 [2024-12-06 17:53:48.234004] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:15968 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.573 [2024-12-06 17:53:48.234041] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.573 [2024-12-06 17:53:48.239534] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.573 [2024-12-06 17:53:48.239582] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:3296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.573 [2024-12-06 17:53:48.239600] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.573 [2024-12-06 17:53:48.244889] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.573 [2024-12-06 17:53:48.244924] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:4928 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.573 [2024-12-06 17:53:48.244956] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.573 [2024-12-06 17:53:48.250044] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.573 [2024-12-06 17:53:48.250078] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:24992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.573 [2024-12-06 17:53:48.250097] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.573 [2024-12-06 17:53:48.255134] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.573 [2024-12-06 17:53:48.255166] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:8512 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.573 [2024-12-06 17:53:48.255183] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.573 [2024-12-06 17:53:48.260323] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.573 [2024-12-06 17:53:48.260356] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:9408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.573 [2024-12-06 17:53:48.260375] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.573 [2024-12-06 17:53:48.265367] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.573 [2024-12-06 17:53:48.265401] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:22304 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.573 [2024-12-06 17:53:48.265419] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.573 [2024-12-06 17:53:48.270609] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.573 [2024-12-06 17:53:48.270644] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:18272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.573 [2024-12-06 17:53:48.270662] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.573 [2024-12-06 17:53:48.275727] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.573 [2024-12-06 17:53:48.275761] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:4320 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.573 [2024-12-06 17:53:48.275799] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.573 [2024-12-06 17:53:48.281020] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.573 [2024-12-06 17:53:48.281068] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:8896 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.573 [2024-12-06 17:53:48.281086] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.573 [2024-12-06 17:53:48.286170] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.573 [2024-12-06 17:53:48.286202] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:5088 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.573 [2024-12-06 17:53:48.286224] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.573 [2024-12-06 17:53:48.291480] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.574 [2024-12-06 17:53:48.291514] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:18656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.574 [2024-12-06 17:53:48.291532] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.574 [2024-12-06 17:53:48.296826] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.574 [2024-12-06 17:53:48.296875] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:2752 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.574 [2024-12-06 17:53:48.296893] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.574 [2024-12-06 17:53:48.301974] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.574 [2024-12-06 17:53:48.302024] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:19392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.574 [2024-12-06 17:53:48.302042] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.574 [2024-12-06 17:53:48.307202] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.574 [2024-12-06 17:53:48.307237] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24000 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.574 [2024-12-06 17:53:48.307255] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.574 [2024-12-06 17:53:48.312288] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.574 [2024-12-06 17:53:48.312321] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:18976 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.574 [2024-12-06 17:53:48.312339] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.574 [2024-12-06 17:53:48.317359] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.574 [2024-12-06 17:53:48.317408] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24128 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.574 [2024-12-06 17:53:48.317426] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.574 [2024-12-06 17:53:48.322382] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.574 [2024-12-06 17:53:48.322416] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.574 [2024-12-06 17:53:48.322435] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.574 [2024-12-06 17:53:48.327635] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.574 [2024-12-06 17:53:48.327677] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:15168 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.574 [2024-12-06 17:53:48.327698] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.574 [2024-12-06 17:53:48.332750] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.574 [2024-12-06 17:53:48.332789] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:19744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.574 [2024-12-06 17:53:48.332823] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.574 [2024-12-06 17:53:48.338021] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.574 [2024-12-06 17:53:48.338054] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:10464 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.574 [2024-12-06 17:53:48.338087] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.574 [2024-12-06 17:53:48.343342] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.574 [2024-12-06 17:53:48.343389] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:4224 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.574 [2024-12-06 17:53:48.343406] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.574 [2024-12-06 17:53:48.348697] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.574 [2024-12-06 17:53:48.348729] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:20736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.574 [2024-12-06 17:53:48.348747] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.574 [2024-12-06 17:53:48.353958] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.574 [2024-12-06 17:53:48.353992] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:15776 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.574 [2024-12-06 17:53:48.354025] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.574 [2024-12-06 17:53:48.359245] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.574 [2024-12-06 17:53:48.359278] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:8736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.574 [2024-12-06 17:53:48.359297] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.574 [2024-12-06 17:53:48.364507] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.574 [2024-12-06 17:53:48.364555] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:15424 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.574 [2024-12-06 17:53:48.364573] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.574 [2024-12-06 17:53:48.369593] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.574 [2024-12-06 17:53:48.369628] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:11520 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.574 [2024-12-06 17:53:48.369647] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.574 [2024-12-06 17:53:48.374747] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.574 [2024-12-06 17:53:48.374782] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:2496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.574 [2024-12-06 17:53:48.374801] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.574 [2024-12-06 17:53:48.378330] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.574 [2024-12-06 17:53:48.378368] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:3872 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.574 [2024-12-06 17:53:48.378388] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.574 [2024-12-06 17:53:48.383816] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.574 [2024-12-06 17:53:48.383850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:12992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.574 [2024-12-06 17:53:48.383868] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.574 [2024-12-06 17:53:48.389286] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.574 [2024-12-06 17:53:48.389332] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:16416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.574 [2024-12-06 17:53:48.389349] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.574 [2024-12-06 17:53:48.394718] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.574 [2024-12-06 17:53:48.394751] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:14688 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.574 [2024-12-06 17:53:48.394768] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.574 [2024-12-06 17:53:48.400806] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.575 [2024-12-06 17:53:48.400839] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:16352 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.575 [2024-12-06 17:53:48.400856] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.575 [2024-12-06 17:53:48.406818] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.575 [2024-12-06 17:53:48.406850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:7936 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.575 [2024-12-06 17:53:48.406868] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.833 [2024-12-06 17:53:48.413123] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.833 [2024-12-06 17:53:48.413153] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:14496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.833 [2024-12-06 17:53:48.413170] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.833 [2024-12-06 17:53:48.418957] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.833 [2024-12-06 17:53:48.419007] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:13920 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.833 [2024-12-06 17:53:48.419025] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.833 [2024-12-06 17:53:48.424933] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.833 [2024-12-06 17:53:48.424982] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:21824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.834 [2024-12-06 17:53:48.425005] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.834 [2024-12-06 17:53:48.431618] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.834 [2024-12-06 17:53:48.431653] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16864 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.834 [2024-12-06 17:53:48.431681] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.834 [2024-12-06 17:53:48.438642] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.834 [2024-12-06 17:53:48.438682] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:23296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.834 [2024-12-06 17:53:48.438703] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.834 [2024-12-06 17:53:48.444223] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.834 [2024-12-06 17:53:48.444255] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:5344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.834 [2024-12-06 17:53:48.444288] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.834 [2024-12-06 17:53:48.449561] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.834 [2024-12-06 17:53:48.449594] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:5920 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.834 [2024-12-06 17:53:48.449614] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:14 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.834 [2024-12-06 17:53:48.454809] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.834 [2024-12-06 17:53:48.454844] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:5728 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.834 [2024-12-06 17:53:48.454862] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.834 [2024-12-06 17:53:48.459932] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.834 [2024-12-06 17:53:48.459969] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:14752 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.834 [2024-12-06 17:53:48.459995] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.834 [2024-12-06 17:53:48.463357] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.834 [2024-12-06 17:53:48.463391] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:6848 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.834 [2024-12-06 17:53:48.463411] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.834 [2024-12-06 17:53:48.468340] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.834 [2024-12-06 17:53:48.468374] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:21152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.834 [2024-12-06 17:53:48.468392] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.834 [2024-12-06 17:53:48.473223] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.834 [2024-12-06 17:53:48.473258] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:1408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.834 [2024-12-06 17:53:48.473278] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.834 [2024-12-06 17:53:48.476634] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.834 [2024-12-06 17:53:48.476690] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:9440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.834 [2024-12-06 17:53:48.476711] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.834 [2024-12-06 17:53:48.481765] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.834 [2024-12-06 17:53:48.481802] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:12224 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.834 [2024-12-06 17:53:48.481825] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.834 [2024-12-06 17:53:48.487132] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.834 [2024-12-06 17:53:48.487166] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:25440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.834 [2024-12-06 17:53:48.487184] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.834 [2024-12-06 17:53:48.492224] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.834 [2024-12-06 17:53:48.492272] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:7680 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.834 [2024-12-06 17:53:48.492290] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.834 [2024-12-06 17:53:48.497365] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.834 [2024-12-06 17:53:48.497413] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:15200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.834 [2024-12-06 17:53:48.497430] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.834 [2024-12-06 17:53:48.502525] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.834 [2024-12-06 17:53:48.502558] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:4608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.834 [2024-12-06 17:53:48.502576] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.834 [2024-12-06 17:53:48.507783] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.834 [2024-12-06 17:53:48.507815] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:5792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.834 [2024-12-06 17:53:48.507833] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.834 [2024-12-06 17:53:48.513023] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.834 [2024-12-06 17:53:48.513057] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:3296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.834 [2024-12-06 17:53:48.513080] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.834 [2024-12-06 17:53:48.518238] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.834 [2024-12-06 17:53:48.518285] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:4448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.834 [2024-12-06 17:53:48.518302] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.834 [2024-12-06 17:53:48.523484] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.834 [2024-12-06 17:53:48.523522] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:16736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.834 [2024-12-06 17:53:48.523541] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.834 [2024-12-06 17:53:48.527156] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.834 [2024-12-06 17:53:48.527213] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:4096 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.834 [2024-12-06 17:53:48.527238] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:06.834 [2024-12-06 17:53:48.531548] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.834 [2024-12-06 17:53:48.531581] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:20768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.834 [2024-12-06 17:53:48.531598] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:06.834 [2024-12-06 17:53:48.536766] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.834 [2024-12-06 17:53:48.536798] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:6496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.834 [2024-12-06 17:53:48.536815] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:06.834 5239.50 IOPS, 654.94 MiB/s [2024-12-06T16:53:48.673Z] [2024-12-06 17:53:48.543296] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x14b3de0) 00:38:06.834 [2024-12-06 17:53:48.543342] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:12224 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:06.834 [2024-12-06 17:53:48.543360] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:06.834 00:38:06.834 Latency(us) 00:38:06.834 [2024-12-06T16:53:48.673Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:38:06.834 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 16, IO size: 131072) 00:38:06.834 nvme0n1 : 2.00 5242.72 655.34 0.00 0.00 3046.38 831.34 9757.58 00:38:06.834 [2024-12-06T16:53:48.673Z] =================================================================================================================== 00:38:06.834 [2024-12-06T16:53:48.673Z] Total : 5242.72 655.34 0.00 0.00 3046.38 831.34 9757.58 00:38:06.834 { 00:38:06.834 "results": [ 00:38:06.834 { 00:38:06.834 "job": "nvme0n1", 00:38:06.834 "core_mask": "0x2", 00:38:06.834 "workload": "randread", 00:38:06.834 "status": "finished", 00:38:06.834 "queue_depth": 16, 00:38:06.834 "io_size": 131072, 00:38:06.834 "runtime": 2.004686, 00:38:06.834 "iops": 5242.71631567238, 00:38:06.834 "mibps": 655.3395394590475, 00:38:06.834 "io_failed": 0, 00:38:06.834 "io_timeout": 0, 00:38:06.834 "avg_latency_us": 3046.3847358071675, 00:38:06.835 "min_latency_us": 831.3362962962963, 00:38:06.835 "max_latency_us": 9757.582222222221 00:38:06.835 } 00:38:06.835 ], 00:38:06.835 "core_count": 1 00:38:06.835 } 00:38:06.835 17:53:48 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # get_transient_errcount nvme0n1 00:38:06.835 17:53:48 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@27 -- # bperf_rpc bdev_get_iostat -b nvme0n1 00:38:06.835 17:53:48 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@28 -- # jq -r '.bdevs[0] 00:38:06.835 | .driver_specific 00:38:06.835 | .nvme_error 00:38:06.835 | .status_code 00:38:06.835 | .command_transient_transport_error' 00:38:06.835 17:53:48 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_get_iostat -b nvme0n1 00:38:07.092 17:53:48 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # (( 340 > 0 )) 00:38:07.092 17:53:48 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@73 -- # killprocess 406706 00:38:07.092 17:53:48 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # '[' -z 406706 ']' 00:38:07.092 17:53:48 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@958 -- # kill -0 406706 00:38:07.092 17:53:48 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@959 -- # uname 00:38:07.092 17:53:48 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:38:07.092 17:53:48 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 406706 00:38:07.092 17:53:48 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:38:07.092 17:53:48 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:38:07.092 17:53:48 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@972 -- # echo 'killing process with pid 406706' 00:38:07.092 killing process with pid 406706 00:38:07.092 17:53:48 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@973 -- # kill 406706 00:38:07.092 Received shutdown signal, test time was about 2.000000 seconds 00:38:07.092 00:38:07.092 Latency(us) 00:38:07.092 [2024-12-06T16:53:48.931Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:38:07.092 [2024-12-06T16:53:48.931Z] =================================================================================================================== 00:38:07.092 [2024-12-06T16:53:48.931Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:38:07.092 17:53:48 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@978 -- # wait 406706 00:38:07.349 17:53:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@114 -- # run_bperf_err randwrite 4096 128 00:38:07.349 17:53:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@54 -- # local rw bs qd 00:38:07.349 17:53:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # rw=randwrite 00:38:07.349 17:53:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # bs=4096 00:38:07.349 17:53:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # qd=128 00:38:07.349 17:53:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@58 -- # bperfpid=407116 00:38:07.349 17:53:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@60 -- # waitforlisten 407116 /var/tmp/bperf.sock 00:38:07.349 17:53:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@835 -- # '[' -z 407116 ']' 00:38:07.349 17:53:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bperf.sock 00:38:07.349 17:53:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randwrite -o 4096 -t 2 -q 128 -z 00:38:07.349 17:53:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@840 -- # local max_retries=100 00:38:07.349 17:53:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:38:07.349 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:38:07.349 17:53:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@844 -- # xtrace_disable 00:38:07.349 17:53:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:38:07.349 [2024-12-06 17:53:49.088477] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:38:07.349 [2024-12-06 17:53:49.088550] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid407116 ] 00:38:07.349 [2024-12-06 17:53:49.159406] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:38:07.607 [2024-12-06 17:53:49.211515] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:38:07.607 17:53:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:38:07.607 17:53:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@868 -- # return 0 00:38:07.607 17:53:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@61 -- # bperf_rpc bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:38:07.607 17:53:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:38:07.865 17:53:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@63 -- # rpc_cmd accel_error_inject_error -o crc32c -t disable 00:38:07.865 17:53:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:07.865 17:53:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:38:07.865 17:53:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:07.865 17:53:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@64 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:38:07.865 17:53:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:38:08.428 nvme0n1 00:38:08.428 17:53:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@67 -- # rpc_cmd accel_error_inject_error -o crc32c -t corrupt -i 256 00:38:08.428 17:53:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:08.428 17:53:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:38:08.428 17:53:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:08.428 17:53:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@69 -- # bperf_py perform_tests 00:38:08.428 17:53:50 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:38:08.428 Running I/O for 2 seconds... 00:38:08.428 [2024-12-06 17:53:50.199738] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ef6020 00:38:08.428 [2024-12-06 17:53:50.200824] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:7542 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.428 [2024-12-06 17:53:50.200864] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:36 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:38:08.428 [2024-12-06 17:53:50.214041] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee49b0 00:38:08.428 [2024-12-06 17:53:50.215716] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:120 nsid:1 lba:14683 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.429 [2024-12-06 17:53:50.215770] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:120 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:38:08.429 [2024-12-06 17:53:50.222413] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016edfdc0 00:38:08.429 [2024-12-06 17:53:50.223179] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:6100 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.429 [2024-12-06 17:53:50.223223] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:122 cdw0:0 sqhd:0008 p:0 m:0 dnr:0 00:38:08.429 [2024-12-06 17:53:50.236746] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ef1430 00:38:08.429 [2024-12-06 17:53:50.238157] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:22700 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.429 [2024-12-06 17:53:50.238203] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:55 cdw0:0 sqhd:004a p:0 m:0 dnr:0 00:38:08.429 [2024-12-06 17:53:50.247641] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016efc560 00:38:08.429 [2024-12-06 17:53:50.248840] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:17444 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.429 [2024-12-06 17:53:50.248871] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:104 cdw0:0 sqhd:006a p:0 m:0 dnr:0 00:38:08.429 [2024-12-06 17:53:50.259492] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ef3e60 00:38:08.429 [2024-12-06 17:53:50.260691] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:613 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.429 [2024-12-06 17:53:50.260721] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:43 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:38:08.685 [2024-12-06 17:53:50.272327] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ef6cc8 00:38:08.685 [2024-12-06 17:53:50.273463] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:9416 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.685 [2024-12-06 17:53:50.273509] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:24 cdw0:0 sqhd:003d p:0 m:0 dnr:0 00:38:08.685 [2024-12-06 17:53:50.286328] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016eea248 00:38:08.685 [2024-12-06 17:53:50.288128] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:11675 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.685 [2024-12-06 17:53:50.288158] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:67 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:38:08.685 [2024-12-06 17:53:50.294739] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016eeaab8 00:38:08.685 [2024-12-06 17:53:50.295590] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:5751 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.685 [2024-12-06 17:53:50.295620] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:66 cdw0:0 sqhd:0011 p:0 m:0 dnr:0 00:38:08.685 [2024-12-06 17:53:50.308386] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.685 [2024-12-06 17:53:50.308699] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:2027 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.685 [2024-12-06 17:53:50.308731] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.685 [2024-12-06 17:53:50.322814] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.685 [2024-12-06 17:53:50.323157] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:2146 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.685 [2024-12-06 17:53:50.323201] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.685 [2024-12-06 17:53:50.337078] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.685 [2024-12-06 17:53:50.337432] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:6952 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.685 [2024-12-06 17:53:50.337475] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.685 [2024-12-06 17:53:50.351416] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.685 [2024-12-06 17:53:50.351733] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:18282 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.685 [2024-12-06 17:53:50.351763] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.685 [2024-12-06 17:53:50.365680] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.685 [2024-12-06 17:53:50.365976] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:8720 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.685 [2024-12-06 17:53:50.366019] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.685 [2024-12-06 17:53:50.379891] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.685 [2024-12-06 17:53:50.380252] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:7250 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.685 [2024-12-06 17:53:50.380281] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.685 [2024-12-06 17:53:50.394242] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.685 [2024-12-06 17:53:50.394596] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:12275 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.685 [2024-12-06 17:53:50.394640] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.685 [2024-12-06 17:53:50.408294] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.685 [2024-12-06 17:53:50.408580] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:18664 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.685 [2024-12-06 17:53:50.408608] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.685 [2024-12-06 17:53:50.422639] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.685 [2024-12-06 17:53:50.422985] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:7018 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.685 [2024-12-06 17:53:50.423014] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.685 [2024-12-06 17:53:50.436989] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.685 [2024-12-06 17:53:50.437338] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:13371 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.685 [2024-12-06 17:53:50.437366] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.685 [2024-12-06 17:53:50.451503] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.685 [2024-12-06 17:53:50.451790] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:11272 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.685 [2024-12-06 17:53:50.451823] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.685 [2024-12-06 17:53:50.465396] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.685 [2024-12-06 17:53:50.465718] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:20936 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.685 [2024-12-06 17:53:50.465747] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.685 [2024-12-06 17:53:50.479506] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.685 [2024-12-06 17:53:50.479812] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:11008 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.685 [2024-12-06 17:53:50.479841] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.685 [2024-12-06 17:53:50.493811] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.685 [2024-12-06 17:53:50.494146] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:8110 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.685 [2024-12-06 17:53:50.494175] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.685 [2024-12-06 17:53:50.508148] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.685 [2024-12-06 17:53:50.508497] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:15075 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.685 [2024-12-06 17:53:50.508525] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.685 [2024-12-06 17:53:50.522239] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.685 [2024-12-06 17:53:50.522555] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:7126 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.685 [2024-12-06 17:53:50.522583] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.941 [2024-12-06 17:53:50.536126] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.941 [2024-12-06 17:53:50.536406] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:20038 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.941 [2024-12-06 17:53:50.536434] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.941 [2024-12-06 17:53:50.550177] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.941 [2024-12-06 17:53:50.550501] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:8236 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.941 [2024-12-06 17:53:50.550529] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.941 [2024-12-06 17:53:50.564327] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.941 [2024-12-06 17:53:50.564676] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:16757 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.941 [2024-12-06 17:53:50.564709] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.941 [2024-12-06 17:53:50.578605] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.941 [2024-12-06 17:53:50.578943] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:9673 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.941 [2024-12-06 17:53:50.578972] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.941 [2024-12-06 17:53:50.592958] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.941 [2024-12-06 17:53:50.593261] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:24354 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.941 [2024-12-06 17:53:50.593305] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.941 [2024-12-06 17:53:50.607062] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.941 [2024-12-06 17:53:50.607358] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:18214 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.941 [2024-12-06 17:53:50.607402] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.941 [2024-12-06 17:53:50.621330] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.941 [2024-12-06 17:53:50.621662] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:2073 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.941 [2024-12-06 17:53:50.621700] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.941 [2024-12-06 17:53:50.635522] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.941 [2024-12-06 17:53:50.635864] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:6476 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.941 [2024-12-06 17:53:50.635892] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.941 [2024-12-06 17:53:50.649846] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.941 [2024-12-06 17:53:50.650148] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:2538 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.941 [2024-12-06 17:53:50.650190] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.941 [2024-12-06 17:53:50.664118] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.941 [2024-12-06 17:53:50.664435] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:22528 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.941 [2024-12-06 17:53:50.664478] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.941 [2024-12-06 17:53:50.678369] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.941 [2024-12-06 17:53:50.678693] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:9373 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.941 [2024-12-06 17:53:50.678721] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.941 [2024-12-06 17:53:50.692534] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.941 [2024-12-06 17:53:50.692871] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:13560 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.941 [2024-12-06 17:53:50.692901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.941 [2024-12-06 17:53:50.706750] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.941 [2024-12-06 17:53:50.707069] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:867 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.941 [2024-12-06 17:53:50.707097] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.941 [2024-12-06 17:53:50.720679] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.941 [2024-12-06 17:53:50.721048] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:4617 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.941 [2024-12-06 17:53:50.721094] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.941 [2024-12-06 17:53:50.734905] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.941 [2024-12-06 17:53:50.735227] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:2112 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.941 [2024-12-06 17:53:50.735256] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.941 [2024-12-06 17:53:50.748963] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.941 [2024-12-06 17:53:50.749261] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:3670 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.941 [2024-12-06 17:53:50.749288] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.941 [2024-12-06 17:53:50.763138] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.941 [2024-12-06 17:53:50.763457] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:10945 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.941 [2024-12-06 17:53:50.763500] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:08.941 [2024-12-06 17:53:50.777399] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:08.941 [2024-12-06 17:53:50.777644] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:1667 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:08.941 [2024-12-06 17:53:50.777680] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.199 [2024-12-06 17:53:50.791191] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.199 [2024-12-06 17:53:50.791530] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:14765 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.199 [2024-12-06 17:53:50.791572] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.199 [2024-12-06 17:53:50.805412] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.199 [2024-12-06 17:53:50.805721] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:2017 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.199 [2024-12-06 17:53:50.805749] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.199 [2024-12-06 17:53:50.819660] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.199 [2024-12-06 17:53:50.819965] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:12204 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.199 [2024-12-06 17:53:50.819993] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.199 [2024-12-06 17:53:50.833790] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.199 [2024-12-06 17:53:50.834047] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:9889 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.199 [2024-12-06 17:53:50.834074] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.199 [2024-12-06 17:53:50.848003] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.199 [2024-12-06 17:53:50.848300] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:7939 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.199 [2024-12-06 17:53:50.848343] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.199 [2024-12-06 17:53:50.862141] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.199 [2024-12-06 17:53:50.862435] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:18177 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.199 [2024-12-06 17:53:50.862477] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.199 [2024-12-06 17:53:50.876367] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.199 [2024-12-06 17:53:50.876704] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:13369 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.199 [2024-12-06 17:53:50.876733] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.199 [2024-12-06 17:53:50.890570] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.199 [2024-12-06 17:53:50.890828] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:12674 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.199 [2024-12-06 17:53:50.890856] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.199 [2024-12-06 17:53:50.904821] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.199 [2024-12-06 17:53:50.905115] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:8143 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.199 [2024-12-06 17:53:50.905143] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.199 [2024-12-06 17:53:50.918826] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.199 [2024-12-06 17:53:50.919108] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:9744 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.199 [2024-12-06 17:53:50.919149] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.199 [2024-12-06 17:53:50.932789] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.199 [2024-12-06 17:53:50.933030] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:1773 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.199 [2024-12-06 17:53:50.933082] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.199 [2024-12-06 17:53:50.946491] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.199 [2024-12-06 17:53:50.946750] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:19001 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.199 [2024-12-06 17:53:50.946779] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.199 [2024-12-06 17:53:50.960373] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.199 [2024-12-06 17:53:50.960655] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:17168 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.199 [2024-12-06 17:53:50.960690] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.199 [2024-12-06 17:53:50.974148] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.199 [2024-12-06 17:53:50.974384] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:1624 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.199 [2024-12-06 17:53:50.974411] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.199 [2024-12-06 17:53:50.988119] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.199 [2024-12-06 17:53:50.988363] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:11452 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.199 [2024-12-06 17:53:50.988390] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.199 [2024-12-06 17:53:51.002090] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.199 [2024-12-06 17:53:51.002338] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:5576 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.199 [2024-12-06 17:53:51.002378] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.200 [2024-12-06 17:53:51.015912] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.200 [2024-12-06 17:53:51.016163] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:12593 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.200 [2024-12-06 17:53:51.016190] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.200 [2024-12-06 17:53:51.029856] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.200 [2024-12-06 17:53:51.030111] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:22034 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.200 [2024-12-06 17:53:51.030151] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.457 [2024-12-06 17:53:51.043415] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.458 [2024-12-06 17:53:51.043648] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:15119 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.458 [2024-12-06 17:53:51.043697] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.458 [2024-12-06 17:53:51.057375] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.458 [2024-12-06 17:53:51.057618] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:5408 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.458 [2024-12-06 17:53:51.057645] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.458 [2024-12-06 17:53:51.071341] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.458 [2024-12-06 17:53:51.071577] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:14865 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.458 [2024-12-06 17:53:51.071604] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.458 [2024-12-06 17:53:51.085163] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.458 [2024-12-06 17:53:51.085403] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:7785 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.458 [2024-12-06 17:53:51.085430] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.458 [2024-12-06 17:53:51.098973] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.458 [2024-12-06 17:53:51.099211] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:20920 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.458 [2024-12-06 17:53:51.099251] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.458 [2024-12-06 17:53:51.112943] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.458 [2024-12-06 17:53:51.113194] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:13090 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.458 [2024-12-06 17:53:51.113221] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.458 [2024-12-06 17:53:51.126888] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.458 [2024-12-06 17:53:51.127125] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:21245 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.458 [2024-12-06 17:53:51.127152] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.458 [2024-12-06 17:53:51.140784] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.458 [2024-12-06 17:53:51.141032] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:14707 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.458 [2024-12-06 17:53:51.141058] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.458 [2024-12-06 17:53:51.154681] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.458 [2024-12-06 17:53:51.154933] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:24831 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.458 [2024-12-06 17:53:51.154961] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.458 [2024-12-06 17:53:51.168791] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.458 [2024-12-06 17:53:51.169028] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:18080 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.458 [2024-12-06 17:53:51.169055] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.458 [2024-12-06 17:53:51.182606] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.458 [2024-12-06 17:53:51.182846] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:2227 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.458 [2024-12-06 17:53:51.182889] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.458 18430.00 IOPS, 71.99 MiB/s [2024-12-06T16:53:51.297Z] [2024-12-06 17:53:51.196349] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.458 [2024-12-06 17:53:51.196605] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:21564 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.458 [2024-12-06 17:53:51.196648] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.458 [2024-12-06 17:53:51.210303] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.458 [2024-12-06 17:53:51.210526] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:22424 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.458 [2024-12-06 17:53:51.210567] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.458 [2024-12-06 17:53:51.223652] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.458 [2024-12-06 17:53:51.223882] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:3164 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.458 [2024-12-06 17:53:51.223910] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.458 [2024-12-06 17:53:51.237384] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.458 [2024-12-06 17:53:51.237613] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:19808 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.458 [2024-12-06 17:53:51.237656] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.458 [2024-12-06 17:53:51.251234] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.458 [2024-12-06 17:53:51.251458] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:13502 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.458 [2024-12-06 17:53:51.251500] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.458 [2024-12-06 17:53:51.265129] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.458 [2024-12-06 17:53:51.265366] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:11432 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.458 [2024-12-06 17:53:51.265394] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.458 [2024-12-06 17:53:51.279038] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.458 [2024-12-06 17:53:51.279287] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:4357 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.458 [2024-12-06 17:53:51.279314] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.458 [2024-12-06 17:53:51.292940] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.458 [2024-12-06 17:53:51.293192] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:24227 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.458 [2024-12-06 17:53:51.293227] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.716 [2024-12-06 17:53:51.306605] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.716 [2024-12-06 17:53:51.306867] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:18401 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.716 [2024-12-06 17:53:51.306896] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.716 [2024-12-06 17:53:51.320594] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.716 [2024-12-06 17:53:51.320855] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:12743 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.716 [2024-12-06 17:53:51.320883] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.716 [2024-12-06 17:53:51.334510] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.716 [2024-12-06 17:53:51.334754] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:23360 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.716 [2024-12-06 17:53:51.334797] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.716 [2024-12-06 17:53:51.348457] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.716 [2024-12-06 17:53:51.348710] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:11142 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.716 [2024-12-06 17:53:51.348738] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.716 [2024-12-06 17:53:51.362486] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.716 [2024-12-06 17:53:51.362750] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:16255 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.716 [2024-12-06 17:53:51.362778] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.717 [2024-12-06 17:53:51.376456] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.717 [2024-12-06 17:53:51.376710] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:7416 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.717 [2024-12-06 17:53:51.376739] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.717 [2024-12-06 17:53:51.390436] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.717 [2024-12-06 17:53:51.390697] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:11676 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.717 [2024-12-06 17:53:51.390740] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.717 [2024-12-06 17:53:51.404310] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.717 [2024-12-06 17:53:51.404545] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:25226 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.717 [2024-12-06 17:53:51.404571] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.717 [2024-12-06 17:53:51.418416] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.717 [2024-12-06 17:53:51.418756] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:5219 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.717 [2024-12-06 17:53:51.418785] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.717 [2024-12-06 17:53:51.432381] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.717 [2024-12-06 17:53:51.432620] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:7858 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.717 [2024-12-06 17:53:51.432648] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.717 [2024-12-06 17:53:51.446292] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.717 [2024-12-06 17:53:51.446534] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:16540 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.717 [2024-12-06 17:53:51.446561] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.717 [2024-12-06 17:53:51.460068] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.717 [2024-12-06 17:53:51.460322] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:12187 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.717 [2024-12-06 17:53:51.460349] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.717 [2024-12-06 17:53:51.473702] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.717 [2024-12-06 17:53:51.473955] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:3139 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.717 [2024-12-06 17:53:51.473998] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.717 [2024-12-06 17:53:51.487672] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.717 [2024-12-06 17:53:51.487905] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:5082 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.717 [2024-12-06 17:53:51.487947] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.717 [2024-12-06 17:53:51.501641] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.717 [2024-12-06 17:53:51.501891] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:12465 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.717 [2024-12-06 17:53:51.501919] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.717 [2024-12-06 17:53:51.515300] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.717 [2024-12-06 17:53:51.515539] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:3608 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.717 [2024-12-06 17:53:51.515566] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.717 [2024-12-06 17:53:51.528988] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.717 [2024-12-06 17:53:51.529238] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:18380 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.717 [2024-12-06 17:53:51.529265] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.717 [2024-12-06 17:53:51.542745] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.717 [2024-12-06 17:53:51.542977] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:12230 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.717 [2024-12-06 17:53:51.543020] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.975 [2024-12-06 17:53:51.556468] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.975 [2024-12-06 17:53:51.556739] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:20540 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.975 [2024-12-06 17:53:51.556768] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.975 [2024-12-06 17:53:51.570511] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.975 [2024-12-06 17:53:51.570777] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:16075 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.975 [2024-12-06 17:53:51.570806] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.975 [2024-12-06 17:53:51.584459] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.975 [2024-12-06 17:53:51.584686] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:150 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.975 [2024-12-06 17:53:51.584715] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.976 [2024-12-06 17:53:51.598326] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.976 [2024-12-06 17:53:51.598563] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:23098 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.976 [2024-12-06 17:53:51.598590] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.976 [2024-12-06 17:53:51.612163] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.976 [2024-12-06 17:53:51.612402] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:22927 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.976 [2024-12-06 17:53:51.612429] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.976 [2024-12-06 17:53:51.626660] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.976 [2024-12-06 17:53:51.626894] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:501 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.976 [2024-12-06 17:53:51.626922] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.976 [2024-12-06 17:53:51.640793] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.976 [2024-12-06 17:53:51.641050] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:20266 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.976 [2024-12-06 17:53:51.641092] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.976 [2024-12-06 17:53:51.654810] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.976 [2024-12-06 17:53:51.655050] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:23917 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.976 [2024-12-06 17:53:51.655084] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.976 [2024-12-06 17:53:51.668580] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.976 [2024-12-06 17:53:51.668844] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:20715 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.976 [2024-12-06 17:53:51.668871] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.976 [2024-12-06 17:53:51.682554] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.976 [2024-12-06 17:53:51.682817] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:21827 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.976 [2024-12-06 17:53:51.682846] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.976 [2024-12-06 17:53:51.696485] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.976 [2024-12-06 17:53:51.696748] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:8450 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.976 [2024-12-06 17:53:51.696776] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.976 [2024-12-06 17:53:51.710450] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.976 [2024-12-06 17:53:51.710687] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:8195 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.976 [2024-12-06 17:53:51.710729] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.976 [2024-12-06 17:53:51.724346] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.976 [2024-12-06 17:53:51.724580] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:8834 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.976 [2024-12-06 17:53:51.724623] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.976 [2024-12-06 17:53:51.737991] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.976 [2024-12-06 17:53:51.738246] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:4560 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.976 [2024-12-06 17:53:51.738288] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.976 [2024-12-06 17:53:51.751912] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.976 [2024-12-06 17:53:51.752160] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:11912 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.976 [2024-12-06 17:53:51.752187] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.976 [2024-12-06 17:53:51.765771] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.976 [2024-12-06 17:53:51.766048] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:17172 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.976 [2024-12-06 17:53:51.766075] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.976 [2024-12-06 17:53:51.779722] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.976 [2024-12-06 17:53:51.779952] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:21622 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.976 [2024-12-06 17:53:51.779994] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.976 [2024-12-06 17:53:51.793560] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.976 [2024-12-06 17:53:51.793850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:556 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.976 [2024-12-06 17:53:51.793878] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:09.976 [2024-12-06 17:53:51.807446] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:09.976 [2024-12-06 17:53:51.807674] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:9493 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:09.976 [2024-12-06 17:53:51.807702] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:10.234 [2024-12-06 17:53:51.820899] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:10.234 [2024-12-06 17:53:51.821151] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:24611 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:10.234 [2024-12-06 17:53:51.821179] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:10.234 [2024-12-06 17:53:51.834977] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:10.234 [2024-12-06 17:53:51.835230] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:17622 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:10.234 [2024-12-06 17:53:51.835272] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:10.234 [2024-12-06 17:53:51.848927] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:10.235 [2024-12-06 17:53:51.849173] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:24660 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:10.235 [2024-12-06 17:53:51.849214] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:10.235 [2024-12-06 17:53:51.862964] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:10.235 [2024-12-06 17:53:51.863215] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:13555 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:10.235 [2024-12-06 17:53:51.863242] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:10.235 [2024-12-06 17:53:51.876945] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:10.235 [2024-12-06 17:53:51.877194] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:3879 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:10.235 [2024-12-06 17:53:51.877220] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:10.235 [2024-12-06 17:53:51.890915] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:10.235 [2024-12-06 17:53:51.891162] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:10151 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:10.235 [2024-12-06 17:53:51.891203] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:10.235 [2024-12-06 17:53:51.904882] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:10.235 [2024-12-06 17:53:51.905160] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:1974 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:10.235 [2024-12-06 17:53:51.905202] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:10.235 [2024-12-06 17:53:51.918896] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:10.235 [2024-12-06 17:53:51.919145] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:11099 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:10.235 [2024-12-06 17:53:51.919172] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:10.235 [2024-12-06 17:53:51.932863] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:10.235 [2024-12-06 17:53:51.933100] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:22660 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:10.235 [2024-12-06 17:53:51.933126] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:10.235 [2024-12-06 17:53:51.946802] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:10.235 [2024-12-06 17:53:51.947055] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:3289 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:10.235 [2024-12-06 17:53:51.947096] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:10.235 [2024-12-06 17:53:51.960770] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:10.235 [2024-12-06 17:53:51.961007] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:813 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:10.235 [2024-12-06 17:53:51.961034] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:10.235 [2024-12-06 17:53:51.974610] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:10.235 [2024-12-06 17:53:51.974845] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:1142 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:10.235 [2024-12-06 17:53:51.974873] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:10.235 [2024-12-06 17:53:51.988191] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:10.235 [2024-12-06 17:53:51.988427] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:12050 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:10.235 [2024-12-06 17:53:51.988470] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:10.235 [2024-12-06 17:53:52.002019] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:10.235 [2024-12-06 17:53:52.002267] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:15948 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:10.235 [2024-12-06 17:53:52.002308] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:10.235 [2024-12-06 17:53:52.015922] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:10.235 [2024-12-06 17:53:52.016167] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:4873 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:10.235 [2024-12-06 17:53:52.016216] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:10.235 [2024-12-06 17:53:52.029737] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:10.235 [2024-12-06 17:53:52.029986] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:11615 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:10.235 [2024-12-06 17:53:52.030026] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:10.235 [2024-12-06 17:53:52.043721] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:10.235 [2024-12-06 17:53:52.043957] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:4627 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:10.235 [2024-12-06 17:53:52.043983] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:10.235 [2024-12-06 17:53:52.057543] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:10.235 [2024-12-06 17:53:52.057804] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:13779 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:10.235 [2024-12-06 17:53:52.057832] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:10.235 [2024-12-06 17:53:52.071166] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:10.235 [2024-12-06 17:53:52.071403] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:11015 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:10.235 [2024-12-06 17:53:52.071431] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:10.494 [2024-12-06 17:53:52.084727] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:10.494 [2024-12-06 17:53:52.084961] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:3615 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:10.494 [2024-12-06 17:53:52.084987] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:10.494 [2024-12-06 17:53:52.098678] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:10.494 [2024-12-06 17:53:52.098902] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:2881 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:10.494 [2024-12-06 17:53:52.098929] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:10.494 [2024-12-06 17:53:52.112818] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:10.494 [2024-12-06 17:53:52.113060] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:21297 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:10.494 [2024-12-06 17:53:52.113087] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:10.494 [2024-12-06 17:53:52.126820] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:10.494 [2024-12-06 17:53:52.127099] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:3156 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:10.494 [2024-12-06 17:53:52.127140] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:10.494 [2024-12-06 17:53:52.140828] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:10.494 [2024-12-06 17:53:52.141083] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:23470 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:10.494 [2024-12-06 17:53:52.141123] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:10.494 [2024-12-06 17:53:52.154817] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:10.494 [2024-12-06 17:53:52.155095] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:14525 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:10.494 [2024-12-06 17:53:52.155136] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:10.494 [2024-12-06 17:53:52.168805] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:10.494 [2024-12-06 17:53:52.169052] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:5653 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:10.494 [2024-12-06 17:53:52.169093] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:10.494 [2024-12-06 17:53:52.182643] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b3e70) with pdu=0x200016ee88f8 00:38:10.494 [2024-12-06 17:53:52.182888] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:1934 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:10.494 [2024-12-06 17:53:52.182915] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:10.494 18415.00 IOPS, 71.93 MiB/s 00:38:10.494 Latency(us) 00:38:10.494 [2024-12-06T16:53:52.333Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:38:10.494 Job: nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:38:10.494 nvme0n1 : 2.01 18418.45 71.95 0.00 0.00 6934.17 2694.26 14854.83 00:38:10.494 [2024-12-06T16:53:52.333Z] =================================================================================================================== 00:38:10.494 [2024-12-06T16:53:52.333Z] Total : 18418.45 71.95 0.00 0.00 6934.17 2694.26 14854.83 00:38:10.494 { 00:38:10.494 "results": [ 00:38:10.494 { 00:38:10.494 "job": "nvme0n1", 00:38:10.494 "core_mask": "0x2", 00:38:10.494 "workload": "randwrite", 00:38:10.494 "status": "finished", 00:38:10.494 "queue_depth": 128, 00:38:10.494 "io_size": 4096, 00:38:10.494 "runtime": 2.006575, 00:38:10.494 "iops": 18418.44934776921, 00:38:10.494 "mibps": 71.94706776472347, 00:38:10.494 "io_failed": 0, 00:38:10.494 "io_timeout": 0, 00:38:10.494 "avg_latency_us": 6934.1676924156145, 00:38:10.494 "min_latency_us": 2694.257777777778, 00:38:10.494 "max_latency_us": 14854.826666666666 00:38:10.494 } 00:38:10.494 ], 00:38:10.494 "core_count": 1 00:38:10.494 } 00:38:10.494 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # get_transient_errcount nvme0n1 00:38:10.494 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@27 -- # bperf_rpc bdev_get_iostat -b nvme0n1 00:38:10.494 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@28 -- # jq -r '.bdevs[0] 00:38:10.494 | .driver_specific 00:38:10.494 | .nvme_error 00:38:10.494 | .status_code 00:38:10.494 | .command_transient_transport_error' 00:38:10.494 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_get_iostat -b nvme0n1 00:38:10.752 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # (( 144 > 0 )) 00:38:10.752 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@73 -- # killprocess 407116 00:38:10.752 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # '[' -z 407116 ']' 00:38:10.752 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@958 -- # kill -0 407116 00:38:10.752 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@959 -- # uname 00:38:10.752 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:38:10.752 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 407116 00:38:10.752 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:38:10.752 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:38:10.752 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@972 -- # echo 'killing process with pid 407116' 00:38:10.752 killing process with pid 407116 00:38:10.752 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@973 -- # kill 407116 00:38:10.752 Received shutdown signal, test time was about 2.000000 seconds 00:38:10.752 00:38:10.752 Latency(us) 00:38:10.752 [2024-12-06T16:53:52.591Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:38:10.752 [2024-12-06T16:53:52.591Z] =================================================================================================================== 00:38:10.752 [2024-12-06T16:53:52.591Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:38:10.752 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@978 -- # wait 407116 00:38:11.015 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@115 -- # run_bperf_err randwrite 131072 16 00:38:11.015 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@54 -- # local rw bs qd 00:38:11.015 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # rw=randwrite 00:38:11.015 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # bs=131072 00:38:11.015 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # qd=16 00:38:11.015 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@58 -- # bperfpid=407635 00:38:11.015 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randwrite -o 131072 -t 2 -q 16 -z 00:38:11.015 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@60 -- # waitforlisten 407635 /var/tmp/bperf.sock 00:38:11.015 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@835 -- # '[' -z 407635 ']' 00:38:11.015 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bperf.sock 00:38:11.015 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@840 -- # local max_retries=100 00:38:11.015 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:38:11.015 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:38:11.015 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@844 -- # xtrace_disable 00:38:11.015 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:38:11.015 [2024-12-06 17:53:52.758147] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:38:11.015 [2024-12-06 17:53:52.758230] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid407635 ] 00:38:11.015 I/O size of 131072 is greater than zero copy threshold (65536). 00:38:11.015 Zero copy mechanism will not be used. 00:38:11.015 [2024-12-06 17:53:52.824670] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:38:11.274 [2024-12-06 17:53:52.870701] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:38:11.274 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:38:11.274 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@868 -- # return 0 00:38:11.274 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@61 -- # bperf_rpc bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:38:11.274 17:53:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:38:11.531 17:53:53 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@63 -- # rpc_cmd accel_error_inject_error -o crc32c -t disable 00:38:11.531 17:53:53 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:11.531 17:53:53 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:38:11.531 17:53:53 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:11.531 17:53:53 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@64 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:38:11.531 17:53:53 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:38:12.097 nvme0n1 00:38:12.097 17:53:53 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@67 -- # rpc_cmd accel_error_inject_error -o crc32c -t corrupt -i 32 00:38:12.097 17:53:53 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:12.097 17:53:53 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:38:12.097 17:53:53 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:12.097 17:53:53 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@69 -- # bperf_py perform_tests 00:38:12.097 17:53:53 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:38:12.097 I/O size of 131072 is greater than zero copy threshold (65536). 00:38:12.097 Zero copy mechanism will not be used. 00:38:12.097 Running I/O for 2 seconds... 00:38:12.097 [2024-12-06 17:53:53.782930] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.097 [2024-12-06 17:53:53.783074] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:2496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.097 [2024-12-06 17:53:53.783112] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.097 [2024-12-06 17:53:53.788818] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.097 [2024-12-06 17:53:53.788916] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.097 [2024-12-06 17:53:53.788949] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.097 [2024-12-06 17:53:53.794599] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.097 [2024-12-06 17:53:53.794733] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:21344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.097 [2024-12-06 17:53:53.794764] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.097 [2024-12-06 17:53:53.800776] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.097 [2024-12-06 17:53:53.800865] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:18336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.097 [2024-12-06 17:53:53.800894] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.097 [2024-12-06 17:53:53.806726] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.097 [2024-12-06 17:53:53.806804] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:16832 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.097 [2024-12-06 17:53:53.806832] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.097 [2024-12-06 17:53:53.812475] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.097 [2024-12-06 17:53:53.812548] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:18432 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.097 [2024-12-06 17:53:53.812577] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.097 [2024-12-06 17:53:53.817604] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.097 [2024-12-06 17:53:53.817717] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:18656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.097 [2024-12-06 17:53:53.817746] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.097 [2024-12-06 17:53:53.822935] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.097 [2024-12-06 17:53:53.823061] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:2144 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.097 [2024-12-06 17:53:53.823089] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.097 [2024-12-06 17:53:53.828599] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.097 [2024-12-06 17:53:53.828700] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:21664 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.097 [2024-12-06 17:53:53.828728] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.097 [2024-12-06 17:53:53.833850] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.097 [2024-12-06 17:53:53.833925] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14528 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.097 [2024-12-06 17:53:53.833952] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.097 [2024-12-06 17:53:53.838974] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.097 [2024-12-06 17:53:53.839075] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:16864 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.098 [2024-12-06 17:53:53.839103] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.098 [2024-12-06 17:53:53.844123] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.098 [2024-12-06 17:53:53.844209] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:17632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.098 [2024-12-06 17:53:53.844236] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.098 [2024-12-06 17:53:53.849197] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.098 [2024-12-06 17:53:53.849292] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:11456 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.098 [2024-12-06 17:53:53.849328] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.098 [2024-12-06 17:53:53.854206] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.098 [2024-12-06 17:53:53.854288] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:11712 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.098 [2024-12-06 17:53:53.854316] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.098 [2024-12-06 17:53:53.859643] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.098 [2024-12-06 17:53:53.859732] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:13728 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.098 [2024-12-06 17:53:53.859760] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.098 [2024-12-06 17:53:53.865477] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.098 [2024-12-06 17:53:53.865607] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:18688 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.098 [2024-12-06 17:53:53.865636] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.098 [2024-12-06 17:53:53.872046] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.098 [2024-12-06 17:53:53.872162] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20480 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.098 [2024-12-06 17:53:53.872191] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.098 [2024-12-06 17:53:53.879006] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.098 [2024-12-06 17:53:53.879092] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:13792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.098 [2024-12-06 17:53:53.879120] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.098 [2024-12-06 17:53:53.885471] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.098 [2024-12-06 17:53:53.885605] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:18560 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.098 [2024-12-06 17:53:53.885635] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.098 [2024-12-06 17:53:53.890259] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.098 [2024-12-06 17:53:53.890529] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:11072 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.098 [2024-12-06 17:53:53.890559] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.098 [2024-12-06 17:53:53.895024] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.098 [2024-12-06 17:53:53.895335] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:2592 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.098 [2024-12-06 17:53:53.895365] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.098 [2024-12-06 17:53:53.900314] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.098 [2024-12-06 17:53:53.900602] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19040 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.098 [2024-12-06 17:53:53.900632] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.098 [2024-12-06 17:53:53.906153] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.098 [2024-12-06 17:53:53.906490] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:22816 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.098 [2024-12-06 17:53:53.906520] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.098 [2024-12-06 17:53:53.910700] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.098 [2024-12-06 17:53:53.911047] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:9184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.098 [2024-12-06 17:53:53.911076] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.098 [2024-12-06 17:53:53.916315] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.098 [2024-12-06 17:53:53.916618] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:22688 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.098 [2024-12-06 17:53:53.916647] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.098 [2024-12-06 17:53:53.921904] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.098 [2024-12-06 17:53:53.922206] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24160 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.098 [2024-12-06 17:53:53.922235] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.098 [2024-12-06 17:53:53.927524] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.098 [2024-12-06 17:53:53.927803] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:13408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.098 [2024-12-06 17:53:53.927832] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.098 [2024-12-06 17:53:53.933087] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.098 [2024-12-06 17:53:53.933397] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14528 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.098 [2024-12-06 17:53:53.933425] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.357 [2024-12-06 17:53:53.938683] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.357 [2024-12-06 17:53:53.938985] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:2880 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.357 [2024-12-06 17:53:53.939014] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.357 [2024-12-06 17:53:53.944330] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.357 [2024-12-06 17:53:53.944590] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:11104 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.357 [2024-12-06 17:53:53.944618] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.357 [2024-12-06 17:53:53.950642] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.357 [2024-12-06 17:53:53.950981] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:3296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.357 [2024-12-06 17:53:53.951010] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.357 [2024-12-06 17:53:53.956874] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.357 [2024-12-06 17:53:53.957121] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:12896 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.357 [2024-12-06 17:53:53.957150] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.357 [2024-12-06 17:53:53.963114] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.357 [2024-12-06 17:53:53.963460] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14304 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.357 [2024-12-06 17:53:53.963489] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.357 [2024-12-06 17:53:53.968875] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.357 [2024-12-06 17:53:53.969121] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:23552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.357 [2024-12-06 17:53:53.969149] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.357 [2024-12-06 17:53:53.973531] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.357 [2024-12-06 17:53:53.973795] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:21344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.357 [2024-12-06 17:53:53.973824] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.357 [2024-12-06 17:53:53.977978] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.357 [2024-12-06 17:53:53.978244] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:2752 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.357 [2024-12-06 17:53:53.978273] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.357 [2024-12-06 17:53:53.982565] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.357 [2024-12-06 17:53:53.982811] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.357 [2024-12-06 17:53:53.982840] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.357 [2024-12-06 17:53:53.987770] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.357 [2024-12-06 17:53:53.988014] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:16896 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.357 [2024-12-06 17:53:53.988042] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.357 [2024-12-06 17:53:53.992973] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.357 [2024-12-06 17:53:53.993234] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:4736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.357 [2024-12-06 17:53:53.993267] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.357 [2024-12-06 17:53:53.998256] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.357 [2024-12-06 17:53:53.998589] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:11776 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.357 [2024-12-06 17:53:53.998618] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.357 [2024-12-06 17:53:54.003597] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.357 [2024-12-06 17:53:54.003924] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:11040 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.357 [2024-12-06 17:53:54.003952] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.357 [2024-12-06 17:53:54.008842] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.357 [2024-12-06 17:53:54.009123] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:20544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.357 [2024-12-06 17:53:54.009151] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.357 [2024-12-06 17:53:54.014944] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.357 [2024-12-06 17:53:54.015191] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:18144 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.357 [2024-12-06 17:53:54.015219] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.357 [2024-12-06 17:53:54.020156] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.357 [2024-12-06 17:53:54.020401] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:6784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.357 [2024-12-06 17:53:54.020429] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.358 [2024-12-06 17:53:54.025221] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.358 [2024-12-06 17:53:54.025499] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:9056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.358 [2024-12-06 17:53:54.025528] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.358 [2024-12-06 17:53:54.030428] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.358 [2024-12-06 17:53:54.030719] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:10272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.358 [2024-12-06 17:53:54.030748] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.358 [2024-12-06 17:53:54.035766] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.358 [2024-12-06 17:53:54.035933] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:17216 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.358 [2024-12-06 17:53:54.035962] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.358 [2024-12-06 17:53:54.040858] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.358 [2024-12-06 17:53:54.041087] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:6240 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.358 [2024-12-06 17:53:54.041116] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.358 [2024-12-06 17:53:54.046078] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.358 [2024-12-06 17:53:54.046311] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:1056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.358 [2024-12-06 17:53:54.046340] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.358 [2024-12-06 17:53:54.051073] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.358 [2024-12-06 17:53:54.051236] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:14464 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.358 [2024-12-06 17:53:54.051264] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.358 [2024-12-06 17:53:54.056207] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.358 [2024-12-06 17:53:54.056439] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:23648 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.358 [2024-12-06 17:53:54.056468] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.358 [2024-12-06 17:53:54.061209] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.358 [2024-12-06 17:53:54.061390] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:16832 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.358 [2024-12-06 17:53:54.061418] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.358 [2024-12-06 17:53:54.066258] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.358 [2024-12-06 17:53:54.066475] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:11584 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.358 [2024-12-06 17:53:54.066504] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.358 [2024-12-06 17:53:54.072135] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.358 [2024-12-06 17:53:54.072315] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:19168 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.358 [2024-12-06 17:53:54.072343] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.358 [2024-12-06 17:53:54.077196] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.358 [2024-12-06 17:53:54.077418] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:7360 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.358 [2024-12-06 17:53:54.077447] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.358 [2024-12-06 17:53:54.082326] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.358 [2024-12-06 17:53:54.082497] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:15072 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.358 [2024-12-06 17:53:54.082525] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.358 [2024-12-06 17:53:54.087547] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.358 [2024-12-06 17:53:54.087720] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:18464 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.358 [2024-12-06 17:53:54.087749] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.358 [2024-12-06 17:53:54.092653] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.358 [2024-12-06 17:53:54.092908] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:19040 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.358 [2024-12-06 17:53:54.092936] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.358 [2024-12-06 17:53:54.097783] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.358 [2024-12-06 17:53:54.097974] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:2400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.358 [2024-12-06 17:53:54.098003] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.358 [2024-12-06 17:53:54.104002] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.358 [2024-12-06 17:53:54.104215] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:7648 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.358 [2024-12-06 17:53:54.104244] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.358 [2024-12-06 17:53:54.109168] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.358 [2024-12-06 17:53:54.109428] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:8384 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.358 [2024-12-06 17:53:54.109456] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.358 [2024-12-06 17:53:54.114365] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.358 [2024-12-06 17:53:54.114645] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:14336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.358 [2024-12-06 17:53:54.114681] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.358 [2024-12-06 17:53:54.119476] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.358 [2024-12-06 17:53:54.119660] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:12416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.358 [2024-12-06 17:53:54.119695] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.358 [2024-12-06 17:53:54.124671] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.358 [2024-12-06 17:53:54.124924] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:3872 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.358 [2024-12-06 17:53:54.124952] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.358 [2024-12-06 17:53:54.129786] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.358 [2024-12-06 17:53:54.129999] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:8384 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.358 [2024-12-06 17:53:54.130032] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.358 [2024-12-06 17:53:54.135126] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.358 [2024-12-06 17:53:54.135335] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:20288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.358 [2024-12-06 17:53:54.135364] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.358 [2024-12-06 17:53:54.140207] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.358 [2024-12-06 17:53:54.140416] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24384 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.358 [2024-12-06 17:53:54.140444] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.358 [2024-12-06 17:53:54.145415] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.358 [2024-12-06 17:53:54.145632] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:8320 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.358 [2024-12-06 17:53:54.145660] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.358 [2024-12-06 17:53:54.150480] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.358 [2024-12-06 17:53:54.150706] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:5088 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.358 [2024-12-06 17:53:54.150735] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.358 [2024-12-06 17:53:54.155644] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.358 [2024-12-06 17:53:54.155908] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.358 [2024-12-06 17:53:54.155937] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.358 [2024-12-06 17:53:54.160757] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.358 [2024-12-06 17:53:54.160990] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:7168 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.358 [2024-12-06 17:53:54.161019] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.358 [2024-12-06 17:53:54.165845] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.359 [2024-12-06 17:53:54.166070] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:12928 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.359 [2024-12-06 17:53:54.166099] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.359 [2024-12-06 17:53:54.171019] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.359 [2024-12-06 17:53:54.171274] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24480 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.359 [2024-12-06 17:53:54.171303] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.359 [2024-12-06 17:53:54.176167] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.359 [2024-12-06 17:53:54.176352] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:8352 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.359 [2024-12-06 17:53:54.176381] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.359 [2024-12-06 17:53:54.181405] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.359 [2024-12-06 17:53:54.181566] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19360 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.359 [2024-12-06 17:53:54.181594] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.359 [2024-12-06 17:53:54.186576] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.359 [2024-12-06 17:53:54.186786] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:12032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.359 [2024-12-06 17:53:54.186815] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.359 [2024-12-06 17:53:54.191715] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.359 [2024-12-06 17:53:54.191977] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:1984 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.359 [2024-12-06 17:53:54.192005] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.617 [2024-12-06 17:53:54.196844] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.617 [2024-12-06 17:53:54.197062] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:17216 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.617 [2024-12-06 17:53:54.197091] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.617 [2024-12-06 17:53:54.202198] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.617 [2024-12-06 17:53:54.202392] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:2240 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.617 [2024-12-06 17:53:54.202421] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.617 [2024-12-06 17:53:54.207395] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.617 [2024-12-06 17:53:54.207590] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:12928 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.617 [2024-12-06 17:53:54.207618] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.617 [2024-12-06 17:53:54.212632] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.617 [2024-12-06 17:53:54.212822] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:5664 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.617 [2024-12-06 17:53:54.212851] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.617 [2024-12-06 17:53:54.217709] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.617 [2024-12-06 17:53:54.217884] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:17408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.617 [2024-12-06 17:53:54.217913] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.618 [2024-12-06 17:53:54.222935] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.618 [2024-12-06 17:53:54.223090] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:3104 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.618 [2024-12-06 17:53:54.223118] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.618 [2024-12-06 17:53:54.228091] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.618 [2024-12-06 17:53:54.228286] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:7968 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.618 [2024-12-06 17:53:54.228315] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.618 [2024-12-06 17:53:54.233290] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.618 [2024-12-06 17:53:54.233443] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:17952 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.618 [2024-12-06 17:53:54.233471] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.618 [2024-12-06 17:53:54.238465] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.618 [2024-12-06 17:53:54.238708] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.618 [2024-12-06 17:53:54.238737] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.618 [2024-12-06 17:53:54.243606] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.618 [2024-12-06 17:53:54.243798] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:7232 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.618 [2024-12-06 17:53:54.243825] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.618 [2024-12-06 17:53:54.248870] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.618 [2024-12-06 17:53:54.249116] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.618 [2024-12-06 17:53:54.249145] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.618 [2024-12-06 17:53:54.254084] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.618 [2024-12-06 17:53:54.254306] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:1440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.618 [2024-12-06 17:53:54.254334] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.618 [2024-12-06 17:53:54.259172] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.618 [2024-12-06 17:53:54.259404] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:21280 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.618 [2024-12-06 17:53:54.259433] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.618 [2024-12-06 17:53:54.264414] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.618 [2024-12-06 17:53:54.264618] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24064 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.618 [2024-12-06 17:53:54.264651] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.618 [2024-12-06 17:53:54.269491] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.618 [2024-12-06 17:53:54.269706] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:15840 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.618 [2024-12-06 17:53:54.269734] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.618 [2024-12-06 17:53:54.274728] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.618 [2024-12-06 17:53:54.274905] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:8256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.618 [2024-12-06 17:53:54.274933] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.618 [2024-12-06 17:53:54.279900] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.618 [2024-12-06 17:53:54.280107] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:16800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.618 [2024-12-06 17:53:54.280134] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.618 [2024-12-06 17:53:54.285014] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.618 [2024-12-06 17:53:54.285182] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:1024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.618 [2024-12-06 17:53:54.285210] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.618 [2024-12-06 17:53:54.290283] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.618 [2024-12-06 17:53:54.290523] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.618 [2024-12-06 17:53:54.290552] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.618 [2024-12-06 17:53:54.295729] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.618 [2024-12-06 17:53:54.295980] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:12224 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.618 [2024-12-06 17:53:54.296010] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.618 [2024-12-06 17:53:54.300155] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.618 [2024-12-06 17:53:54.300288] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:21312 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.618 [2024-12-06 17:53:54.300316] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.618 [2024-12-06 17:53:54.304444] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.618 [2024-12-06 17:53:54.304587] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:13056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.618 [2024-12-06 17:53:54.304615] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.618 [2024-12-06 17:53:54.309204] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.618 [2024-12-06 17:53:54.309331] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.618 [2024-12-06 17:53:54.309366] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.618 [2024-12-06 17:53:54.314493] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.618 [2024-12-06 17:53:54.314631] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:10208 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.618 [2024-12-06 17:53:54.314660] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.618 [2024-12-06 17:53:54.318933] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.618 [2024-12-06 17:53:54.319054] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:16608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.618 [2024-12-06 17:53:54.319081] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.618 [2024-12-06 17:53:54.323257] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.618 [2024-12-06 17:53:54.323391] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:8576 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.618 [2024-12-06 17:53:54.323419] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.618 [2024-12-06 17:53:54.327618] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.618 [2024-12-06 17:53:54.327785] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:11744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.618 [2024-12-06 17:53:54.327812] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.618 [2024-12-06 17:53:54.332126] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.618 [2024-12-06 17:53:54.332292] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:13824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.618 [2024-12-06 17:53:54.332320] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.618 [2024-12-06 17:53:54.337230] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.618 [2024-12-06 17:53:54.337412] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:7616 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.618 [2024-12-06 17:53:54.337441] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.618 [2024-12-06 17:53:54.342358] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.618 [2024-12-06 17:53:54.342600] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:96 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.618 [2024-12-06 17:53:54.342628] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.618 [2024-12-06 17:53:54.348353] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.618 [2024-12-06 17:53:54.348546] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:6784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.618 [2024-12-06 17:53:54.348574] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.618 [2024-12-06 17:53:54.352967] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.618 [2024-12-06 17:53:54.353113] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:12832 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.618 [2024-12-06 17:53:54.353141] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.618 [2024-12-06 17:53:54.357200] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.619 [2024-12-06 17:53:54.357334] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:22368 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.619 [2024-12-06 17:53:54.357362] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.619 [2024-12-06 17:53:54.361576] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.619 [2024-12-06 17:53:54.361743] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:9760 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.619 [2024-12-06 17:53:54.361771] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.619 [2024-12-06 17:53:54.365913] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.619 [2024-12-06 17:53:54.366080] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:21408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.619 [2024-12-06 17:53:54.366108] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.619 [2024-12-06 17:53:54.370076] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.619 [2024-12-06 17:53:54.370214] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:1504 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.619 [2024-12-06 17:53:54.370242] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.619 [2024-12-06 17:53:54.374344] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.619 [2024-12-06 17:53:54.374541] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:3200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.619 [2024-12-06 17:53:54.374569] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.619 [2024-12-06 17:53:54.379444] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.619 [2024-12-06 17:53:54.379629] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:10176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.619 [2024-12-06 17:53:54.379657] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.619 [2024-12-06 17:53:54.384641] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.619 [2024-12-06 17:53:54.384846] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:2880 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.619 [2024-12-06 17:53:54.384874] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.619 [2024-12-06 17:53:54.390377] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.619 [2024-12-06 17:53:54.390553] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:10464 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.619 [2024-12-06 17:53:54.390582] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.619 [2024-12-06 17:53:54.395529] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.619 [2024-12-06 17:53:54.395722] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:18080 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.619 [2024-12-06 17:53:54.395750] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.619 [2024-12-06 17:53:54.400743] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.619 [2024-12-06 17:53:54.400933] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:12192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.619 [2024-12-06 17:53:54.400961] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.619 [2024-12-06 17:53:54.405810] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.619 [2024-12-06 17:53:54.406015] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:8608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.619 [2024-12-06 17:53:54.406043] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.619 [2024-12-06 17:53:54.410988] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.619 [2024-12-06 17:53:54.411210] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:19680 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.619 [2024-12-06 17:53:54.411238] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.619 [2024-12-06 17:53:54.416112] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.619 [2024-12-06 17:53:54.416289] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:5312 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.619 [2024-12-06 17:53:54.416317] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.619 [2024-12-06 17:53:54.421227] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.619 [2024-12-06 17:53:54.421371] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:23200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.619 [2024-12-06 17:53:54.421398] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.619 [2024-12-06 17:53:54.426364] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.619 [2024-12-06 17:53:54.426587] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:12512 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.619 [2024-12-06 17:53:54.426615] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.619 [2024-12-06 17:53:54.431493] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.619 [2024-12-06 17:53:54.431638] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:18816 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.619 [2024-12-06 17:53:54.431673] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.619 [2024-12-06 17:53:54.437405] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.619 [2024-12-06 17:53:54.437587] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:20544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.619 [2024-12-06 17:53:54.437621] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.619 [2024-12-06 17:53:54.442915] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.619 [2024-12-06 17:53:54.443081] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:20320 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.619 [2024-12-06 17:53:54.443109] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.619 [2024-12-06 17:53:54.448140] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.619 [2024-12-06 17:53:54.448288] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:2368 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.619 [2024-12-06 17:53:54.448316] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.619 [2024-12-06 17:53:54.453301] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.619 [2024-12-06 17:53:54.453504] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:5248 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.619 [2024-12-06 17:53:54.453533] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.878 [2024-12-06 17:53:54.458408] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.878 [2024-12-06 17:53:54.458594] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:13344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.878 [2024-12-06 17:53:54.458622] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.878 [2024-12-06 17:53:54.463599] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.878 [2024-12-06 17:53:54.463798] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:20096 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.878 [2024-12-06 17:53:54.463826] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.878 [2024-12-06 17:53:54.468708] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.878 [2024-12-06 17:53:54.468875] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:19680 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.878 [2024-12-06 17:53:54.468903] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.878 [2024-12-06 17:53:54.474149] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.878 [2024-12-06 17:53:54.474351] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:16800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.878 [2024-12-06 17:53:54.474379] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.878 [2024-12-06 17:53:54.479744] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.878 [2024-12-06 17:53:54.479930] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:11680 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.878 [2024-12-06 17:53:54.479958] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.878 [2024-12-06 17:53:54.484979] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.878 [2024-12-06 17:53:54.485206] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:7872 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.878 [2024-12-06 17:53:54.485234] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.878 [2024-12-06 17:53:54.490108] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.878 [2024-12-06 17:53:54.490327] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:10240 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.878 [2024-12-06 17:53:54.490355] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.878 [2024-12-06 17:53:54.495360] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.878 [2024-12-06 17:53:54.495543] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:12864 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.878 [2024-12-06 17:53:54.495570] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.878 [2024-12-06 17:53:54.500516] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.878 [2024-12-06 17:53:54.500761] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:22496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.878 [2024-12-06 17:53:54.500789] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.878 [2024-12-06 17:53:54.505738] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.878 [2024-12-06 17:53:54.505964] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:22752 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.878 [2024-12-06 17:53:54.505991] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.878 [2024-12-06 17:53:54.510857] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.878 [2024-12-06 17:53:54.511040] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:9440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.878 [2024-12-06 17:53:54.511068] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.878 [2024-12-06 17:53:54.515895] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.878 [2024-12-06 17:53:54.516120] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:9824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.878 [2024-12-06 17:53:54.516147] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.878 [2024-12-06 17:53:54.521029] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.878 [2024-12-06 17:53:54.521205] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:18496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.878 [2024-12-06 17:53:54.521233] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.878 [2024-12-06 17:53:54.526294] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.878 [2024-12-06 17:53:54.526551] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:1888 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.879 [2024-12-06 17:53:54.526578] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.879 [2024-12-06 17:53:54.531502] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.879 [2024-12-06 17:53:54.531754] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24480 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.879 [2024-12-06 17:53:54.531783] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.879 [2024-12-06 17:53:54.536720] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.879 [2024-12-06 17:53:54.536963] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:23456 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.879 [2024-12-06 17:53:54.536993] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.879 [2024-12-06 17:53:54.541780] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.879 [2024-12-06 17:53:54.542033] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:21984 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.879 [2024-12-06 17:53:54.542062] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.879 [2024-12-06 17:53:54.547164] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.879 [2024-12-06 17:53:54.547381] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:22336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.879 [2024-12-06 17:53:54.547414] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.879 [2024-12-06 17:53:54.552365] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.879 [2024-12-06 17:53:54.552575] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:16960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.879 [2024-12-06 17:53:54.552604] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.879 [2024-12-06 17:53:54.557814] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.879 [2024-12-06 17:53:54.558050] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:19872 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.879 [2024-12-06 17:53:54.558079] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.879 [2024-12-06 17:53:54.563400] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.879 [2024-12-06 17:53:54.563648] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:11744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.879 [2024-12-06 17:53:54.563685] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.879 [2024-12-06 17:53:54.568568] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.879 [2024-12-06 17:53:54.568736] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:10976 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.879 [2024-12-06 17:53:54.568765] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.879 [2024-12-06 17:53:54.573740] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.879 [2024-12-06 17:53:54.573934] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:2368 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.879 [2024-12-06 17:53:54.573969] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.879 [2024-12-06 17:53:54.578844] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.879 [2024-12-06 17:53:54.579009] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24096 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.879 [2024-12-06 17:53:54.579037] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.879 [2024-12-06 17:53:54.584058] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.879 [2024-12-06 17:53:54.584224] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:7008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.879 [2024-12-06 17:53:54.584252] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.879 [2024-12-06 17:53:54.589182] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.879 [2024-12-06 17:53:54.589427] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:11040 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.879 [2024-12-06 17:53:54.589455] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.879 [2024-12-06 17:53:54.594254] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.879 [2024-12-06 17:53:54.594466] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:16032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.879 [2024-12-06 17:53:54.594494] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.879 [2024-12-06 17:53:54.599327] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.879 [2024-12-06 17:53:54.599588] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23712 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.879 [2024-12-06 17:53:54.599617] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.879 [2024-12-06 17:53:54.604558] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.879 [2024-12-06 17:53:54.604745] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:22432 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.879 [2024-12-06 17:53:54.604772] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.879 [2024-12-06 17:53:54.609827] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.879 [2024-12-06 17:53:54.610002] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.879 [2024-12-06 17:53:54.610031] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.879 [2024-12-06 17:53:54.614954] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.879 [2024-12-06 17:53:54.615208] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:12928 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.879 [2024-12-06 17:53:54.615236] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.879 [2024-12-06 17:53:54.620092] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.879 [2024-12-06 17:53:54.620298] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:1824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.879 [2024-12-06 17:53:54.620327] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.879 [2024-12-06 17:53:54.625166] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.879 [2024-12-06 17:53:54.625375] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19936 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.879 [2024-12-06 17:53:54.625403] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.879 [2024-12-06 17:53:54.630416] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.879 [2024-12-06 17:53:54.630660] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:13824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.879 [2024-12-06 17:53:54.630696] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.879 [2024-12-06 17:53:54.635571] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.879 [2024-12-06 17:53:54.635822] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19328 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.879 [2024-12-06 17:53:54.635851] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.879 [2024-12-06 17:53:54.640675] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.879 [2024-12-06 17:53:54.640876] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:16672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.879 [2024-12-06 17:53:54.640904] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.879 [2024-12-06 17:53:54.645760] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.879 [2024-12-06 17:53:54.645942] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:1856 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.879 [2024-12-06 17:53:54.645970] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.879 [2024-12-06 17:53:54.650876] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.879 [2024-12-06 17:53:54.651113] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:15328 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.879 [2024-12-06 17:53:54.651141] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.879 [2024-12-06 17:53:54.656016] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.879 [2024-12-06 17:53:54.656238] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.879 [2024-12-06 17:53:54.656267] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.879 [2024-12-06 17:53:54.661236] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.879 [2024-12-06 17:53:54.661455] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:18528 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.879 [2024-12-06 17:53:54.661483] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.879 [2024-12-06 17:53:54.666343] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.879 [2024-12-06 17:53:54.666519] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:7904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.880 [2024-12-06 17:53:54.666547] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.880 [2024-12-06 17:53:54.671764] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.880 [2024-12-06 17:53:54.672003] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.880 [2024-12-06 17:53:54.672031] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.880 [2024-12-06 17:53:54.676969] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.880 [2024-12-06 17:53:54.677209] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:4640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.880 [2024-12-06 17:53:54.677238] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.880 [2024-12-06 17:53:54.682157] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.880 [2024-12-06 17:53:54.682354] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:2144 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.880 [2024-12-06 17:53:54.682382] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.880 [2024-12-06 17:53:54.687410] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.880 [2024-12-06 17:53:54.687626] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.880 [2024-12-06 17:53:54.687654] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.880 [2024-12-06 17:53:54.692671] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.880 [2024-12-06 17:53:54.692845] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:16544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.880 [2024-12-06 17:53:54.692873] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:12.880 [2024-12-06 17:53:54.697848] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.880 [2024-12-06 17:53:54.698036] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:18016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.880 [2024-12-06 17:53:54.698065] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:12.880 [2024-12-06 17:53:54.702991] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.880 [2024-12-06 17:53:54.703236] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.880 [2024-12-06 17:53:54.703265] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:12.880 [2024-12-06 17:53:54.708178] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.880 [2024-12-06 17:53:54.708403] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.880 [2024-12-06 17:53:54.708440] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:12.880 [2024-12-06 17:53:54.713185] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:12.880 [2024-12-06 17:53:54.713317] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:7840 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:12.880 [2024-12-06 17:53:54.713345] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.139 [2024-12-06 17:53:54.717759] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.139 [2024-12-06 17:53:54.717901] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:19520 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.139 [2024-12-06 17:53:54.717930] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.139 [2024-12-06 17:53:54.722928] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.139 [2024-12-06 17:53:54.723115] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:15424 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.139 [2024-12-06 17:53:54.723143] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.139 [2024-12-06 17:53:54.727625] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.139 [2024-12-06 17:53:54.727787] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:16864 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.139 [2024-12-06 17:53:54.727816] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.139 [2024-12-06 17:53:54.732746] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.139 [2024-12-06 17:53:54.732938] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:15872 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.139 [2024-12-06 17:53:54.732967] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.139 [2024-12-06 17:53:54.737830] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.139 [2024-12-06 17:53:54.737996] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.139 [2024-12-06 17:53:54.738025] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.139 [2024-12-06 17:53:54.743041] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.139 [2024-12-06 17:53:54.743208] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:22816 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.139 [2024-12-06 17:53:54.743237] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.139 [2024-12-06 17:53:54.748208] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.139 [2024-12-06 17:53:54.748411] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:7872 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.139 [2024-12-06 17:53:54.748439] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.139 [2024-12-06 17:53:54.753271] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.139 [2024-12-06 17:53:54.753488] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.139 [2024-12-06 17:53:54.753517] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.139 [2024-12-06 17:53:54.758393] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.139 [2024-12-06 17:53:54.758565] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:3904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.139 [2024-12-06 17:53:54.758593] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.139 [2024-12-06 17:53:54.763593] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.139 [2024-12-06 17:53:54.763788] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:10688 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.139 [2024-12-06 17:53:54.763816] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.139 [2024-12-06 17:53:54.768825] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.139 [2024-12-06 17:53:54.768965] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:5088 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.139 [2024-12-06 17:53:54.768993] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.139 [2024-12-06 17:53:54.774010] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.139 [2024-12-06 17:53:54.774191] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:11680 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.139 [2024-12-06 17:53:54.774220] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.139 [2024-12-06 17:53:54.779135] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.139 [2024-12-06 17:53:54.779283] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:23264 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.139 [2024-12-06 17:53:54.779311] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.139 5931.00 IOPS, 741.38 MiB/s [2024-12-06T16:53:54.978Z] [2024-12-06 17:53:54.785933] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.139 [2024-12-06 17:53:54.786112] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:19360 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.139 [2024-12-06 17:53:54.786140] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.139 [2024-12-06 17:53:54.791011] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.139 [2024-12-06 17:53:54.791119] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24704 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.139 [2024-12-06 17:53:54.791147] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.139 [2024-12-06 17:53:54.796039] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.139 [2024-12-06 17:53:54.796107] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:4416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.139 [2024-12-06 17:53:54.796134] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.139 [2024-12-06 17:53:54.800848] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.139 [2024-12-06 17:53:54.800926] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:3264 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.139 [2024-12-06 17:53:54.800956] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.139 [2024-12-06 17:53:54.805656] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.139 [2024-12-06 17:53:54.805789] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24224 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.139 [2024-12-06 17:53:54.805818] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.139 [2024-12-06 17:53:54.811206] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.139 [2024-12-06 17:53:54.811371] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:11712 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.811399] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.815845] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.815974] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:5376 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.816003] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.820368] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.820513] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:9504 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.820541] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.824599] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.824775] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:18528 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.824803] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.828833] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.828981] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:16064 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.829010] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.832983] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.833127] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:12672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.833156] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.837156] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.837302] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:14912 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.837336] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.841324] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.841479] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:5216 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.841508] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.845434] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.845590] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:8384 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.845619] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.849629] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.849799] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.849828] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.853764] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.853904] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:17408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.853932] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.857962] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.858127] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:19616 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.858154] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.862181] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.862314] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:1824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.862342] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.866425] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.866554] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:18912 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.866582] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.870613] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.870745] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:1280 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.870773] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.874886] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.874979] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:2208 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.875006] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.879031] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.879155] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:25280 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.879183] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.883601] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.883762] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:10496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.883790] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.888292] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.888451] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:10528 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.888480] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.893592] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.893725] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:22016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.893754] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.898044] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.898179] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:11008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.898207] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.902243] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.902411] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:22464 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.902439] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.906454] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.906570] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:8896 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.906598] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.910709] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.910845] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:1184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.910873] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.914997] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.915145] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:15648 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.915173] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.919142] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.919282] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:15520 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.919311] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.923358] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.923481] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:512 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.923508] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.927511] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.927643] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.927678] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.931691] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.931837] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:6144 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.931865] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.935793] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.935927] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:20576 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.935954] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.939905] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.940036] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:3328 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.940064] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.944056] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.944144] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:7680 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.944171] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.948297] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.948423] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.948457] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.952420] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.952540] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:9696 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.952567] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.956677] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.956788] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:15360 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.956816] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.960869] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.961003] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:2624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.961031] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.965006] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.965172] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:20320 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.965200] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.140 [2024-12-06 17:53:54.969430] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.140 [2024-12-06 17:53:54.969679] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.140 [2024-12-06 17:53:54.969708] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.141 [2024-12-06 17:53:54.974606] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.141 [2024-12-06 17:53:54.974817] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:16672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.141 [2024-12-06 17:53:54.974846] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.400 [2024-12-06 17:53:54.979760] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.400 [2024-12-06 17:53:54.980020] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:25152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.400 [2024-12-06 17:53:54.980048] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.400 [2024-12-06 17:53:54.985288] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.400 [2024-12-06 17:53:54.985482] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14432 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.400 [2024-12-06 17:53:54.985510] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.400 [2024-12-06 17:53:54.990709] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.400 [2024-12-06 17:53:54.990948] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:1696 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.400 [2024-12-06 17:53:54.990977] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.400 [2024-12-06 17:53:54.996764] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.400 [2024-12-06 17:53:54.996939] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:1184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.400 [2024-12-06 17:53:54.996967] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.400 [2024-12-06 17:53:55.002006] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.400 [2024-12-06 17:53:55.002221] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.400 [2024-12-06 17:53:55.002249] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.400 [2024-12-06 17:53:55.007149] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.400 [2024-12-06 17:53:55.007342] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:17152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.400 [2024-12-06 17:53:55.007370] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.400 [2024-12-06 17:53:55.012370] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.400 [2024-12-06 17:53:55.012555] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.400 [2024-12-06 17:53:55.012584] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.400 [2024-12-06 17:53:55.017615] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.400 [2024-12-06 17:53:55.017765] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.400 [2024-12-06 17:53:55.017797] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.400 [2024-12-06 17:53:55.023265] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.400 [2024-12-06 17:53:55.023467] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:4832 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.400 [2024-12-06 17:53:55.023495] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.400 [2024-12-06 17:53:55.028875] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.400 [2024-12-06 17:53:55.029062] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:11424 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.400 [2024-12-06 17:53:55.029090] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.400 [2024-12-06 17:53:55.034017] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.400 [2024-12-06 17:53:55.034266] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:2016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.400 [2024-12-06 17:53:55.034294] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.400 [2024-12-06 17:53:55.039156] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.400 [2024-12-06 17:53:55.039334] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:9600 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.400 [2024-12-06 17:53:55.039363] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.400 [2024-12-06 17:53:55.044293] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.400 [2024-12-06 17:53:55.044541] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:7392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.400 [2024-12-06 17:53:55.044569] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.400 [2024-12-06 17:53:55.049439] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.400 [2024-12-06 17:53:55.049608] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:3104 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.400 [2024-12-06 17:53:55.049636] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.400 [2024-12-06 17:53:55.054585] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.400 [2024-12-06 17:53:55.054835] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:10176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.400 [2024-12-06 17:53:55.054868] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.400 [2024-12-06 17:53:55.059652] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.400 [2024-12-06 17:53:55.059799] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:7456 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.400 [2024-12-06 17:53:55.059827] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.400 [2024-12-06 17:53:55.064378] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.400 [2024-12-06 17:53:55.064594] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:13696 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.400 [2024-12-06 17:53:55.064621] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.400 [2024-12-06 17:53:55.069366] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.400 [2024-12-06 17:53:55.069530] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.400 [2024-12-06 17:53:55.069558] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.400 [2024-12-06 17:53:55.075391] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.400 [2024-12-06 17:53:55.075620] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.400 [2024-12-06 17:53:55.075648] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.400 [2024-12-06 17:53:55.080413] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.400 [2024-12-06 17:53:55.080560] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:21504 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.400 [2024-12-06 17:53:55.080593] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.400 [2024-12-06 17:53:55.084788] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.400 [2024-12-06 17:53:55.084913] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:1632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.400 [2024-12-06 17:53:55.084940] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.400 [2024-12-06 17:53:55.088953] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.400 [2024-12-06 17:53:55.089065] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:21600 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.400 [2024-12-06 17:53:55.089093] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.400 [2024-12-06 17:53:55.093359] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.401 [2024-12-06 17:53:55.093493] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:6016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.401 [2024-12-06 17:53:55.093521] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.401 [2024-12-06 17:53:55.098728] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.401 [2024-12-06 17:53:55.098828] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:15392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.401 [2024-12-06 17:53:55.098857] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.401 [2024-12-06 17:53:55.102992] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.401 [2024-12-06 17:53:55.103108] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:12096 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.401 [2024-12-06 17:53:55.103137] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.401 [2024-12-06 17:53:55.107171] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.401 [2024-12-06 17:53:55.107295] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:9536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.401 [2024-12-06 17:53:55.107323] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.401 [2024-12-06 17:53:55.111550] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.401 [2024-12-06 17:53:55.111696] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20160 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.401 [2024-12-06 17:53:55.111732] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.401 [2024-12-06 17:53:55.115797] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.401 [2024-12-06 17:53:55.115910] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24000 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.401 [2024-12-06 17:53:55.115938] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.401 [2024-12-06 17:53:55.120333] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.401 [2024-12-06 17:53:55.120425] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:8672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.401 [2024-12-06 17:53:55.120451] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.401 [2024-12-06 17:53:55.124783] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.401 [2024-12-06 17:53:55.124899] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:1280 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.401 [2024-12-06 17:53:55.124927] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.401 [2024-12-06 17:53:55.129242] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.401 [2024-12-06 17:53:55.129370] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:25024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.401 [2024-12-06 17:53:55.129397] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.401 [2024-12-06 17:53:55.133623] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.401 [2024-12-06 17:53:55.133726] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:17120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.401 [2024-12-06 17:53:55.133754] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.401 [2024-12-06 17:53:55.138099] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.401 [2024-12-06 17:53:55.138195] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14880 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.401 [2024-12-06 17:53:55.138226] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.401 [2024-12-06 17:53:55.142454] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.401 [2024-12-06 17:53:55.142547] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:4320 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.401 [2024-12-06 17:53:55.142573] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.401 [2024-12-06 17:53:55.146855] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.401 [2024-12-06 17:53:55.146981] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19264 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.401 [2024-12-06 17:53:55.147009] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.401 [2024-12-06 17:53:55.151314] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.401 [2024-12-06 17:53:55.151428] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:32 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.401 [2024-12-06 17:53:55.151456] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.401 [2024-12-06 17:53:55.155800] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.401 [2024-12-06 17:53:55.155887] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24896 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.401 [2024-12-06 17:53:55.155915] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.401 [2024-12-06 17:53:55.160438] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.401 [2024-12-06 17:53:55.160522] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:13280 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.401 [2024-12-06 17:53:55.160549] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.401 [2024-12-06 17:53:55.165114] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.401 [2024-12-06 17:53:55.165206] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:18432 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.401 [2024-12-06 17:53:55.165233] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.401 [2024-12-06 17:53:55.169638] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.401 [2024-12-06 17:53:55.169757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:8992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.401 [2024-12-06 17:53:55.169787] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.401 [2024-12-06 17:53:55.174165] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.401 [2024-12-06 17:53:55.174276] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:2848 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.401 [2024-12-06 17:53:55.174305] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.401 [2024-12-06 17:53:55.178632] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.401 [2024-12-06 17:53:55.178753] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:3904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.401 [2024-12-06 17:53:55.178781] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.401 [2024-12-06 17:53:55.183214] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.401 [2024-12-06 17:53:55.183312] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:4768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.401 [2024-12-06 17:53:55.183340] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.401 [2024-12-06 17:53:55.187829] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.401 [2024-12-06 17:53:55.187911] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:4256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.401 [2024-12-06 17:53:55.187938] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.401 [2024-12-06 17:53:55.192477] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.401 [2024-12-06 17:53:55.192582] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:9600 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.401 [2024-12-06 17:53:55.192610] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.401 [2024-12-06 17:53:55.196877] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.401 [2024-12-06 17:53:55.196983] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:9760 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.401 [2024-12-06 17:53:55.197017] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.401 [2024-12-06 17:53:55.201490] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.401 [2024-12-06 17:53:55.201588] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:17344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.401 [2024-12-06 17:53:55.201616] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.401 [2024-12-06 17:53:55.206102] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.401 [2024-12-06 17:53:55.206194] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:4736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.401 [2024-12-06 17:53:55.206220] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.401 [2024-12-06 17:53:55.210716] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.401 [2024-12-06 17:53:55.210808] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:21536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.402 [2024-12-06 17:53:55.210836] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.402 [2024-12-06 17:53:55.215547] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.402 [2024-12-06 17:53:55.215676] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19936 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.402 [2024-12-06 17:53:55.215704] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.402 [2024-12-06 17:53:55.220117] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.402 [2024-12-06 17:53:55.220221] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:7744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.402 [2024-12-06 17:53:55.220249] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.402 [2024-12-06 17:53:55.224651] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.402 [2024-12-06 17:53:55.224788] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:5344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.402 [2024-12-06 17:53:55.224817] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.402 [2024-12-06 17:53:55.229361] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.402 [2024-12-06 17:53:55.229522] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23680 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.402 [2024-12-06 17:53:55.229549] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.402 [2024-12-06 17:53:55.234217] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.402 [2024-12-06 17:53:55.234319] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:12704 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.402 [2024-12-06 17:53:55.234347] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.660 [2024-12-06 17:53:55.238498] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.660 [2024-12-06 17:53:55.238625] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:2656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.660 [2024-12-06 17:53:55.238660] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.660 [2024-12-06 17:53:55.242888] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.660 [2024-12-06 17:53:55.242995] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:6688 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.660 [2024-12-06 17:53:55.243023] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.660 [2024-12-06 17:53:55.247347] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.660 [2024-12-06 17:53:55.247477] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:1664 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.660 [2024-12-06 17:53:55.247505] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.660 [2024-12-06 17:53:55.251867] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.660 [2024-12-06 17:53:55.252037] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:12896 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.660 [2024-12-06 17:53:55.252065] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.660 [2024-12-06 17:53:55.256180] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.660 [2024-12-06 17:53:55.256332] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.660 [2024-12-06 17:53:55.256360] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.660 [2024-12-06 17:53:55.260687] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.660 [2024-12-06 17:53:55.260829] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:8768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.660 [2024-12-06 17:53:55.260857] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.660 [2024-12-06 17:53:55.265075] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.660 [2024-12-06 17:53:55.265231] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:22432 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.660 [2024-12-06 17:53:55.265258] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.660 [2024-12-06 17:53:55.269494] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.660 [2024-12-06 17:53:55.269653] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19328 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.660 [2024-12-06 17:53:55.269691] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.660 [2024-12-06 17:53:55.273989] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.660 [2024-12-06 17:53:55.274135] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23968 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.660 [2024-12-06 17:53:55.274162] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.660 [2024-12-06 17:53:55.278569] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.660 [2024-12-06 17:53:55.278703] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14240 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.660 [2024-12-06 17:53:55.278731] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.660 [2024-12-06 17:53:55.282900] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.660 [2024-12-06 17:53:55.283012] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:928 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.660 [2024-12-06 17:53:55.283039] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.660 [2024-12-06 17:53:55.287401] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.660 [2024-12-06 17:53:55.287634] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:15904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.660 [2024-12-06 17:53:55.287662] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.660 [2024-12-06 17:53:55.292576] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.660 [2024-12-06 17:53:55.292730] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:13312 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.660 [2024-12-06 17:53:55.292759] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.660 [2024-12-06 17:53:55.297682] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.660 [2024-12-06 17:53:55.297897] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:2656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.660 [2024-12-06 17:53:55.297925] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.660 [2024-12-06 17:53:55.303389] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.660 [2024-12-06 17:53:55.303525] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:17472 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.660 [2024-12-06 17:53:55.303552] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.660 [2024-12-06 17:53:55.308974] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.660 [2024-12-06 17:53:55.309126] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.660 [2024-12-06 17:53:55.309155] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.660 [2024-12-06 17:53:55.313625] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.660 [2024-12-06 17:53:55.313766] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:5888 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.660 [2024-12-06 17:53:55.313795] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.661 [2024-12-06 17:53:55.317855] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.661 [2024-12-06 17:53:55.318008] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:5280 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.661 [2024-12-06 17:53:55.318036] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.661 [2024-12-06 17:53:55.322030] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.661 [2024-12-06 17:53:55.322141] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:10560 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.661 [2024-12-06 17:53:55.322169] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.661 [2024-12-06 17:53:55.326193] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.661 [2024-12-06 17:53:55.326299] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:9920 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.661 [2024-12-06 17:53:55.326327] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.661 [2024-12-06 17:53:55.330480] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.661 [2024-12-06 17:53:55.330575] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:8512 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.661 [2024-12-06 17:53:55.330604] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.661 [2024-12-06 17:53:55.335472] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.661 [2024-12-06 17:53:55.335591] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24224 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.661 [2024-12-06 17:53:55.335619] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.661 [2024-12-06 17:53:55.340821] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.661 [2024-12-06 17:53:55.341045] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:1056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.661 [2024-12-06 17:53:55.341073] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.661 [2024-12-06 17:53:55.346524] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.661 [2024-12-06 17:53:55.346656] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:11552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.661 [2024-12-06 17:53:55.346692] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.661 [2024-12-06 17:53:55.352010] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.661 [2024-12-06 17:53:55.352221] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:928 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.661 [2024-12-06 17:53:55.352249] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.661 [2024-12-06 17:53:55.357035] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.661 [2024-12-06 17:53:55.357211] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:15008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.661 [2024-12-06 17:53:55.357240] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.661 [2024-12-06 17:53:55.361862] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.661 [2024-12-06 17:53:55.361964] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:23712 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.661 [2024-12-06 17:53:55.361997] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.661 [2024-12-06 17:53:55.366553] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.661 [2024-12-06 17:53:55.366727] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:15776 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.661 [2024-12-06 17:53:55.366755] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.661 [2024-12-06 17:53:55.371682] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.661 [2024-12-06 17:53:55.371889] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.661 [2024-12-06 17:53:55.371917] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.661 [2024-12-06 17:53:55.376815] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.661 [2024-12-06 17:53:55.377001] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19616 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.661 [2024-12-06 17:53:55.377030] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.661 [2024-12-06 17:53:55.381888] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.661 [2024-12-06 17:53:55.382089] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.661 [2024-12-06 17:53:55.382117] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.661 [2024-12-06 17:53:55.387028] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.661 [2024-12-06 17:53:55.387165] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:4000 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.661 [2024-12-06 17:53:55.387192] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.661 [2024-12-06 17:53:55.392059] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.661 [2024-12-06 17:53:55.392246] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:4416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.661 [2024-12-06 17:53:55.392275] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.661 [2024-12-06 17:53:55.397159] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.661 [2024-12-06 17:53:55.397334] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:16096 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.661 [2024-12-06 17:53:55.397363] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.661 [2024-12-06 17:53:55.402211] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.661 [2024-12-06 17:53:55.402410] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:18272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.661 [2024-12-06 17:53:55.402438] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.661 [2024-12-06 17:53:55.407363] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.661 [2024-12-06 17:53:55.407596] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:3168 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.661 [2024-12-06 17:53:55.407624] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.661 [2024-12-06 17:53:55.412452] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.661 [2024-12-06 17:53:55.412700] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:15360 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.661 [2024-12-06 17:53:55.412736] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.661 [2024-12-06 17:53:55.417572] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.661 [2024-12-06 17:53:55.417781] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:7648 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.661 [2024-12-06 17:53:55.417815] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.661 [2024-12-06 17:53:55.422643] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.661 [2024-12-06 17:53:55.422867] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:6912 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.661 [2024-12-06 17:53:55.422896] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.661 [2024-12-06 17:53:55.427740] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.661 [2024-12-06 17:53:55.427941] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:5856 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.661 [2024-12-06 17:53:55.427968] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.661 [2024-12-06 17:53:55.432879] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.661 [2024-12-06 17:53:55.433114] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:10624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.661 [2024-12-06 17:53:55.433141] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.661 [2024-12-06 17:53:55.438011] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.661 [2024-12-06 17:53:55.438225] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:8000 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.661 [2024-12-06 17:53:55.438253] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.661 [2024-12-06 17:53:55.443085] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.661 [2024-12-06 17:53:55.443308] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:11040 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.661 [2024-12-06 17:53:55.443337] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.661 [2024-12-06 17:53:55.448291] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.661 [2024-12-06 17:53:55.448512] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:4192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.661 [2024-12-06 17:53:55.448540] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.661 [2024-12-06 17:53:55.453357] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.662 [2024-12-06 17:53:55.453632] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:13440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.662 [2024-12-06 17:53:55.453661] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.662 [2024-12-06 17:53:55.458452] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.662 [2024-12-06 17:53:55.458722] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:1920 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.662 [2024-12-06 17:53:55.458751] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.662 [2024-12-06 17:53:55.463621] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.662 [2024-12-06 17:53:55.463800] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:4384 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.662 [2024-12-06 17:53:55.463828] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.662 [2024-12-06 17:53:55.468653] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.662 [2024-12-06 17:53:55.468868] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20512 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.662 [2024-12-06 17:53:55.468897] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.662 [2024-12-06 17:53:55.474006] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.662 [2024-12-06 17:53:55.474213] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:5536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.662 [2024-12-06 17:53:55.474243] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.662 [2024-12-06 17:53:55.479161] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.662 [2024-12-06 17:53:55.479416] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:12544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.662 [2024-12-06 17:53:55.479444] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.662 [2024-12-06 17:53:55.484240] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.662 [2024-12-06 17:53:55.484488] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:12576 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.662 [2024-12-06 17:53:55.484516] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.662 [2024-12-06 17:53:55.489266] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.662 [2024-12-06 17:53:55.489461] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:21312 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.662 [2024-12-06 17:53:55.489489] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.662 [2024-12-06 17:53:55.494466] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.662 [2024-12-06 17:53:55.494656] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:10176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.662 [2024-12-06 17:53:55.494698] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.921 [2024-12-06 17:53:55.499581] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.921 [2024-12-06 17:53:55.499861] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:10144 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.921 [2024-12-06 17:53:55.499890] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.921 [2024-12-06 17:53:55.504650] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.921 [2024-12-06 17:53:55.504956] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:15488 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.921 [2024-12-06 17:53:55.504984] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.921 [2024-12-06 17:53:55.509788] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.921 [2024-12-06 17:53:55.510023] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:9440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.921 [2024-12-06 17:53:55.510051] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.921 [2024-12-06 17:53:55.514869] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.921 [2024-12-06 17:53:55.515098] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:25408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.921 [2024-12-06 17:53:55.515127] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.921 [2024-12-06 17:53:55.520181] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.921 [2024-12-06 17:53:55.520441] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23968 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.921 [2024-12-06 17:53:55.520470] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.921 [2024-12-06 17:53:55.525180] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.921 [2024-12-06 17:53:55.525408] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:1792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.921 [2024-12-06 17:53:55.525437] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.921 [2024-12-06 17:53:55.530247] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.921 [2024-12-06 17:53:55.530475] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:12896 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.921 [2024-12-06 17:53:55.530504] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.921 [2024-12-06 17:53:55.535427] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.921 [2024-12-06 17:53:55.535685] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19232 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.921 [2024-12-06 17:53:55.535714] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.921 [2024-12-06 17:53:55.540504] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.921 [2024-12-06 17:53:55.540775] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:4928 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.921 [2024-12-06 17:53:55.540803] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.921 [2024-12-06 17:53:55.545540] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.921 [2024-12-06 17:53:55.545636] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:19040 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.921 [2024-12-06 17:53:55.545671] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.921 [2024-12-06 17:53:55.550532] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.921 [2024-12-06 17:53:55.550716] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:3744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.921 [2024-12-06 17:53:55.550744] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.921 [2024-12-06 17:53:55.555566] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.921 [2024-12-06 17:53:55.555685] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:9216 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.921 [2024-12-06 17:53:55.555712] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.921 [2024-12-06 17:53:55.560819] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.921 [2024-12-06 17:53:55.561013] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:17088 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.921 [2024-12-06 17:53:55.561042] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.921 [2024-12-06 17:53:55.565910] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.921 [2024-12-06 17:53:55.566092] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:16960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.921 [2024-12-06 17:53:55.566120] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.921 [2024-12-06 17:53:55.571150] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.921 [2024-12-06 17:53:55.571303] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:12096 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.921 [2024-12-06 17:53:55.571331] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.921 [2024-12-06 17:53:55.576238] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.921 [2024-12-06 17:53:55.576381] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:22656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.921 [2024-12-06 17:53:55.576409] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.921 [2024-12-06 17:53:55.581302] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.921 [2024-12-06 17:53:55.581457] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:20512 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.921 [2024-12-06 17:53:55.581485] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.921 [2024-12-06 17:53:55.586401] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.921 [2024-12-06 17:53:55.586539] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:4192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.921 [2024-12-06 17:53:55.586567] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.921 [2024-12-06 17:53:55.591454] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.921 [2024-12-06 17:53:55.591615] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.921 [2024-12-06 17:53:55.591644] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.921 [2024-12-06 17:53:55.596524] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.921 [2024-12-06 17:53:55.596711] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:9504 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.921 [2024-12-06 17:53:55.596739] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.921 [2024-12-06 17:53:55.601628] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.922 [2024-12-06 17:53:55.601769] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:21952 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.922 [2024-12-06 17:53:55.601798] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.922 [2024-12-06 17:53:55.606721] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.922 [2024-12-06 17:53:55.606857] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:10336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.922 [2024-12-06 17:53:55.606885] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.922 [2024-12-06 17:53:55.611762] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.922 [2024-12-06 17:53:55.611945] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:11136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.922 [2024-12-06 17:53:55.611974] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.922 [2024-12-06 17:53:55.616752] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.922 [2024-12-06 17:53:55.616899] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:22752 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.922 [2024-12-06 17:53:55.616927] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.922 [2024-12-06 17:53:55.621904] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.922 [2024-12-06 17:53:55.622095] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:10144 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.922 [2024-12-06 17:53:55.622124] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.922 [2024-12-06 17:53:55.626976] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.922 [2024-12-06 17:53:55.627181] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:4128 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.922 [2024-12-06 17:53:55.627220] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.922 [2024-12-06 17:53:55.632000] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.922 [2024-12-06 17:53:55.632137] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:4992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.922 [2024-12-06 17:53:55.632165] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.922 [2024-12-06 17:53:55.637199] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.922 [2024-12-06 17:53:55.637346] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:22624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.922 [2024-12-06 17:53:55.637374] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.922 [2024-12-06 17:53:55.642333] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.922 [2024-12-06 17:53:55.642419] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:12320 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.922 [2024-12-06 17:53:55.642446] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.922 [2024-12-06 17:53:55.647399] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.922 [2024-12-06 17:53:55.647505] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:2784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.922 [2024-12-06 17:53:55.647532] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.922 [2024-12-06 17:53:55.652432] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.922 [2024-12-06 17:53:55.652592] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:11360 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.922 [2024-12-06 17:53:55.652620] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.922 [2024-12-06 17:53:55.657512] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.922 [2024-12-06 17:53:55.657673] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:22048 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.922 [2024-12-06 17:53:55.657702] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.922 [2024-12-06 17:53:55.662614] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.922 [2024-12-06 17:53:55.662757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:12128 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.922 [2024-12-06 17:53:55.662786] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.922 [2024-12-06 17:53:55.667704] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.922 [2024-12-06 17:53:55.667836] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:7904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.922 [2024-12-06 17:53:55.667864] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.922 [2024-12-06 17:53:55.672859] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.922 [2024-12-06 17:53:55.673031] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14720 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.922 [2024-12-06 17:53:55.673059] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.922 [2024-12-06 17:53:55.677906] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.922 [2024-12-06 17:53:55.678126] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:15072 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.922 [2024-12-06 17:53:55.678154] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.922 [2024-12-06 17:53:55.682996] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.922 [2024-12-06 17:53:55.683251] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:6016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.922 [2024-12-06 17:53:55.683280] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.922 [2024-12-06 17:53:55.688163] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.922 [2024-12-06 17:53:55.688328] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:10048 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.922 [2024-12-06 17:53:55.688356] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.922 [2024-12-06 17:53:55.693340] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.922 [2024-12-06 17:53:55.693543] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:25536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.922 [2024-12-06 17:53:55.693571] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.922 [2024-12-06 17:53:55.698408] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.922 [2024-12-06 17:53:55.698620] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:12672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.922 [2024-12-06 17:53:55.698650] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.922 [2024-12-06 17:53:55.703471] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.922 [2024-12-06 17:53:55.703704] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.922 [2024-12-06 17:53:55.703733] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.922 [2024-12-06 17:53:55.708624] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.922 [2024-12-06 17:53:55.708787] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14880 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.922 [2024-12-06 17:53:55.708816] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.922 [2024-12-06 17:53:55.713698] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.922 [2024-12-06 17:53:55.713861] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:12928 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.922 [2024-12-06 17:53:55.713889] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.922 [2024-12-06 17:53:55.718767] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.922 [2024-12-06 17:53:55.718938] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24224 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.922 [2024-12-06 17:53:55.718967] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.922 [2024-12-06 17:53:55.723867] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.922 [2024-12-06 17:53:55.724024] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.922 [2024-12-06 17:53:55.724052] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.922 [2024-12-06 17:53:55.728895] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.922 [2024-12-06 17:53:55.729105] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:22272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.922 [2024-12-06 17:53:55.729133] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.922 [2024-12-06 17:53:55.734007] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.922 [2024-12-06 17:53:55.734184] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:18176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.922 [2024-12-06 17:53:55.734213] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:13.922 [2024-12-06 17:53:55.739082] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.922 [2024-12-06 17:53:55.739258] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:9408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.923 [2024-12-06 17:53:55.739287] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:13.923 [2024-12-06 17:53:55.744297] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.923 [2024-12-06 17:53:55.744455] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:9536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.923 [2024-12-06 17:53:55.744485] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:13.923 [2024-12-06 17:53:55.749347] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.923 [2024-12-06 17:53:55.749548] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:21408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.923 [2024-12-06 17:53:55.749576] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:13.923 [2024-12-06 17:53:55.754373] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:13.923 [2024-12-06 17:53:55.754605] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:18464 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:13.923 [2024-12-06 17:53:55.754634] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:14.181 [2024-12-06 17:53:55.759496] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:14.181 [2024-12-06 17:53:55.759684] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:14496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:14.181 [2024-12-06 17:53:55.759719] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:14.181 [2024-12-06 17:53:55.764492] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:14.181 [2024-12-06 17:53:55.764638] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24352 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:14.181 [2024-12-06 17:53:55.764672] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:14.181 [2024-12-06 17:53:55.769685] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:14.181 [2024-12-06 17:53:55.769857] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:17312 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:14.181 [2024-12-06 17:53:55.769886] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:38:14.181 [2024-12-06 17:53:55.774784] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:14.181 [2024-12-06 17:53:55.774924] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:14.181 [2024-12-06 17:53:55.774953] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:14.181 [2024-12-06 17:53:55.779833] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:14.181 [2024-12-06 17:53:55.780006] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:13632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:14.181 [2024-12-06 17:53:55.780034] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:38:14.181 6179.50 IOPS, 772.44 MiB/s [2024-12-06T16:53:56.020Z] [2024-12-06 17:53:55.786046] tcp.c:2241:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x7b41b0) with pdu=0x200016eff3c8 00:38:14.181 [2024-12-06 17:53:55.786221] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:5792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:14.181 [2024-12-06 17:53:55.786250] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:38:14.181 00:38:14.181 Latency(us) 00:38:14.181 [2024-12-06T16:53:56.020Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:38:14.181 Job: nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 16, IO size: 131072) 00:38:14.181 nvme0n1 : 2.00 6175.78 771.97 0.00 0.00 2583.29 1953.94 7767.23 00:38:14.181 [2024-12-06T16:53:56.020Z] =================================================================================================================== 00:38:14.181 [2024-12-06T16:53:56.020Z] Total : 6175.78 771.97 0.00 0.00 2583.29 1953.94 7767.23 00:38:14.181 { 00:38:14.181 "results": [ 00:38:14.181 { 00:38:14.181 "job": "nvme0n1", 00:38:14.181 "core_mask": "0x2", 00:38:14.181 "workload": "randwrite", 00:38:14.181 "status": "finished", 00:38:14.181 "queue_depth": 16, 00:38:14.181 "io_size": 131072, 00:38:14.181 "runtime": 2.004605, 00:38:14.181 "iops": 6175.780265937678, 00:38:14.181 "mibps": 771.9725332422098, 00:38:14.181 "io_failed": 0, 00:38:14.181 "io_timeout": 0, 00:38:14.181 "avg_latency_us": 2583.288111529947, 00:38:14.181 "min_latency_us": 1953.9437037037037, 00:38:14.181 "max_latency_us": 7767.22962962963 00:38:14.181 } 00:38:14.181 ], 00:38:14.181 "core_count": 1 00:38:14.181 } 00:38:14.181 17:53:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # get_transient_errcount nvme0n1 00:38:14.181 17:53:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@27 -- # bperf_rpc bdev_get_iostat -b nvme0n1 00:38:14.181 17:53:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@28 -- # jq -r '.bdevs[0] 00:38:14.181 | .driver_specific 00:38:14.181 | .nvme_error 00:38:14.181 | .status_code 00:38:14.181 | .command_transient_transport_error' 00:38:14.181 17:53:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_get_iostat -b nvme0n1 00:38:14.439 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # (( 400 > 0 )) 00:38:14.439 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@73 -- # killprocess 407635 00:38:14.439 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # '[' -z 407635 ']' 00:38:14.439 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@958 -- # kill -0 407635 00:38:14.439 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@959 -- # uname 00:38:14.439 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:38:14.439 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 407635 00:38:14.439 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:38:14.439 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:38:14.439 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@972 -- # echo 'killing process with pid 407635' 00:38:14.439 killing process with pid 407635 00:38:14.439 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@973 -- # kill 407635 00:38:14.439 Received shutdown signal, test time was about 2.000000 seconds 00:38:14.439 00:38:14.439 Latency(us) 00:38:14.439 [2024-12-06T16:53:56.278Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:38:14.439 [2024-12-06T16:53:56.278Z] =================================================================================================================== 00:38:14.439 [2024-12-06T16:53:56.278Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:38:14.439 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@978 -- # wait 407635 00:38:14.697 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@116 -- # killprocess 406247 00:38:14.697 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # '[' -z 406247 ']' 00:38:14.697 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@958 -- # kill -0 406247 00:38:14.697 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@959 -- # uname 00:38:14.697 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:38:14.697 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 406247 00:38:14.697 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:38:14.697 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:38:14.697 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@972 -- # echo 'killing process with pid 406247' 00:38:14.697 killing process with pid 406247 00:38:14.697 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@973 -- # kill 406247 00:38:14.697 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@978 -- # wait 406247 00:38:14.957 00:38:14.957 real 0m15.368s 00:38:14.957 user 0m30.964s 00:38:14.957 sys 0m4.302s 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@1130 -- # xtrace_disable 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:38:14.957 ************************************ 00:38:14.957 END TEST nvmf_digest_error 00:38:14.957 ************************************ 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@149 -- # trap - SIGINT SIGTERM EXIT 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@150 -- # nvmftestfini 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@516 -- # nvmfcleanup 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@121 -- # sync 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@124 -- # set +e 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@125 -- # for i in {1..20} 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:38:14.957 rmmod nvme_tcp 00:38:14.957 rmmod nvme_fabrics 00:38:14.957 rmmod nvme_keyring 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@128 -- # set -e 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@129 -- # return 0 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@517 -- # '[' -n 406247 ']' 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@518 -- # killprocess 406247 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@954 -- # '[' -z 406247 ']' 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@958 -- # kill -0 406247 00:38:14.957 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 958: kill: (406247) - No such process 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@981 -- # echo 'Process with pid 406247 is not found' 00:38:14.957 Process with pid 406247 is not found 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@297 -- # iptr 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@791 -- # iptables-save 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@791 -- # iptables-restore 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@302 -- # remove_spdk_ns 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:38:14.957 17:53:56 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:38:16.865 17:53:58 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:38:16.865 00:38:16.865 real 0m35.314s 00:38:16.865 user 1m2.652s 00:38:16.865 sys 0m10.248s 00:38:16.865 17:53:58 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1130 -- # xtrace_disable 00:38:16.865 17:53:58 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@10 -- # set +x 00:38:16.865 ************************************ 00:38:16.865 END TEST nvmf_digest 00:38:16.865 ************************************ 00:38:17.124 17:53:58 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@36 -- # [[ 0 -eq 1 ]] 00:38:17.124 17:53:58 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@41 -- # [[ 0 -eq 1 ]] 00:38:17.124 17:53:58 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@46 -- # [[ phy == phy ]] 00:38:17.124 17:53:58 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@47 -- # run_test nvmf_bdevperf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/bdevperf.sh --transport=tcp 00:38:17.124 17:53:58 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:38:17.124 17:53:58 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:38:17.124 17:53:58 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:38:17.124 ************************************ 00:38:17.124 START TEST nvmf_bdevperf 00:38:17.124 ************************************ 00:38:17.124 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/bdevperf.sh --transport=tcp 00:38:17.124 * Looking for test storage... 00:38:17.124 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:38:17.124 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:38:17.124 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1711 -- # lcov --version 00:38:17.124 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:38:17.124 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:38:17.124 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:38:17.124 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@333 -- # local ver1 ver1_l 00:38:17.124 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@334 -- # local ver2 ver2_l 00:38:17.124 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@336 -- # IFS=.-: 00:38:17.124 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@336 -- # read -ra ver1 00:38:17.124 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@337 -- # IFS=.-: 00:38:17.124 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@337 -- # read -ra ver2 00:38:17.124 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@338 -- # local 'op=<' 00:38:17.124 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@340 -- # ver1_l=2 00:38:17.124 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@341 -- # ver2_l=1 00:38:17.124 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:38:17.124 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@344 -- # case "$op" in 00:38:17.124 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@345 -- # : 1 00:38:17.124 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@364 -- # (( v = 0 )) 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@365 -- # decimal 1 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@353 -- # local d=1 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@355 -- # echo 1 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@365 -- # ver1[v]=1 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@366 -- # decimal 2 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@353 -- # local d=2 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@355 -- # echo 2 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@366 -- # ver2[v]=2 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@368 -- # return 0 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:38:17.125 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:17.125 --rc genhtml_branch_coverage=1 00:38:17.125 --rc genhtml_function_coverage=1 00:38:17.125 --rc genhtml_legend=1 00:38:17.125 --rc geninfo_all_blocks=1 00:38:17.125 --rc geninfo_unexecuted_blocks=1 00:38:17.125 00:38:17.125 ' 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:38:17.125 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:17.125 --rc genhtml_branch_coverage=1 00:38:17.125 --rc genhtml_function_coverage=1 00:38:17.125 --rc genhtml_legend=1 00:38:17.125 --rc geninfo_all_blocks=1 00:38:17.125 --rc geninfo_unexecuted_blocks=1 00:38:17.125 00:38:17.125 ' 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:38:17.125 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:17.125 --rc genhtml_branch_coverage=1 00:38:17.125 --rc genhtml_function_coverage=1 00:38:17.125 --rc genhtml_legend=1 00:38:17.125 --rc geninfo_all_blocks=1 00:38:17.125 --rc geninfo_unexecuted_blocks=1 00:38:17.125 00:38:17.125 ' 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:38:17.125 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:17.125 --rc genhtml_branch_coverage=1 00:38:17.125 --rc genhtml_function_coverage=1 00:38:17.125 --rc genhtml_legend=1 00:38:17.125 --rc geninfo_all_blocks=1 00:38:17.125 --rc geninfo_unexecuted_blocks=1 00:38:17.125 00:38:17.125 ' 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@7 -- # uname -s 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@15 -- # shopt -s extglob 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- paths/export.sh@5 -- # export PATH 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@51 -- # : 0 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:38:17.125 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@55 -- # have_pci_nics=0 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@11 -- # MALLOC_BDEV_SIZE=64 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@24 -- # nvmftestinit 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@476 -- # prepare_net_devs 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@438 -- # local -g is_hw=no 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@440 -- # remove_spdk_ns 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@309 -- # xtrace_disable 00:38:17.125 17:53:58 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@315 -- # pci_devs=() 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@315 -- # local -a pci_devs 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@316 -- # pci_net_devs=() 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@317 -- # pci_drivers=() 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@317 -- # local -A pci_drivers 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@319 -- # net_devs=() 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@319 -- # local -ga net_devs 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@320 -- # e810=() 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@320 -- # local -ga e810 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@321 -- # x722=() 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@321 -- # local -ga x722 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@322 -- # mlx=() 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@322 -- # local -ga mlx 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:38:19.657 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:38:19.657 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@418 -- # [[ up == up ]] 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:38:19.657 Found net devices under 0000:0a:00.0: cvl_0_0 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@418 -- # [[ up == up ]] 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:38:19.657 Found net devices under 0000:0a:00.1: cvl_0_1 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@442 -- # is_hw=yes 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:38:19.657 17:54:00 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:38:19.657 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:38:19.657 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:38:19.657 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:38:19.657 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:38:19.657 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:38:19.657 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:38:19.657 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:38:19.657 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:38:19.657 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:38:19.657 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.218 ms 00:38:19.657 00:38:19.657 --- 10.0.0.2 ping statistics --- 00:38:19.657 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:38:19.657 rtt min/avg/max/mdev = 0.218/0.218/0.218/0.000 ms 00:38:19.657 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:38:19.657 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:38:19.657 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.186 ms 00:38:19.657 00:38:19.657 --- 10.0.0.1 ping statistics --- 00:38:19.657 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:38:19.657 rtt min/avg/max/mdev = 0.186/0.186/0.186/0.000 ms 00:38:19.657 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:38:19.657 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@450 -- # return 0 00:38:19.657 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:38:19.657 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:38:19.657 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:38:19.657 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:38:19.657 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:38:19.657 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@25 -- # tgt_init 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@15 -- # nvmfappstart -m 0xE 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@726 -- # xtrace_disable 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@509 -- # nvmfpid=410070 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@510 -- # waitforlisten 410070 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@835 -- # '[' -z 410070 ']' 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@840 -- # local max_retries=100 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:38:19.658 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@844 -- # xtrace_disable 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:38:19.658 [2024-12-06 17:54:01.154946] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:38:19.658 [2024-12-06 17:54:01.155025] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:38:19.658 [2024-12-06 17:54:01.229533] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:38:19.658 [2024-12-06 17:54:01.277325] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:38:19.658 [2024-12-06 17:54:01.277390] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:38:19.658 [2024-12-06 17:54:01.277419] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:38:19.658 [2024-12-06 17:54:01.277431] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:38:19.658 [2024-12-06 17:54:01.277440] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:38:19.658 [2024-12-06 17:54:01.278984] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:38:19.658 [2024-12-06 17:54:01.279052] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:38:19.658 [2024-12-06 17:54:01.279056] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@868 -- # return 0 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@732 -- # xtrace_disable 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:38:19.658 [2024-12-06 17:54:01.420692] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@18 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:38:19.658 Malloc0 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@19 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@20 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@21 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:38:19.658 [2024-12-06 17:54:01.483511] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -q 128 -o 4096 -w verify -t 1 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@27 -- # gen_nvmf_target_json 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@560 -- # config=() 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@560 -- # local subsystem config 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:38:19.658 { 00:38:19.658 "params": { 00:38:19.658 "name": "Nvme$subsystem", 00:38:19.658 "trtype": "$TEST_TRANSPORT", 00:38:19.658 "traddr": "$NVMF_FIRST_TARGET_IP", 00:38:19.658 "adrfam": "ipv4", 00:38:19.658 "trsvcid": "$NVMF_PORT", 00:38:19.658 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:38:19.658 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:38:19.658 "hdgst": ${hdgst:-false}, 00:38:19.658 "ddgst": ${ddgst:-false} 00:38:19.658 }, 00:38:19.658 "method": "bdev_nvme_attach_controller" 00:38:19.658 } 00:38:19.658 EOF 00:38:19.658 )") 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@582 -- # cat 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@584 -- # jq . 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@585 -- # IFS=, 00:38:19.658 17:54:01 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:38:19.658 "params": { 00:38:19.658 "name": "Nvme1", 00:38:19.658 "trtype": "tcp", 00:38:19.658 "traddr": "10.0.0.2", 00:38:19.658 "adrfam": "ipv4", 00:38:19.658 "trsvcid": "4420", 00:38:19.658 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:38:19.658 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:38:19.658 "hdgst": false, 00:38:19.658 "ddgst": false 00:38:19.658 }, 00:38:19.658 "method": "bdev_nvme_attach_controller" 00:38:19.658 }' 00:38:19.916 [2024-12-06 17:54:01.532412] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:38:19.916 [2024-12-06 17:54:01.532487] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid410114 ] 00:38:19.916 [2024-12-06 17:54:01.603606] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:38:19.916 [2024-12-06 17:54:01.651901] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:38:20.173 Running I/O for 1 seconds... 00:38:21.106 8533.00 IOPS, 33.33 MiB/s 00:38:21.106 Latency(us) 00:38:21.106 [2024-12-06T16:54:02.945Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:38:21.106 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:38:21.106 Verification LBA range: start 0x0 length 0x4000 00:38:21.106 Nvme1n1 : 1.01 8613.66 33.65 0.00 0.00 14756.34 2063.17 16117.00 00:38:21.106 [2024-12-06T16:54:02.945Z] =================================================================================================================== 00:38:21.106 [2024-12-06T16:54:02.945Z] Total : 8613.66 33.65 0.00 0.00 14756.34 2063.17 16117.00 00:38:21.364 17:54:03 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@30 -- # bdevperfpid=410394 00:38:21.364 17:54:03 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@32 -- # sleep 3 00:38:21.364 17:54:03 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@29 -- # gen_nvmf_target_json 00:38:21.364 17:54:03 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/63 -q 128 -o 4096 -w verify -t 15 -f 00:38:21.364 17:54:03 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@560 -- # config=() 00:38:21.364 17:54:03 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@560 -- # local subsystem config 00:38:21.364 17:54:03 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:38:21.364 17:54:03 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:38:21.364 { 00:38:21.364 "params": { 00:38:21.364 "name": "Nvme$subsystem", 00:38:21.364 "trtype": "$TEST_TRANSPORT", 00:38:21.364 "traddr": "$NVMF_FIRST_TARGET_IP", 00:38:21.364 "adrfam": "ipv4", 00:38:21.364 "trsvcid": "$NVMF_PORT", 00:38:21.364 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:38:21.364 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:38:21.364 "hdgst": ${hdgst:-false}, 00:38:21.364 "ddgst": ${ddgst:-false} 00:38:21.364 }, 00:38:21.364 "method": "bdev_nvme_attach_controller" 00:38:21.364 } 00:38:21.364 EOF 00:38:21.364 )") 00:38:21.364 17:54:03 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@582 -- # cat 00:38:21.364 17:54:03 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@584 -- # jq . 00:38:21.364 17:54:03 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@585 -- # IFS=, 00:38:21.364 17:54:03 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:38:21.364 "params": { 00:38:21.364 "name": "Nvme1", 00:38:21.364 "trtype": "tcp", 00:38:21.364 "traddr": "10.0.0.2", 00:38:21.364 "adrfam": "ipv4", 00:38:21.364 "trsvcid": "4420", 00:38:21.365 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:38:21.365 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:38:21.365 "hdgst": false, 00:38:21.365 "ddgst": false 00:38:21.365 }, 00:38:21.365 "method": "bdev_nvme_attach_controller" 00:38:21.365 }' 00:38:21.365 [2024-12-06 17:54:03.120549] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:38:21.365 [2024-12-06 17:54:03.120636] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid410394 ] 00:38:21.365 [2024-12-06 17:54:03.189547] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:38:21.623 [2024-12-06 17:54:03.235545] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:38:21.623 Running I/O for 15 seconds... 00:38:23.927 8567.00 IOPS, 33.46 MiB/s [2024-12-06T16:54:06.332Z] 8572.50 IOPS, 33.49 MiB/s [2024-12-06T16:54:06.332Z] 17:54:06 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@33 -- # kill -9 410070 00:38:24.493 17:54:06 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@35 -- # sleep 3 00:38:24.493 [2024-12-06 17:54:06.092798] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:49040 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:24.493 [2024-12-06 17:54:06.092851] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.493 [2024-12-06 17:54:06.092885] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:49048 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:24.493 [2024-12-06 17:54:06.092904] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.493 [2024-12-06 17:54:06.092922] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:49056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:24.493 [2024-12-06 17:54:06.092939] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.493 [2024-12-06 17:54:06.092972] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:49064 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:24.493 [2024-12-06 17:54:06.092988] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.493 [2024-12-06 17:54:06.093008] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:49072 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:24.493 [2024-12-06 17:54:06.093044] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.493 [2024-12-06 17:54:06.093063] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:49080 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:24.493 [2024-12-06 17:54:06.093078] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.493 [2024-12-06 17:54:06.093093] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:48088 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.493 [2024-12-06 17:54:06.093124] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.493 [2024-12-06 17:54:06.093140] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:48096 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.493 [2024-12-06 17:54:06.093163] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.493 [2024-12-06 17:54:06.093194] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:48104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.493 [2024-12-06 17:54:06.093208] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.493 [2024-12-06 17:54:06.093233] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:48112 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.493 [2024-12-06 17:54:06.093251] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.493 [2024-12-06 17:54:06.093266] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:125 nsid:1 lba:48120 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.493 [2024-12-06 17:54:06.093279] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.493 [2024-12-06 17:54:06.093293] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:83 nsid:1 lba:48128 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.493 [2024-12-06 17:54:06.093305] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.493 [2024-12-06 17:54:06.093318] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:114 nsid:1 lba:48136 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.493 [2024-12-06 17:54:06.093333] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.493 [2024-12-06 17:54:06.093350] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:117 nsid:1 lba:48144 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.493 [2024-12-06 17:54:06.093365] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.493 [2024-12-06 17:54:06.093381] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:89 nsid:1 lba:48152 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.493 [2024-12-06 17:54:06.093397] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.493 [2024-12-06 17:54:06.093413] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:68 nsid:1 lba:48160 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.493 [2024-12-06 17:54:06.093429] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.493 [2024-12-06 17:54:06.093453] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:48168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.493 [2024-12-06 17:54:06.093469] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.493 [2024-12-06 17:54:06.093486] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:48176 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.493 [2024-12-06 17:54:06.093501] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.493 [2024-12-06 17:54:06.093530] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:48184 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.493 [2024-12-06 17:54:06.093543] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.493 [2024-12-06 17:54:06.093557] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:48192 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.493 [2024-12-06 17:54:06.093577] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.493 [2024-12-06 17:54:06.093595] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:48200 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.493 [2024-12-06 17:54:06.093609] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.493 [2024-12-06 17:54:06.093624] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:49088 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:24.494 [2024-12-06 17:54:06.093637] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.093683] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:64 nsid:1 lba:48208 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.093711] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.093726] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:48216 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.093741] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.093756] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:73 nsid:1 lba:48224 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.093775] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.093791] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:119 nsid:1 lba:48232 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.093805] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.093820] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:48240 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.093834] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.093849] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:115 nsid:1 lba:48248 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.093862] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.093880] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:66 nsid:1 lba:48256 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.093899] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.093922] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:48264 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.093936] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.093967] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:48272 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.093980] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.093994] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:110 nsid:1 lba:48280 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094007] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094036] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:48288 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094052] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094066] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:112 nsid:1 lba:48296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094079] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094100] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:48304 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094116] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094131] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:48312 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094143] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094157] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:48320 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094169] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094183] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:48328 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094196] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094216] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:48336 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094230] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094243] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:48344 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094255] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094269] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:82 nsid:1 lba:48352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094281] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094298] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:48360 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094315] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094329] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:48368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094342] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094356] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:48376 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094368] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094382] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:48384 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094394] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094408] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:86 nsid:1 lba:48392 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094424] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094438] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:48400 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094450] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094464] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:48408 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094476] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094489] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:48416 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094504] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094523] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:48424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094536] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094550] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:48432 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094563] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094576] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:74 nsid:1 lba:48440 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094589] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094602] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:48448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094615] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094628] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:48456 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094640] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094691] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:48464 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094711] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094726] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:48472 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094740] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094755] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:102 nsid:1 lba:48480 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094770] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094785] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:48488 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094799] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094819] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:48496 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094834] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094849] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:48504 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094863] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.494 [2024-12-06 17:54:06.094881] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:48512 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.494 [2024-12-06 17:54:06.094902] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.094919] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:48520 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.094934] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.094949] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:98 nsid:1 lba:48528 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.094976] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.094990] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:104 nsid:1 lba:48536 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095002] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095015] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:48544 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095028] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095042] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:48552 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095054] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095067] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:48560 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095080] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095096] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:48568 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095114] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095129] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:123 nsid:1 lba:48576 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095141] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095155] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:79 nsid:1 lba:48584 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095167] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095186] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:48592 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095207] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095222] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:48600 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095234] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095247] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:48608 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095259] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095273] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:48616 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095286] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095299] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:48624 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095317] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095334] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:48632 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095347] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095361] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:48640 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095373] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095387] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:48648 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095399] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095413] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:48656 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095425] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095439] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:93 nsid:1 lba:48664 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095451] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095464] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:48672 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095477] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095491] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:48680 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095503] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095517] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:48688 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095537] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095556] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:96 nsid:1 lba:48696 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095570] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095584] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:48704 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095596] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095609] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:48712 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095621] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095636] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:48720 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095670] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095696] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:48728 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095711] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095733] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:48736 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095747] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095762] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:48744 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095776] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095800] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:48752 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095814] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095829] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:48760 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095843] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095858] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:71 nsid:1 lba:48768 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095872] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095893] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:75 nsid:1 lba:48776 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095911] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095928] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:78 nsid:1 lba:48784 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095941] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.095971] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:48792 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.095987] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.096002] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:48800 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.096029] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.096043] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:48808 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.096056] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.096069] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:48816 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.096087] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.096108] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:48824 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.495 [2024-12-06 17:54:06.096124] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.495 [2024-12-06 17:54:06.096138] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:48832 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.496 [2024-12-06 17:54:06.096151] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.496 [2024-12-06 17:54:06.096164] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:48840 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.496 [2024-12-06 17:54:06.096177] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.496 [2024-12-06 17:54:06.096191] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:107 nsid:1 lba:49096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:24.496 [2024-12-06 17:54:06.096203] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.496 [2024-12-06 17:54:06.096216] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:49104 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:38:24.496 [2024-12-06 17:54:06.096229] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.496 [2024-12-06 17:54:06.096242] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:85 nsid:1 lba:48848 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.496 [2024-12-06 17:54:06.096255] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.496 [2024-12-06 17:54:06.096268] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:77 nsid:1 lba:48856 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.496 [2024-12-06 17:54:06.096280] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.496 [2024-12-06 17:54:06.096294] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:48864 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.496 [2024-12-06 17:54:06.096307] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.496 [2024-12-06 17:54:06.096330] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:48872 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.496 [2024-12-06 17:54:06.096345] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.496 [2024-12-06 17:54:06.096359] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:48880 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.496 [2024-12-06 17:54:06.096375] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.496 [2024-12-06 17:54:06.096389] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:48888 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.496 [2024-12-06 17:54:06.096402] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.496 [2024-12-06 17:54:06.096415] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:48896 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.496 [2024-12-06 17:54:06.096427] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.496 [2024-12-06 17:54:06.096441] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:48904 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.496 [2024-12-06 17:54:06.096453] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.496 [2024-12-06 17:54:06.096466] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:81 nsid:1 lba:48912 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.496 [2024-12-06 17:54:06.096479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.496 [2024-12-06 17:54:06.096492] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:48920 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.496 [2024-12-06 17:54:06.096504] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.496 [2024-12-06 17:54:06.096517] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:48928 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.496 [2024-12-06 17:54:06.096536] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.496 [2024-12-06 17:54:06.096553] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:76 nsid:1 lba:48936 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.496 [2024-12-06 17:54:06.096566] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.496 [2024-12-06 17:54:06.096579] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:116 nsid:1 lba:48944 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.496 [2024-12-06 17:54:06.096591] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.496 [2024-12-06 17:54:06.096604] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:48952 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.496 [2024-12-06 17:54:06.096616] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.496 [2024-12-06 17:54:06.096629] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:48960 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.496 [2024-12-06 17:54:06.096640] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.496 [2024-12-06 17:54:06.096680] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:48968 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.496 [2024-12-06 17:54:06.096700] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.496 [2024-12-06 17:54:06.096716] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:48976 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.496 [2024-12-06 17:54:06.096729] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.496 [2024-12-06 17:54:06.096749] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:48984 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.496 [2024-12-06 17:54:06.096763] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.496 [2024-12-06 17:54:06.096778] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:48992 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.496 [2024-12-06 17:54:06.096792] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.496 [2024-12-06 17:54:06.096806] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:49000 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.496 [2024-12-06 17:54:06.096820] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.496 [2024-12-06 17:54:06.096835] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:49008 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.496 [2024-12-06 17:54:06.096849] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.496 [2024-12-06 17:54:06.096863] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:49016 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.496 [2024-12-06 17:54:06.096877] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.496 [2024-12-06 17:54:06.096902] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:106 nsid:1 lba:49024 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:38:24.496 [2024-12-06 17:54:06.096919] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.496 [2024-12-06 17:54:06.096933] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x94bb30 is same with the state(6) to be set 00:38:24.496 [2024-12-06 17:54:06.096964] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:38:24.496 [2024-12-06 17:54:06.096975] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:38:24.496 [2024-12-06 17:54:06.096985] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:49032 len:8 PRP1 0x0 PRP2 0x0 00:38:24.496 [2024-12-06 17:54:06.096996] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:38:24.496 [2024-12-06 17:54:06.100517] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.496 [2024-12-06 17:54:06.100598] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.496 [2024-12-06 17:54:06.101344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.496 [2024-12-06 17:54:06.101390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.496 [2024-12-06 17:54:06.101408] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.496 [2024-12-06 17:54:06.101692] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.496 [2024-12-06 17:54:06.101929] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.496 [2024-12-06 17:54:06.101952] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.496 [2024-12-06 17:54:06.101980] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.496 [2024-12-06 17:54:06.101996] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.496 [2024-12-06 17:54:06.114185] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.496 [2024-12-06 17:54:06.114560] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.496 [2024-12-06 17:54:06.114589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.496 [2024-12-06 17:54:06.114606] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.496 [2024-12-06 17:54:06.114877] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.496 [2024-12-06 17:54:06.115114] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.496 [2024-12-06 17:54:06.115134] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.496 [2024-12-06 17:54:06.115147] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.496 [2024-12-06 17:54:06.115158] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.496 [2024-12-06 17:54:06.127515] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.496 [2024-12-06 17:54:06.128052] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.496 [2024-12-06 17:54:06.128081] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.496 [2024-12-06 17:54:06.128098] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.496 [2024-12-06 17:54:06.128353] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.497 [2024-12-06 17:54:06.128557] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.497 [2024-12-06 17:54:06.128576] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.497 [2024-12-06 17:54:06.128589] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.497 [2024-12-06 17:54:06.128600] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.497 [2024-12-06 17:54:06.140820] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.497 [2024-12-06 17:54:06.141218] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.497 [2024-12-06 17:54:06.141261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.497 [2024-12-06 17:54:06.141277] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.497 [2024-12-06 17:54:06.141550] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.497 [2024-12-06 17:54:06.141814] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.497 [2024-12-06 17:54:06.141837] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.497 [2024-12-06 17:54:06.141851] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.497 [2024-12-06 17:54:06.141864] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.497 [2024-12-06 17:54:06.154083] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.497 [2024-12-06 17:54:06.154520] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.497 [2024-12-06 17:54:06.154551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.497 [2024-12-06 17:54:06.154577] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.497 [2024-12-06 17:54:06.154850] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.497 [2024-12-06 17:54:06.155092] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.497 [2024-12-06 17:54:06.155112] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.497 [2024-12-06 17:54:06.155133] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.497 [2024-12-06 17:54:06.155146] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.497 [2024-12-06 17:54:06.167363] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.497 [2024-12-06 17:54:06.167766] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.497 [2024-12-06 17:54:06.167809] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.497 [2024-12-06 17:54:06.167825] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.497 [2024-12-06 17:54:06.168060] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.497 [2024-12-06 17:54:06.168301] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.497 [2024-12-06 17:54:06.168322] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.497 [2024-12-06 17:54:06.168335] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.497 [2024-12-06 17:54:06.168347] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.497 [2024-12-06 17:54:06.180639] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.497 [2024-12-06 17:54:06.181030] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.497 [2024-12-06 17:54:06.181059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.497 [2024-12-06 17:54:06.181075] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.497 [2024-12-06 17:54:06.181301] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.497 [2024-12-06 17:54:06.181542] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.497 [2024-12-06 17:54:06.181563] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.497 [2024-12-06 17:54:06.181576] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.497 [2024-12-06 17:54:06.181587] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.497 [2024-12-06 17:54:06.194087] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.497 [2024-12-06 17:54:06.194526] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.497 [2024-12-06 17:54:06.194569] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.497 [2024-12-06 17:54:06.194586] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.497 [2024-12-06 17:54:06.194845] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.497 [2024-12-06 17:54:06.195087] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.497 [2024-12-06 17:54:06.195111] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.497 [2024-12-06 17:54:06.195127] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.497 [2024-12-06 17:54:06.195138] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.497 [2024-12-06 17:54:06.207295] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.497 [2024-12-06 17:54:06.207676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.497 [2024-12-06 17:54:06.207720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.497 [2024-12-06 17:54:06.207737] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.497 [2024-12-06 17:54:06.207988] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.497 [2024-12-06 17:54:06.208207] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.497 [2024-12-06 17:54:06.208227] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.497 [2024-12-06 17:54:06.208239] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.497 [2024-12-06 17:54:06.208250] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.497 [2024-12-06 17:54:06.220598] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.497 [2024-12-06 17:54:06.221059] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.497 [2024-12-06 17:54:06.221090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.497 [2024-12-06 17:54:06.221111] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.497 [2024-12-06 17:54:06.221372] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.497 [2024-12-06 17:54:06.221576] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.497 [2024-12-06 17:54:06.221595] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.497 [2024-12-06 17:54:06.221608] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.497 [2024-12-06 17:54:06.221619] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.497 [2024-12-06 17:54:06.234044] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.497 [2024-12-06 17:54:06.234422] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.497 [2024-12-06 17:54:06.234466] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.497 [2024-12-06 17:54:06.234491] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.497 [2024-12-06 17:54:06.234767] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.497 [2024-12-06 17:54:06.235017] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.497 [2024-12-06 17:54:06.235036] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.497 [2024-12-06 17:54:06.235054] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.497 [2024-12-06 17:54:06.235066] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.497 [2024-12-06 17:54:06.247487] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.497 [2024-12-06 17:54:06.247849] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.498 [2024-12-06 17:54:06.247898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.498 [2024-12-06 17:54:06.247919] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.498 [2024-12-06 17:54:06.248150] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.498 [2024-12-06 17:54:06.248366] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.498 [2024-12-06 17:54:06.248386] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.498 [2024-12-06 17:54:06.248399] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.498 [2024-12-06 17:54:06.248411] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.498 [2024-12-06 17:54:06.260802] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.498 [2024-12-06 17:54:06.261159] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.498 [2024-12-06 17:54:06.261187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.498 [2024-12-06 17:54:06.261203] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.498 [2024-12-06 17:54:06.261429] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.498 [2024-12-06 17:54:06.261658] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.498 [2024-12-06 17:54:06.261708] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.498 [2024-12-06 17:54:06.261722] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.498 [2024-12-06 17:54:06.261734] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.498 [2024-12-06 17:54:06.274512] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.498 [2024-12-06 17:54:06.274903] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.498 [2024-12-06 17:54:06.274934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.498 [2024-12-06 17:54:06.274951] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.498 [2024-12-06 17:54:06.275216] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.498 [2024-12-06 17:54:06.275426] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.498 [2024-12-06 17:54:06.275446] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.498 [2024-12-06 17:54:06.275459] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.498 [2024-12-06 17:54:06.275470] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.498 [2024-12-06 17:54:06.287862] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.498 [2024-12-06 17:54:06.288343] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.498 [2024-12-06 17:54:06.288385] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.498 [2024-12-06 17:54:06.288402] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.498 [2024-12-06 17:54:06.288648] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.498 [2024-12-06 17:54:06.288882] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.498 [2024-12-06 17:54:06.288904] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.498 [2024-12-06 17:54:06.288918] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.498 [2024-12-06 17:54:06.288930] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.498 [2024-12-06 17:54:06.301269] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.498 [2024-12-06 17:54:06.301648] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.498 [2024-12-06 17:54:06.301703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.498 [2024-12-06 17:54:06.301721] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.498 [2024-12-06 17:54:06.301984] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.498 [2024-12-06 17:54:06.302201] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.498 [2024-12-06 17:54:06.302220] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.498 [2024-12-06 17:54:06.302232] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.498 [2024-12-06 17:54:06.302244] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.498 [2024-12-06 17:54:06.314519] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.498 [2024-12-06 17:54:06.314970] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.498 [2024-12-06 17:54:06.315013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.498 [2024-12-06 17:54:06.315030] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.498 [2024-12-06 17:54:06.315282] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.498 [2024-12-06 17:54:06.315498] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.498 [2024-12-06 17:54:06.315517] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.498 [2024-12-06 17:54:06.315539] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.498 [2024-12-06 17:54:06.315553] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.498 [2024-12-06 17:54:06.328242] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.498 [2024-12-06 17:54:06.328602] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.498 [2024-12-06 17:54:06.328632] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.498 [2024-12-06 17:54:06.328655] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.756 [2024-12-06 17:54:06.328909] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.756 [2024-12-06 17:54:06.329153] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.756 [2024-12-06 17:54:06.329174] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.756 [2024-12-06 17:54:06.329187] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.756 [2024-12-06 17:54:06.329199] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.756 [2024-12-06 17:54:06.341726] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.756 [2024-12-06 17:54:06.342140] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.756 [2024-12-06 17:54:06.342184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.756 [2024-12-06 17:54:06.342199] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.756 [2024-12-06 17:54:06.342481] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.756 [2024-12-06 17:54:06.342706] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.756 [2024-12-06 17:54:06.342728] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.756 [2024-12-06 17:54:06.342741] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.756 [2024-12-06 17:54:06.342753] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.756 [2024-12-06 17:54:06.355124] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.756 [2024-12-06 17:54:06.355485] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.756 [2024-12-06 17:54:06.355529] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.756 [2024-12-06 17:54:06.355546] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.756 [2024-12-06 17:54:06.355780] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.756 [2024-12-06 17:54:06.356034] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.756 [2024-12-06 17:54:06.356055] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.757 [2024-12-06 17:54:06.356078] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.757 [2024-12-06 17:54:06.356093] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.757 [2024-12-06 17:54:06.368875] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.757 [2024-12-06 17:54:06.369221] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.757 [2024-12-06 17:54:06.369249] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.757 [2024-12-06 17:54:06.369265] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.757 [2024-12-06 17:54:06.369493] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.757 [2024-12-06 17:54:06.369761] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.757 [2024-12-06 17:54:06.369783] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.757 [2024-12-06 17:54:06.369797] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.757 [2024-12-06 17:54:06.369810] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.757 [2024-12-06 17:54:06.382354] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.757 [2024-12-06 17:54:06.382762] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.757 [2024-12-06 17:54:06.382792] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.757 [2024-12-06 17:54:06.382809] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.757 [2024-12-06 17:54:06.383057] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.757 [2024-12-06 17:54:06.383274] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.757 [2024-12-06 17:54:06.383294] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.757 [2024-12-06 17:54:06.383306] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.757 [2024-12-06 17:54:06.383317] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.757 [2024-12-06 17:54:06.395644] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.757 [2024-12-06 17:54:06.396050] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.757 [2024-12-06 17:54:06.396079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.757 [2024-12-06 17:54:06.396095] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.757 [2024-12-06 17:54:06.396336] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.757 [2024-12-06 17:54:06.396537] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.757 [2024-12-06 17:54:06.396561] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.757 [2024-12-06 17:54:06.396577] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.757 [2024-12-06 17:54:06.396589] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.757 [2024-12-06 17:54:06.408867] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.757 [2024-12-06 17:54:06.409319] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.757 [2024-12-06 17:54:06.409347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.757 [2024-12-06 17:54:06.409364] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.757 [2024-12-06 17:54:06.409593] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.757 [2024-12-06 17:54:06.409852] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.757 [2024-12-06 17:54:06.409874] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.757 [2024-12-06 17:54:06.409894] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.757 [2024-12-06 17:54:06.409915] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.757 [2024-12-06 17:54:06.422281] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.757 [2024-12-06 17:54:06.422790] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.757 [2024-12-06 17:54:06.422820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.757 [2024-12-06 17:54:06.422836] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.757 [2024-12-06 17:54:06.423083] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.757 [2024-12-06 17:54:06.423297] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.757 [2024-12-06 17:54:06.423316] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.757 [2024-12-06 17:54:06.423328] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.757 [2024-12-06 17:54:06.423346] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.757 [2024-12-06 17:54:06.435545] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.757 [2024-12-06 17:54:06.435910] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.757 [2024-12-06 17:54:06.435940] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.757 [2024-12-06 17:54:06.435957] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.757 [2024-12-06 17:54:06.436194] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.757 [2024-12-06 17:54:06.436418] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.757 [2024-12-06 17:54:06.436438] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.757 [2024-12-06 17:54:06.436450] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.757 [2024-12-06 17:54:06.436462] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.757 [2024-12-06 17:54:06.449171] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.757 [2024-12-06 17:54:06.449623] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.757 [2024-12-06 17:54:06.449652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.757 [2024-12-06 17:54:06.449679] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.757 [2024-12-06 17:54:06.449924] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.757 [2024-12-06 17:54:06.450165] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.757 [2024-12-06 17:54:06.450185] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.757 [2024-12-06 17:54:06.450198] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.757 [2024-12-06 17:54:06.450209] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.757 [2024-12-06 17:54:06.462715] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.757 [2024-12-06 17:54:06.463157] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.757 [2024-12-06 17:54:06.463199] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.757 [2024-12-06 17:54:06.463215] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.757 [2024-12-06 17:54:06.463464] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.758 [2024-12-06 17:54:06.463705] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.758 [2024-12-06 17:54:06.463741] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.758 [2024-12-06 17:54:06.463754] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.758 [2024-12-06 17:54:06.463767] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.758 7465.00 IOPS, 29.16 MiB/s [2024-12-06T16:54:06.597Z] [2024-12-06 17:54:06.476226] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.758 [2024-12-06 17:54:06.476606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.758 [2024-12-06 17:54:06.476634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.758 [2024-12-06 17:54:06.476677] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.758 [2024-12-06 17:54:06.476918] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.758 [2024-12-06 17:54:06.477169] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.758 [2024-12-06 17:54:06.477190] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.758 [2024-12-06 17:54:06.477203] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.758 [2024-12-06 17:54:06.477214] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.758 [2024-12-06 17:54:06.489636] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.758 [2024-12-06 17:54:06.490050] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.758 [2024-12-06 17:54:06.490094] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.758 [2024-12-06 17:54:06.490113] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.758 [2024-12-06 17:54:06.490360] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.758 [2024-12-06 17:54:06.490566] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.758 [2024-12-06 17:54:06.490587] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.758 [2024-12-06 17:54:06.490599] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.758 [2024-12-06 17:54:06.490612] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.758 [2024-12-06 17:54:06.503090] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.758 [2024-12-06 17:54:06.503474] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.758 [2024-12-06 17:54:06.503503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.758 [2024-12-06 17:54:06.503525] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.758 [2024-12-06 17:54:06.503774] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.758 [2024-12-06 17:54:06.504009] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.758 [2024-12-06 17:54:06.504030] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.758 [2024-12-06 17:54:06.504042] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.758 [2024-12-06 17:54:06.504054] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.758 [2024-12-06 17:54:06.516429] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.758 [2024-12-06 17:54:06.516809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.758 [2024-12-06 17:54:06.516838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.758 [2024-12-06 17:54:06.516855] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.758 [2024-12-06 17:54:06.517093] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.758 [2024-12-06 17:54:06.517315] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.758 [2024-12-06 17:54:06.517336] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.758 [2024-12-06 17:54:06.517355] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.758 [2024-12-06 17:54:06.517368] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.758 [2024-12-06 17:54:06.529812] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.758 [2024-12-06 17:54:06.530229] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.758 [2024-12-06 17:54:06.530273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.758 [2024-12-06 17:54:06.530290] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.758 [2024-12-06 17:54:06.530538] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.758 [2024-12-06 17:54:06.530797] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.758 [2024-12-06 17:54:06.530819] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.758 [2024-12-06 17:54:06.530833] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.758 [2024-12-06 17:54:06.530845] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.758 [2024-12-06 17:54:06.543210] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.758 [2024-12-06 17:54:06.543645] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.758 [2024-12-06 17:54:06.543685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.758 [2024-12-06 17:54:06.543727] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.758 [2024-12-06 17:54:06.543979] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.758 [2024-12-06 17:54:06.544210] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.758 [2024-12-06 17:54:06.544230] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.758 [2024-12-06 17:54:06.544243] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.758 [2024-12-06 17:54:06.544255] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.758 [2024-12-06 17:54:06.556655] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.758 [2024-12-06 17:54:06.557105] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.758 [2024-12-06 17:54:06.557151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.758 [2024-12-06 17:54:06.557171] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.758 [2024-12-06 17:54:06.557423] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.758 [2024-12-06 17:54:06.557629] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.758 [2024-12-06 17:54:06.557675] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.758 [2024-12-06 17:54:06.557693] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.758 [2024-12-06 17:54:06.557706] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.758 [2024-12-06 17:54:06.570140] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.758 [2024-12-06 17:54:06.570489] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.759 [2024-12-06 17:54:06.570519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.759 [2024-12-06 17:54:06.570538] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.759 [2024-12-06 17:54:06.570795] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.759 [2024-12-06 17:54:06.571050] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.759 [2024-12-06 17:54:06.571070] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.759 [2024-12-06 17:54:06.571082] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.759 [2024-12-06 17:54:06.571094] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:24.759 [2024-12-06 17:54:06.583544] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:24.759 [2024-12-06 17:54:06.583955] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:24.759 [2024-12-06 17:54:06.583999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:24.759 [2024-12-06 17:54:06.584015] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:24.759 [2024-12-06 17:54:06.584279] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:24.759 [2024-12-06 17:54:06.584484] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:24.759 [2024-12-06 17:54:06.584504] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:24.759 [2024-12-06 17:54:06.584523] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:24.759 [2024-12-06 17:54:06.584543] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.018 [2024-12-06 17:54:06.597107] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.018 [2024-12-06 17:54:06.597433] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.018 [2024-12-06 17:54:06.597461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.018 [2024-12-06 17:54:06.597477] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.018 [2024-12-06 17:54:06.597730] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.018 [2024-12-06 17:54:06.597959] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.018 [2024-12-06 17:54:06.597982] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.018 [2024-12-06 17:54:06.597996] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.018 [2024-12-06 17:54:06.598008] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.018 [2024-12-06 17:54:06.610583] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.018 [2024-12-06 17:54:06.610983] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.018 [2024-12-06 17:54:06.611013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.018 [2024-12-06 17:54:06.611029] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.018 [2024-12-06 17:54:06.611267] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.018 [2024-12-06 17:54:06.611507] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.018 [2024-12-06 17:54:06.611529] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.018 [2024-12-06 17:54:06.611551] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.018 [2024-12-06 17:54:06.611566] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.018 [2024-12-06 17:54:06.624126] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.018 [2024-12-06 17:54:06.624510] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.018 [2024-12-06 17:54:06.624546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.018 [2024-12-06 17:54:06.624571] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.018 [2024-12-06 17:54:06.624831] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.018 [2024-12-06 17:54:06.625077] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.018 [2024-12-06 17:54:06.625098] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.018 [2024-12-06 17:54:06.625118] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.018 [2024-12-06 17:54:06.625132] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.018 [2024-12-06 17:54:06.637606] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.018 [2024-12-06 17:54:06.638017] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.018 [2024-12-06 17:54:06.638046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.018 [2024-12-06 17:54:06.638063] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.018 [2024-12-06 17:54:06.638301] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.018 [2024-12-06 17:54:06.638553] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.018 [2024-12-06 17:54:06.638574] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.018 [2024-12-06 17:54:06.638587] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.018 [2024-12-06 17:54:06.638598] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.018 [2024-12-06 17:54:06.651061] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.018 [2024-12-06 17:54:06.651442] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.018 [2024-12-06 17:54:06.651485] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.018 [2024-12-06 17:54:06.651502] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.018 [2024-12-06 17:54:06.651797] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.018 [2024-12-06 17:54:06.652018] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.018 [2024-12-06 17:54:06.652046] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.018 [2024-12-06 17:54:06.652079] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.018 [2024-12-06 17:54:06.652092] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.018 [2024-12-06 17:54:06.664417] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.018 [2024-12-06 17:54:06.664806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.018 [2024-12-06 17:54:06.664835] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.018 [2024-12-06 17:54:06.664851] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.018 [2024-12-06 17:54:06.665104] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.018 [2024-12-06 17:54:06.665314] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.018 [2024-12-06 17:54:06.665334] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.018 [2024-12-06 17:54:06.665347] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.019 [2024-12-06 17:54:06.665358] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.019 [2024-12-06 17:54:06.678026] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.019 [2024-12-06 17:54:06.678396] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.019 [2024-12-06 17:54:06.678425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.019 [2024-12-06 17:54:06.678448] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.019 [2024-12-06 17:54:06.678703] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.019 [2024-12-06 17:54:06.678933] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.019 [2024-12-06 17:54:06.678954] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.019 [2024-12-06 17:54:06.678967] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.019 [2024-12-06 17:54:06.678979] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.019 [2024-12-06 17:54:06.691377] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.019 [2024-12-06 17:54:06.691739] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.019 [2024-12-06 17:54:06.691768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.019 [2024-12-06 17:54:06.691784] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.019 [2024-12-06 17:54:06.692015] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.019 [2024-12-06 17:54:06.692239] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.019 [2024-12-06 17:54:06.692259] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.019 [2024-12-06 17:54:06.692273] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.019 [2024-12-06 17:54:06.692285] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.019 [2024-12-06 17:54:06.704735] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.019 [2024-12-06 17:54:06.705138] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.019 [2024-12-06 17:54:06.705166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.019 [2024-12-06 17:54:06.705183] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.019 [2024-12-06 17:54:06.705426] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.019 [2024-12-06 17:54:06.705633] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.019 [2024-12-06 17:54:06.705678] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.019 [2024-12-06 17:54:06.705694] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.019 [2024-12-06 17:54:06.705706] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.019 [2024-12-06 17:54:06.718164] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.019 [2024-12-06 17:54:06.718523] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.019 [2024-12-06 17:54:06.718557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.019 [2024-12-06 17:54:06.718574] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.019 [2024-12-06 17:54:06.718805] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.019 [2024-12-06 17:54:06.719052] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.019 [2024-12-06 17:54:06.719073] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.019 [2024-12-06 17:54:06.719086] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.019 [2024-12-06 17:54:06.719101] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.019 [2024-12-06 17:54:06.731756] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.019 [2024-12-06 17:54:06.732226] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.019 [2024-12-06 17:54:06.732255] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.019 [2024-12-06 17:54:06.732272] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.019 [2024-12-06 17:54:06.732521] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.019 [2024-12-06 17:54:06.732772] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.019 [2024-12-06 17:54:06.732795] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.019 [2024-12-06 17:54:06.732808] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.019 [2024-12-06 17:54:06.732821] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.019 [2024-12-06 17:54:06.745244] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.019 [2024-12-06 17:54:06.745591] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.019 [2024-12-06 17:54:06.745619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.019 [2024-12-06 17:54:06.745635] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.019 [2024-12-06 17:54:06.745916] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.019 [2024-12-06 17:54:06.746130] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.019 [2024-12-06 17:54:06.746150] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.019 [2024-12-06 17:54:06.746164] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.019 [2024-12-06 17:54:06.746176] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.019 [2024-12-06 17:54:06.758701] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.019 [2024-12-06 17:54:06.759041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.019 [2024-12-06 17:54:06.759069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.019 [2024-12-06 17:54:06.759086] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.019 [2024-12-06 17:54:06.759318] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.019 [2024-12-06 17:54:06.759541] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.019 [2024-12-06 17:54:06.759561] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.019 [2024-12-06 17:54:06.759574] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.019 [2024-12-06 17:54:06.759590] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.019 [2024-12-06 17:54:06.772067] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.019 [2024-12-06 17:54:06.772451] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.019 [2024-12-06 17:54:06.772479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.019 [2024-12-06 17:54:06.772496] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.019 [2024-12-06 17:54:06.772753] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.019 [2024-12-06 17:54:06.773002] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.019 [2024-12-06 17:54:06.773024] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.019 [2024-12-06 17:54:06.773038] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.019 [2024-12-06 17:54:06.773051] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.019 [2024-12-06 17:54:06.785391] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.019 [2024-12-06 17:54:06.785769] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.019 [2024-12-06 17:54:06.785798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.019 [2024-12-06 17:54:06.785815] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.019 [2024-12-06 17:54:06.786044] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.019 [2024-12-06 17:54:06.786267] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.020 [2024-12-06 17:54:06.786286] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.020 [2024-12-06 17:54:06.786299] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.020 [2024-12-06 17:54:06.786312] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.020 [2024-12-06 17:54:06.798914] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.020 [2024-12-06 17:54:06.799341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.020 [2024-12-06 17:54:06.799369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.020 [2024-12-06 17:54:06.799385] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.020 [2024-12-06 17:54:06.799615] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.020 [2024-12-06 17:54:06.799866] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.020 [2024-12-06 17:54:06.799887] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.020 [2024-12-06 17:54:06.799901] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.020 [2024-12-06 17:54:06.799914] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.020 [2024-12-06 17:54:06.812360] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.020 [2024-12-06 17:54:06.812739] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.020 [2024-12-06 17:54:06.812767] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.020 [2024-12-06 17:54:06.812784] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.020 [2024-12-06 17:54:06.813014] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.020 [2024-12-06 17:54:06.813237] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.020 [2024-12-06 17:54:06.813257] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.020 [2024-12-06 17:54:06.813270] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.020 [2024-12-06 17:54:06.813282] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.020 [2024-12-06 17:54:06.825989] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.020 [2024-12-06 17:54:06.826406] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.020 [2024-12-06 17:54:06.826435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.020 [2024-12-06 17:54:06.826451] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.020 [2024-12-06 17:54:06.826710] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.020 [2024-12-06 17:54:06.826931] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.020 [2024-12-06 17:54:06.826952] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.020 [2024-12-06 17:54:06.826966] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.020 [2024-12-06 17:54:06.826995] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.020 [2024-12-06 17:54:06.839482] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.020 [2024-12-06 17:54:06.839938] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.020 [2024-12-06 17:54:06.839968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.020 [2024-12-06 17:54:06.839985] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.020 [2024-12-06 17:54:06.840223] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.020 [2024-12-06 17:54:06.840445] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.020 [2024-12-06 17:54:06.840466] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.020 [2024-12-06 17:54:06.840478] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.020 [2024-12-06 17:54:06.840490] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.020 [2024-12-06 17:54:06.853208] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.020 [2024-12-06 17:54:06.853761] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.020 [2024-12-06 17:54:06.853791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.020 [2024-12-06 17:54:06.853808] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.020 [2024-12-06 17:54:06.854050] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.020 [2024-12-06 17:54:06.854295] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.020 [2024-12-06 17:54:06.854317] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.020 [2024-12-06 17:54:06.854330] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.020 [2024-12-06 17:54:06.854353] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.278 [2024-12-06 17:54:06.866672] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.278 [2024-12-06 17:54:06.867056] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.278 [2024-12-06 17:54:06.867085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.278 [2024-12-06 17:54:06.867101] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.278 [2024-12-06 17:54:06.867316] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.278 [2024-12-06 17:54:06.867536] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.278 [2024-12-06 17:54:06.867557] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.278 [2024-12-06 17:54:06.867572] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.278 [2024-12-06 17:54:06.867585] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.278 [2024-12-06 17:54:06.880252] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.278 [2024-12-06 17:54:06.880576] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.279 [2024-12-06 17:54:06.880619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.279 [2024-12-06 17:54:06.880636] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.279 [2024-12-06 17:54:06.880894] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.279 [2024-12-06 17:54:06.881124] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.279 [2024-12-06 17:54:06.881145] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.279 [2024-12-06 17:54:06.881158] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.279 [2024-12-06 17:54:06.881170] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.279 [2024-12-06 17:54:06.893615] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.279 [2024-12-06 17:54:06.894032] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.279 [2024-12-06 17:54:06.894062] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.279 [2024-12-06 17:54:06.894079] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.279 [2024-12-06 17:54:06.894328] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.279 [2024-12-06 17:54:06.894534] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.279 [2024-12-06 17:54:06.894561] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.279 [2024-12-06 17:54:06.894574] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.279 [2024-12-06 17:54:06.894586] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.279 [2024-12-06 17:54:06.907053] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.279 [2024-12-06 17:54:06.907472] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.279 [2024-12-06 17:54:06.907499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.279 [2024-12-06 17:54:06.907537] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.279 [2024-12-06 17:54:06.907813] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.279 [2024-12-06 17:54:06.908048] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.279 [2024-12-06 17:54:06.908068] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.279 [2024-12-06 17:54:06.908081] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.279 [2024-12-06 17:54:06.908092] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.279 [2024-12-06 17:54:06.920448] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.279 [2024-12-06 17:54:06.920845] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.279 [2024-12-06 17:54:06.920875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.279 [2024-12-06 17:54:06.920891] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.279 [2024-12-06 17:54:06.921143] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.279 [2024-12-06 17:54:06.921350] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.279 [2024-12-06 17:54:06.921370] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.279 [2024-12-06 17:54:06.921383] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.279 [2024-12-06 17:54:06.921394] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.279 [2024-12-06 17:54:06.933799] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.279 [2024-12-06 17:54:06.934171] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.279 [2024-12-06 17:54:06.934198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.279 [2024-12-06 17:54:06.934214] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.279 [2024-12-06 17:54:06.934444] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.279 [2024-12-06 17:54:06.934697] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.279 [2024-12-06 17:54:06.934718] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.279 [2024-12-06 17:54:06.934731] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.279 [2024-12-06 17:54:06.934763] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.279 [2024-12-06 17:54:06.947325] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.279 [2024-12-06 17:54:06.947734] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.279 [2024-12-06 17:54:06.947764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.279 [2024-12-06 17:54:06.947781] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.279 [2024-12-06 17:54:06.948021] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.279 [2024-12-06 17:54:06.948245] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.279 [2024-12-06 17:54:06.948265] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.279 [2024-12-06 17:54:06.948278] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.279 [2024-12-06 17:54:06.948290] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.279 [2024-12-06 17:54:06.960784] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.279 [2024-12-06 17:54:06.961192] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.279 [2024-12-06 17:54:06.961221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.279 [2024-12-06 17:54:06.961238] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.279 [2024-12-06 17:54:06.961488] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.279 [2024-12-06 17:54:06.961722] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.279 [2024-12-06 17:54:06.961743] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.279 [2024-12-06 17:54:06.961756] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.279 [2024-12-06 17:54:06.961768] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.279 [2024-12-06 17:54:06.974353] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.279 [2024-12-06 17:54:06.974737] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.279 [2024-12-06 17:54:06.974768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.279 [2024-12-06 17:54:06.974784] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.279 [2024-12-06 17:54:06.975015] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.279 [2024-12-06 17:54:06.975233] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.279 [2024-12-06 17:54:06.975253] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.279 [2024-12-06 17:54:06.975265] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.279 [2024-12-06 17:54:06.975276] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.279 [2024-12-06 17:54:06.987751] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.279 [2024-12-06 17:54:06.988142] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.279 [2024-12-06 17:54:06.988170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.279 [2024-12-06 17:54:06.988186] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.279 [2024-12-06 17:54:06.988410] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.279 [2024-12-06 17:54:06.988626] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.279 [2024-12-06 17:54:06.988660] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.279 [2024-12-06 17:54:06.988686] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.279 [2024-12-06 17:54:06.988699] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.279 [2024-12-06 17:54:07.001069] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.279 [2024-12-06 17:54:07.001463] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.279 [2024-12-06 17:54:07.001491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.279 [2024-12-06 17:54:07.001523] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.279 [2024-12-06 17:54:07.001765] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.279 [2024-12-06 17:54:07.001994] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.279 [2024-12-06 17:54:07.002014] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.279 [2024-12-06 17:54:07.002043] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.280 [2024-12-06 17:54:07.002055] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.280 [2024-12-06 17:54:07.014364] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.280 [2024-12-06 17:54:07.014802] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.280 [2024-12-06 17:54:07.014846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.280 [2024-12-06 17:54:07.014863] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.280 [2024-12-06 17:54:07.015093] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.280 [2024-12-06 17:54:07.015311] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.280 [2024-12-06 17:54:07.015330] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.280 [2024-12-06 17:54:07.015343] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.280 [2024-12-06 17:54:07.015355] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.280 [2024-12-06 17:54:07.027744] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.280 [2024-12-06 17:54:07.028137] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.280 [2024-12-06 17:54:07.028167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.280 [2024-12-06 17:54:07.028183] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.280 [2024-12-06 17:54:07.028426] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.280 [2024-12-06 17:54:07.028675] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.280 [2024-12-06 17:54:07.028697] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.280 [2024-12-06 17:54:07.028725] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.280 [2024-12-06 17:54:07.028739] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.280 [2024-12-06 17:54:07.041126] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.280 [2024-12-06 17:54:07.041485] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.280 [2024-12-06 17:54:07.041516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.280 [2024-12-06 17:54:07.041538] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.280 [2024-12-06 17:54:07.041771] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.280 [2024-12-06 17:54:07.042026] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.280 [2024-12-06 17:54:07.042046] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.280 [2024-12-06 17:54:07.042059] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.280 [2024-12-06 17:54:07.042070] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.280 [2024-12-06 17:54:07.054594] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.280 [2024-12-06 17:54:07.054972] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.280 [2024-12-06 17:54:07.055001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.280 [2024-12-06 17:54:07.055020] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.280 [2024-12-06 17:54:07.055274] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.280 [2024-12-06 17:54:07.055484] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.280 [2024-12-06 17:54:07.055504] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.280 [2024-12-06 17:54:07.055517] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.280 [2024-12-06 17:54:07.055528] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.280 [2024-12-06 17:54:07.067940] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.280 [2024-12-06 17:54:07.068368] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.280 [2024-12-06 17:54:07.068398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.280 [2024-12-06 17:54:07.068414] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.280 [2024-12-06 17:54:07.068679] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.280 [2024-12-06 17:54:07.068915] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.280 [2024-12-06 17:54:07.068942] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.280 [2024-12-06 17:54:07.068971] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.280 [2024-12-06 17:54:07.068984] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.280 [2024-12-06 17:54:07.081311] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.280 [2024-12-06 17:54:07.081675] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.280 [2024-12-06 17:54:07.081706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.280 [2024-12-06 17:54:07.081723] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.280 [2024-12-06 17:54:07.081972] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.280 [2024-12-06 17:54:07.082180] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.280 [2024-12-06 17:54:07.082200] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.280 [2024-12-06 17:54:07.082213] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.280 [2024-12-06 17:54:07.082224] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.280 [2024-12-06 17:54:07.094705] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.280 [2024-12-06 17:54:07.095129] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.280 [2024-12-06 17:54:07.095159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.280 [2024-12-06 17:54:07.095175] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.280 [2024-12-06 17:54:07.095428] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.280 [2024-12-06 17:54:07.095681] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.280 [2024-12-06 17:54:07.095704] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.280 [2024-12-06 17:54:07.095733] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.280 [2024-12-06 17:54:07.095746] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.280 [2024-12-06 17:54:07.108057] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.280 [2024-12-06 17:54:07.108444] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.280 [2024-12-06 17:54:07.108490] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.280 [2024-12-06 17:54:07.108507] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.280 [2024-12-06 17:54:07.108777] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.280 [2024-12-06 17:54:07.109032] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.280 [2024-12-06 17:54:07.109053] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.280 [2024-12-06 17:54:07.109066] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.280 [2024-12-06 17:54:07.109107] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.541 [2024-12-06 17:54:07.121628] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.541 [2024-12-06 17:54:07.122020] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.541 [2024-12-06 17:54:07.122049] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.541 [2024-12-06 17:54:07.122066] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.541 [2024-12-06 17:54:07.122302] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.541 [2024-12-06 17:54:07.122523] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.541 [2024-12-06 17:54:07.122562] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.541 [2024-12-06 17:54:07.122577] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.541 [2024-12-06 17:54:07.122590] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.541 [2024-12-06 17:54:07.135202] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.541 [2024-12-06 17:54:07.135588] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.541 [2024-12-06 17:54:07.135631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.541 [2024-12-06 17:54:07.135648] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.541 [2024-12-06 17:54:07.135908] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.541 [2024-12-06 17:54:07.136150] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.541 [2024-12-06 17:54:07.136170] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.541 [2024-12-06 17:54:07.136183] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.541 [2024-12-06 17:54:07.136194] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.541 [2024-12-06 17:54:07.148593] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.541 [2024-12-06 17:54:07.149003] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.541 [2024-12-06 17:54:07.149032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.541 [2024-12-06 17:54:07.149049] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.541 [2024-12-06 17:54:07.149294] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.541 [2024-12-06 17:54:07.149499] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.541 [2024-12-06 17:54:07.149524] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.541 [2024-12-06 17:54:07.149537] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.541 [2024-12-06 17:54:07.149549] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.541 [2024-12-06 17:54:07.161923] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.541 [2024-12-06 17:54:07.162386] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.541 [2024-12-06 17:54:07.162435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.541 [2024-12-06 17:54:07.162451] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.541 [2024-12-06 17:54:07.162743] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.541 [2024-12-06 17:54:07.162981] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.541 [2024-12-06 17:54:07.163002] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.541 [2024-12-06 17:54:07.163015] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.541 [2024-12-06 17:54:07.163041] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.541 [2024-12-06 17:54:07.175366] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.541 [2024-12-06 17:54:07.175752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.541 [2024-12-06 17:54:07.175783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.541 [2024-12-06 17:54:07.175800] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.541 [2024-12-06 17:54:07.176049] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.541 [2024-12-06 17:54:07.176255] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.541 [2024-12-06 17:54:07.176276] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.541 [2024-12-06 17:54:07.176288] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.541 [2024-12-06 17:54:07.176300] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.541 [2024-12-06 17:54:07.188809] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.541 [2024-12-06 17:54:07.189212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.541 [2024-12-06 17:54:07.189255] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.541 [2024-12-06 17:54:07.189271] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.541 [2024-12-06 17:54:07.189532] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.541 [2024-12-06 17:54:07.189783] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.541 [2024-12-06 17:54:07.189805] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.541 [2024-12-06 17:54:07.189819] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.541 [2024-12-06 17:54:07.189831] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.541 [2024-12-06 17:54:07.202195] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.541 [2024-12-06 17:54:07.202642] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.541 [2024-12-06 17:54:07.202681] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.541 [2024-12-06 17:54:07.202703] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.541 [2024-12-06 17:54:07.202947] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.541 [2024-12-06 17:54:07.203174] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.541 [2024-12-06 17:54:07.203194] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.541 [2024-12-06 17:54:07.203207] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.541 [2024-12-06 17:54:07.203219] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.541 [2024-12-06 17:54:07.215660] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.541 [2024-12-06 17:54:07.216018] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.542 [2024-12-06 17:54:07.216047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.542 [2024-12-06 17:54:07.216063] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.542 [2024-12-06 17:54:07.216280] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.542 [2024-12-06 17:54:07.216502] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.542 [2024-12-06 17:54:07.216522] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.542 [2024-12-06 17:54:07.216541] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.542 [2024-12-06 17:54:07.216558] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.542 [2024-12-06 17:54:07.229080] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.542 [2024-12-06 17:54:07.229519] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.542 [2024-12-06 17:54:07.229549] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.542 [2024-12-06 17:54:07.229566] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.542 [2024-12-06 17:54:07.229816] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.542 [2024-12-06 17:54:07.230063] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.542 [2024-12-06 17:54:07.230083] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.542 [2024-12-06 17:54:07.230097] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.542 [2024-12-06 17:54:07.230115] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.542 [2024-12-06 17:54:07.242467] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.542 [2024-12-06 17:54:07.242847] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.542 [2024-12-06 17:54:07.242875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.542 [2024-12-06 17:54:07.242892] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.542 [2024-12-06 17:54:07.243134] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.542 [2024-12-06 17:54:07.243336] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.542 [2024-12-06 17:54:07.243360] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.542 [2024-12-06 17:54:07.243373] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.542 [2024-12-06 17:54:07.243385] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.542 [2024-12-06 17:54:07.255808] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.542 [2024-12-06 17:54:07.256191] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.542 [2024-12-06 17:54:07.256218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.542 [2024-12-06 17:54:07.256234] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.542 [2024-12-06 17:54:07.256457] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.542 [2024-12-06 17:54:07.256702] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.542 [2024-12-06 17:54:07.256732] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.542 [2024-12-06 17:54:07.256746] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.542 [2024-12-06 17:54:07.256758] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.542 [2024-12-06 17:54:07.269164] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.542 [2024-12-06 17:54:07.269601] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.542 [2024-12-06 17:54:07.269630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.542 [2024-12-06 17:54:07.269646] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.542 [2024-12-06 17:54:07.269896] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.542 [2024-12-06 17:54:07.270125] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.542 [2024-12-06 17:54:07.270144] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.542 [2024-12-06 17:54:07.270157] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.542 [2024-12-06 17:54:07.270168] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.542 [2024-12-06 17:54:07.282441] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.542 [2024-12-06 17:54:07.282854] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.542 [2024-12-06 17:54:07.282883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.542 [2024-12-06 17:54:07.282900] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.542 [2024-12-06 17:54:07.283131] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.542 [2024-12-06 17:54:07.283347] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.542 [2024-12-06 17:54:07.283366] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.542 [2024-12-06 17:54:07.283378] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.542 [2024-12-06 17:54:07.283389] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.542 [2024-12-06 17:54:07.295631] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.542 [2024-12-06 17:54:07.296020] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.542 [2024-12-06 17:54:07.296049] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.542 [2024-12-06 17:54:07.296065] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.542 [2024-12-06 17:54:07.296281] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.542 [2024-12-06 17:54:07.296520] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.542 [2024-12-06 17:54:07.296539] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.542 [2024-12-06 17:54:07.296552] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.542 [2024-12-06 17:54:07.296563] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.542 [2024-12-06 17:54:07.309042] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.542 [2024-12-06 17:54:07.309479] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.542 [2024-12-06 17:54:07.309508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.542 [2024-12-06 17:54:07.309524] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.542 [2024-12-06 17:54:07.309764] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.542 [2024-12-06 17:54:07.309991] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.542 [2024-12-06 17:54:07.310011] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.542 [2024-12-06 17:54:07.310024] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.542 [2024-12-06 17:54:07.310035] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.542 [2024-12-06 17:54:07.322324] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.542 [2024-12-06 17:54:07.322784] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.542 [2024-12-06 17:54:07.322813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.542 [2024-12-06 17:54:07.322830] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.542 [2024-12-06 17:54:07.323070] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.542 [2024-12-06 17:54:07.323287] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.542 [2024-12-06 17:54:07.323306] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.542 [2024-12-06 17:54:07.323319] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.542 [2024-12-06 17:54:07.323330] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.542 [2024-12-06 17:54:07.335771] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.542 [2024-12-06 17:54:07.336170] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.542 [2024-12-06 17:54:07.336203] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.542 [2024-12-06 17:54:07.336220] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.542 [2024-12-06 17:54:07.336463] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.542 [2024-12-06 17:54:07.336690] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.542 [2024-12-06 17:54:07.336719] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.542 [2024-12-06 17:54:07.336732] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.542 [2024-12-06 17:54:07.336744] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.542 [2024-12-06 17:54:07.349124] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.543 [2024-12-06 17:54:07.349439] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.543 [2024-12-06 17:54:07.349482] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.543 [2024-12-06 17:54:07.349497] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.543 [2024-12-06 17:54:07.349749] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.543 [2024-12-06 17:54:07.349971] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.543 [2024-12-06 17:54:07.349991] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.543 [2024-12-06 17:54:07.350020] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.543 [2024-12-06 17:54:07.350031] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.543 [2024-12-06 17:54:07.362364] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.543 [2024-12-06 17:54:07.362786] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.543 [2024-12-06 17:54:07.362815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.543 [2024-12-06 17:54:07.362831] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.543 [2024-12-06 17:54:07.363059] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.543 [2024-12-06 17:54:07.363274] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.543 [2024-12-06 17:54:07.363294] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.543 [2024-12-06 17:54:07.363306] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.543 [2024-12-06 17:54:07.363317] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.863 [2024-12-06 17:54:07.376094] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.863 [2024-12-06 17:54:07.376446] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.863 [2024-12-06 17:54:07.376474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.863 [2024-12-06 17:54:07.376490] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.863 [2024-12-06 17:54:07.376720] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.864 [2024-12-06 17:54:07.376941] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.864 [2024-12-06 17:54:07.376963] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.864 [2024-12-06 17:54:07.376977] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.864 [2024-12-06 17:54:07.376990] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.864 [2024-12-06 17:54:07.389738] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.864 [2024-12-06 17:54:07.390122] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.864 [2024-12-06 17:54:07.390150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.864 [2024-12-06 17:54:07.390166] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.864 [2024-12-06 17:54:07.390381] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.864 [2024-12-06 17:54:07.390609] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.864 [2024-12-06 17:54:07.390630] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.864 [2024-12-06 17:54:07.390657] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.864 [2024-12-06 17:54:07.390681] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.864 [2024-12-06 17:54:07.403139] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.864 [2024-12-06 17:54:07.403539] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.864 [2024-12-06 17:54:07.403582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.864 [2024-12-06 17:54:07.403599] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.864 [2024-12-06 17:54:07.403824] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.864 [2024-12-06 17:54:07.404072] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.864 [2024-12-06 17:54:07.404092] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.864 [2024-12-06 17:54:07.404104] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.864 [2024-12-06 17:54:07.404115] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.864 [2024-12-06 17:54:07.416604] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.864 [2024-12-06 17:54:07.416980] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.864 [2024-12-06 17:54:07.417009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.864 [2024-12-06 17:54:07.417025] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.864 [2024-12-06 17:54:07.417254] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.864 [2024-12-06 17:54:07.417472] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.864 [2024-12-06 17:54:07.417492] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.864 [2024-12-06 17:54:07.417509] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.864 [2024-12-06 17:54:07.417521] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.864 [2024-12-06 17:54:07.430395] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.864 [2024-12-06 17:54:07.430804] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.864 [2024-12-06 17:54:07.430832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.864 [2024-12-06 17:54:07.430849] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.864 [2024-12-06 17:54:07.431079] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.864 [2024-12-06 17:54:07.431303] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.864 [2024-12-06 17:54:07.431323] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.864 [2024-12-06 17:54:07.431336] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.864 [2024-12-06 17:54:07.431347] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.864 [2024-12-06 17:54:07.444004] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.864 [2024-12-06 17:54:07.444367] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.864 [2024-12-06 17:54:07.444395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.864 [2024-12-06 17:54:07.444411] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.864 [2024-12-06 17:54:07.444626] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.864 [2024-12-06 17:54:07.444854] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.864 [2024-12-06 17:54:07.444876] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.864 [2024-12-06 17:54:07.444890] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.864 [2024-12-06 17:54:07.444903] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.864 [2024-12-06 17:54:07.457492] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.864 [2024-12-06 17:54:07.457931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.864 [2024-12-06 17:54:07.457968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.864 [2024-12-06 17:54:07.457985] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.864 [2024-12-06 17:54:07.458234] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.864 [2024-12-06 17:54:07.458434] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.864 [2024-12-06 17:54:07.458454] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.864 [2024-12-06 17:54:07.458466] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.864 [2024-12-06 17:54:07.458478] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.864 5598.75 IOPS, 21.87 MiB/s [2024-12-06T16:54:07.703Z] [2024-12-06 17:54:07.472447] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.864 [2024-12-06 17:54:07.472812] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.864 [2024-12-06 17:54:07.472841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.864 [2024-12-06 17:54:07.472857] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.864 [2024-12-06 17:54:07.473088] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.864 [2024-12-06 17:54:07.473305] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.864 [2024-12-06 17:54:07.473324] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.864 [2024-12-06 17:54:07.473337] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.865 [2024-12-06 17:54:07.473348] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.865 [2024-12-06 17:54:07.485834] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.865 [2024-12-06 17:54:07.486182] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.865 [2024-12-06 17:54:07.486227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.865 [2024-12-06 17:54:07.486243] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.865 [2024-12-06 17:54:07.486466] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.865 [2024-12-06 17:54:07.486711] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.865 [2024-12-06 17:54:07.486733] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.865 [2024-12-06 17:54:07.486746] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.865 [2024-12-06 17:54:07.486758] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.865 [2024-12-06 17:54:07.499248] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.865 [2024-12-06 17:54:07.499631] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.865 [2024-12-06 17:54:07.499677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.865 [2024-12-06 17:54:07.499698] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.865 [2024-12-06 17:54:07.499934] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.865 [2024-12-06 17:54:07.500190] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.865 [2024-12-06 17:54:07.500219] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.865 [2024-12-06 17:54:07.500232] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.865 [2024-12-06 17:54:07.500243] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.865 [2024-12-06 17:54:07.512644] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.865 [2024-12-06 17:54:07.513054] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.865 [2024-12-06 17:54:07.513088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.865 [2024-12-06 17:54:07.513112] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.865 [2024-12-06 17:54:07.513366] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.865 [2024-12-06 17:54:07.513579] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.865 [2024-12-06 17:54:07.513600] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.865 [2024-12-06 17:54:07.513613] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.865 [2024-12-06 17:54:07.513624] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.865 [2024-12-06 17:54:07.526135] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.865 [2024-12-06 17:54:07.526476] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.865 [2024-12-06 17:54:07.526508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.865 [2024-12-06 17:54:07.526524] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.865 [2024-12-06 17:54:07.526795] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.865 [2024-12-06 17:54:07.527022] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.865 [2024-12-06 17:54:07.527042] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.865 [2024-12-06 17:54:07.527055] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.865 [2024-12-06 17:54:07.527067] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.865 [2024-12-06 17:54:07.539476] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.865 [2024-12-06 17:54:07.539922] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.865 [2024-12-06 17:54:07.539966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.865 [2024-12-06 17:54:07.539983] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.865 [2024-12-06 17:54:07.540228] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.865 [2024-12-06 17:54:07.540441] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.865 [2024-12-06 17:54:07.540461] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.865 [2024-12-06 17:54:07.540480] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.865 [2024-12-06 17:54:07.540495] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.865 [2024-12-06 17:54:07.553027] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.865 [2024-12-06 17:54:07.553455] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.865 [2024-12-06 17:54:07.553483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.865 [2024-12-06 17:54:07.553513] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.865 [2024-12-06 17:54:07.553770] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.865 [2024-12-06 17:54:07.554006] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.865 [2024-12-06 17:54:07.554026] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.865 [2024-12-06 17:54:07.554039] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.865 [2024-12-06 17:54:07.554050] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.865 [2024-12-06 17:54:07.566419] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.865 [2024-12-06 17:54:07.566783] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.865 [2024-12-06 17:54:07.566828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.865 [2024-12-06 17:54:07.566844] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.865 [2024-12-06 17:54:07.567095] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.865 [2024-12-06 17:54:07.567323] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.865 [2024-12-06 17:54:07.567345] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.865 [2024-12-06 17:54:07.567358] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.865 [2024-12-06 17:54:07.567369] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.866 [2024-12-06 17:54:07.579910] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.866 [2024-12-06 17:54:07.580225] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.866 [2024-12-06 17:54:07.580267] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.866 [2024-12-06 17:54:07.580283] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.866 [2024-12-06 17:54:07.580506] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.866 [2024-12-06 17:54:07.580751] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.866 [2024-12-06 17:54:07.580772] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.866 [2024-12-06 17:54:07.580785] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.866 [2024-12-06 17:54:07.580797] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.866 [2024-12-06 17:54:07.593272] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.866 [2024-12-06 17:54:07.593648] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.866 [2024-12-06 17:54:07.593700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.866 [2024-12-06 17:54:07.593717] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.866 [2024-12-06 17:54:07.593974] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.866 [2024-12-06 17:54:07.594190] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.866 [2024-12-06 17:54:07.594210] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.866 [2024-12-06 17:54:07.594227] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.866 [2024-12-06 17:54:07.594239] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.866 [2024-12-06 17:54:07.606608] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.866 [2024-12-06 17:54:07.606991] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.866 [2024-12-06 17:54:07.607020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.866 [2024-12-06 17:54:07.607037] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.866 [2024-12-06 17:54:07.607267] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.866 [2024-12-06 17:54:07.607483] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.866 [2024-12-06 17:54:07.607502] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.866 [2024-12-06 17:54:07.607514] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.866 [2024-12-06 17:54:07.607526] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.866 [2024-12-06 17:54:07.619929] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.866 [2024-12-06 17:54:07.620383] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.866 [2024-12-06 17:54:07.620411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.866 [2024-12-06 17:54:07.620427] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.866 [2024-12-06 17:54:07.620682] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.866 [2024-12-06 17:54:07.620903] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.866 [2024-12-06 17:54:07.620925] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.866 [2024-12-06 17:54:07.620954] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.866 [2024-12-06 17:54:07.620967] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.866 [2024-12-06 17:54:07.633149] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.866 [2024-12-06 17:54:07.633519] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.866 [2024-12-06 17:54:07.633547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.866 [2024-12-06 17:54:07.633564] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.866 [2024-12-06 17:54:07.633804] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.866 [2024-12-06 17:54:07.634032] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.866 [2024-12-06 17:54:07.634067] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.866 [2024-12-06 17:54:07.634081] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.866 [2024-12-06 17:54:07.634093] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.866 [2024-12-06 17:54:07.646600] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.866 [2024-12-06 17:54:07.647041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.866 [2024-12-06 17:54:07.647079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.866 [2024-12-06 17:54:07.647095] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.866 [2024-12-06 17:54:07.647325] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.866 [2024-12-06 17:54:07.647540] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.866 [2024-12-06 17:54:07.647560] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.866 [2024-12-06 17:54:07.647572] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.866 [2024-12-06 17:54:07.647584] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.866 [2024-12-06 17:54:07.659884] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.866 [2024-12-06 17:54:07.660317] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.866 [2024-12-06 17:54:07.660343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.866 [2024-12-06 17:54:07.660374] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.866 [2024-12-06 17:54:07.660597] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.866 [2024-12-06 17:54:07.660848] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.867 [2024-12-06 17:54:07.660869] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.867 [2024-12-06 17:54:07.660882] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.867 [2024-12-06 17:54:07.660895] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.867 [2024-12-06 17:54:07.673406] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.867 [2024-12-06 17:54:07.673741] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.867 [2024-12-06 17:54:07.673769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.867 [2024-12-06 17:54:07.673786] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.867 [2024-12-06 17:54:07.674002] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.867 [2024-12-06 17:54:07.674224] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.867 [2024-12-06 17:54:07.674244] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.867 [2024-12-06 17:54:07.674256] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.867 [2024-12-06 17:54:07.674268] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:25.867 [2024-12-06 17:54:07.686696] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:25.867 [2024-12-06 17:54:07.687142] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:25.867 [2024-12-06 17:54:07.687190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:25.867 [2024-12-06 17:54:07.687211] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:25.867 [2024-12-06 17:54:07.687473] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:25.867 [2024-12-06 17:54:07.687696] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:25.867 [2024-12-06 17:54:07.687719] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:25.867 [2024-12-06 17:54:07.687732] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:25.867 [2024-12-06 17:54:07.687744] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.148 [2024-12-06 17:54:07.700355] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.148 [2024-12-06 17:54:07.700706] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.148 [2024-12-06 17:54:07.700735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.148 [2024-12-06 17:54:07.700751] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.148 [2024-12-06 17:54:07.700967] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.148 [2024-12-06 17:54:07.701196] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.148 [2024-12-06 17:54:07.701217] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.148 [2024-12-06 17:54:07.701232] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.148 [2024-12-06 17:54:07.701244] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.148 [2024-12-06 17:54:07.713765] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.148 [2024-12-06 17:54:07.714166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.148 [2024-12-06 17:54:07.714212] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.148 [2024-12-06 17:54:07.714228] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.148 [2024-12-06 17:54:07.714497] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.148 [2024-12-06 17:54:07.714720] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.148 [2024-12-06 17:54:07.714741] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.148 [2024-12-06 17:54:07.714755] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.148 [2024-12-06 17:54:07.714767] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.148 [2024-12-06 17:54:07.726943] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.148 [2024-12-06 17:54:07.727293] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.148 [2024-12-06 17:54:07.727341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.148 [2024-12-06 17:54:07.727376] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.148 [2024-12-06 17:54:07.727592] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.148 [2024-12-06 17:54:07.727867] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.148 [2024-12-06 17:54:07.727890] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.148 [2024-12-06 17:54:07.727904] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.148 [2024-12-06 17:54:07.727917] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.148 [2024-12-06 17:54:07.740134] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.148 [2024-12-06 17:54:07.740530] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.148 [2024-12-06 17:54:07.740558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.148 [2024-12-06 17:54:07.740573] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.148 [2024-12-06 17:54:07.740814] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.148 [2024-12-06 17:54:07.741067] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.148 [2024-12-06 17:54:07.741086] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.148 [2024-12-06 17:54:07.741099] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.148 [2024-12-06 17:54:07.741110] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.148 [2024-12-06 17:54:07.753380] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.148 [2024-12-06 17:54:07.753817] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.148 [2024-12-06 17:54:07.753861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.148 [2024-12-06 17:54:07.753880] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.148 [2024-12-06 17:54:07.754135] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.148 [2024-12-06 17:54:07.754350] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.149 [2024-12-06 17:54:07.754370] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.149 [2024-12-06 17:54:07.754382] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.149 [2024-12-06 17:54:07.754393] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.149 [2024-12-06 17:54:07.766703] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.149 [2024-12-06 17:54:07.767123] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.149 [2024-12-06 17:54:07.767151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.149 [2024-12-06 17:54:07.767167] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.149 [2024-12-06 17:54:07.767390] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.149 [2024-12-06 17:54:07.767610] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.149 [2024-12-06 17:54:07.767629] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.149 [2024-12-06 17:54:07.767673] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.149 [2024-12-06 17:54:07.767690] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.149 [2024-12-06 17:54:07.779966] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.149 [2024-12-06 17:54:07.780333] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.149 [2024-12-06 17:54:07.780361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.149 [2024-12-06 17:54:07.780377] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.149 [2024-12-06 17:54:07.780613] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.149 [2024-12-06 17:54:07.780851] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.149 [2024-12-06 17:54:07.780871] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.149 [2024-12-06 17:54:07.780883] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.149 [2024-12-06 17:54:07.780895] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.149 [2024-12-06 17:54:07.793248] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.149 [2024-12-06 17:54:07.793742] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.149 [2024-12-06 17:54:07.793771] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.149 [2024-12-06 17:54:07.793787] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.149 [2024-12-06 17:54:07.794040] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.149 [2024-12-06 17:54:07.794235] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.149 [2024-12-06 17:54:07.794254] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.149 [2024-12-06 17:54:07.794274] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.149 [2024-12-06 17:54:07.794285] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.149 [2024-12-06 17:54:07.806494] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.149 [2024-12-06 17:54:07.806900] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.149 [2024-12-06 17:54:07.806929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.149 [2024-12-06 17:54:07.806945] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.149 [2024-12-06 17:54:07.807193] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.149 [2024-12-06 17:54:07.807388] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.149 [2024-12-06 17:54:07.807406] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.149 [2024-12-06 17:54:07.807418] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.149 [2024-12-06 17:54:07.807429] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.149 [2024-12-06 17:54:07.819875] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.149 [2024-12-06 17:54:07.820197] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.149 [2024-12-06 17:54:07.820238] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.149 [2024-12-06 17:54:07.820254] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.149 [2024-12-06 17:54:07.820456] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.149 [2024-12-06 17:54:07.820708] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.149 [2024-12-06 17:54:07.820729] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.149 [2024-12-06 17:54:07.820741] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.149 [2024-12-06 17:54:07.820753] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.149 [2024-12-06 17:54:07.833178] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.149 [2024-12-06 17:54:07.833542] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.149 [2024-12-06 17:54:07.833582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.149 [2024-12-06 17:54:07.833597] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.149 [2024-12-06 17:54:07.833854] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.149 [2024-12-06 17:54:07.834067] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.149 [2024-12-06 17:54:07.834086] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.149 [2024-12-06 17:54:07.834098] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.149 [2024-12-06 17:54:07.834109] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.149 [2024-12-06 17:54:07.846473] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.149 [2024-12-06 17:54:07.847010] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.149 [2024-12-06 17:54:07.847056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.149 [2024-12-06 17:54:07.847072] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.149 [2024-12-06 17:54:07.847336] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.149 [2024-12-06 17:54:07.847531] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.149 [2024-12-06 17:54:07.847550] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.149 [2024-12-06 17:54:07.847563] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.149 [2024-12-06 17:54:07.847574] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.149 [2024-12-06 17:54:07.859776] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.149 [2024-12-06 17:54:07.860187] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.149 [2024-12-06 17:54:07.860229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.150 [2024-12-06 17:54:07.860251] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.150 [2024-12-06 17:54:07.860496] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.150 [2024-12-06 17:54:07.860724] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.150 [2024-12-06 17:54:07.860746] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.150 [2024-12-06 17:54:07.860759] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.150 [2024-12-06 17:54:07.860771] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.150 [2024-12-06 17:54:07.873188] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.150 [2024-12-06 17:54:07.873581] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.150 [2024-12-06 17:54:07.873625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.150 [2024-12-06 17:54:07.873642] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.150 [2024-12-06 17:54:07.873921] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.150 [2024-12-06 17:54:07.874144] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.150 [2024-12-06 17:54:07.874180] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.150 [2024-12-06 17:54:07.874193] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.150 [2024-12-06 17:54:07.874205] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.150 [2024-12-06 17:54:07.886480] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.150 [2024-12-06 17:54:07.886956] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.150 [2024-12-06 17:54:07.886986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.150 [2024-12-06 17:54:07.887003] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.150 [2024-12-06 17:54:07.887265] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.150 [2024-12-06 17:54:07.887493] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.150 [2024-12-06 17:54:07.887514] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.150 [2024-12-06 17:54:07.887528] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.150 [2024-12-06 17:54:07.887540] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.150 [2024-12-06 17:54:07.899845] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.150 [2024-12-06 17:54:07.900273] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.150 [2024-12-06 17:54:07.900326] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.150 [2024-12-06 17:54:07.900343] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.150 [2024-12-06 17:54:07.900604] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.150 [2024-12-06 17:54:07.900834] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.150 [2024-12-06 17:54:07.900854] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.150 [2024-12-06 17:54:07.900867] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.150 [2024-12-06 17:54:07.900879] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.150 [2024-12-06 17:54:07.913201] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.150 [2024-12-06 17:54:07.913575] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.150 [2024-12-06 17:54:07.913603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.150 [2024-12-06 17:54:07.913620] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.150 [2024-12-06 17:54:07.913904] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.150 [2024-12-06 17:54:07.914136] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.150 [2024-12-06 17:54:07.914155] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.150 [2024-12-06 17:54:07.914168] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.150 [2024-12-06 17:54:07.914179] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.150 [2024-12-06 17:54:07.926327] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.150 [2024-12-06 17:54:07.926688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.150 [2024-12-06 17:54:07.926765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.150 [2024-12-06 17:54:07.926783] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.150 [2024-12-06 17:54:07.927027] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.150 [2024-12-06 17:54:07.927279] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.150 [2024-12-06 17:54:07.927300] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.150 [2024-12-06 17:54:07.927314] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.150 [2024-12-06 17:54:07.927326] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.150 [2024-12-06 17:54:07.939485] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.150 [2024-12-06 17:54:07.939944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.150 [2024-12-06 17:54:07.939988] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.150 [2024-12-06 17:54:07.940005] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.150 [2024-12-06 17:54:07.940272] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.150 [2024-12-06 17:54:07.940466] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.150 [2024-12-06 17:54:07.940486] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.150 [2024-12-06 17:54:07.940503] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.150 [2024-12-06 17:54:07.940514] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.150 [2024-12-06 17:54:07.953245] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.150 [2024-12-06 17:54:07.953641] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.150 [2024-12-06 17:54:07.953678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.150 [2024-12-06 17:54:07.953696] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.150 [2024-12-06 17:54:07.953913] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.150 [2024-12-06 17:54:07.954155] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.150 [2024-12-06 17:54:07.954176] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.150 [2024-12-06 17:54:07.954203] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.151 [2024-12-06 17:54:07.954216] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.151 [2024-12-06 17:54:07.966906] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.151 [2024-12-06 17:54:07.967223] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.151 [2024-12-06 17:54:07.967249] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.151 [2024-12-06 17:54:07.967265] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.151 [2024-12-06 17:54:07.967468] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.151 [2024-12-06 17:54:07.967728] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.151 [2024-12-06 17:54:07.967751] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.151 [2024-12-06 17:54:07.967765] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.151 [2024-12-06 17:54:07.967778] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.151 [2024-12-06 17:54:07.980449] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.151 [2024-12-06 17:54:07.980823] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.151 [2024-12-06 17:54:07.980859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.151 [2024-12-06 17:54:07.980882] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.151 [2024-12-06 17:54:07.981122] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.151 [2024-12-06 17:54:07.981381] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.151 [2024-12-06 17:54:07.981400] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.151 [2024-12-06 17:54:07.981413] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.151 [2024-12-06 17:54:07.981425] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.410 [2024-12-06 17:54:07.994044] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.410 [2024-12-06 17:54:07.994432] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.410 [2024-12-06 17:54:07.994475] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.410 [2024-12-06 17:54:07.994491] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.410 [2024-12-06 17:54:07.994742] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.410 [2024-12-06 17:54:07.994978] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.410 [2024-12-06 17:54:07.994998] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.410 [2024-12-06 17:54:07.995011] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.410 [2024-12-06 17:54:07.995038] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.410 [2024-12-06 17:54:08.007343] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.410 [2024-12-06 17:54:08.007719] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.410 [2024-12-06 17:54:08.007746] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.410 [2024-12-06 17:54:08.007763] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.410 [2024-12-06 17:54:08.007998] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.410 [2024-12-06 17:54:08.008193] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.410 [2024-12-06 17:54:08.008212] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.410 [2024-12-06 17:54:08.008225] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.410 [2024-12-06 17:54:08.008236] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.410 [2024-12-06 17:54:08.020602] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.410 [2024-12-06 17:54:08.021078] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.410 [2024-12-06 17:54:08.021127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.410 [2024-12-06 17:54:08.021143] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.410 [2024-12-06 17:54:08.021412] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.410 [2024-12-06 17:54:08.021607] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.410 [2024-12-06 17:54:08.021626] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.410 [2024-12-06 17:54:08.021638] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.410 [2024-12-06 17:54:08.021673] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.410 [2024-12-06 17:54:08.033897] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.410 [2024-12-06 17:54:08.034359] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.410 [2024-12-06 17:54:08.034407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.410 [2024-12-06 17:54:08.034428] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.410 [2024-12-06 17:54:08.034706] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.410 [2024-12-06 17:54:08.034927] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.410 [2024-12-06 17:54:08.034947] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.410 [2024-12-06 17:54:08.034960] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.410 [2024-12-06 17:54:08.034972] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.410 [2024-12-06 17:54:08.047292] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.410 [2024-12-06 17:54:08.047747] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.410 [2024-12-06 17:54:08.047791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.410 [2024-12-06 17:54:08.047808] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.410 [2024-12-06 17:54:08.048057] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.410 [2024-12-06 17:54:08.048271] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.410 [2024-12-06 17:54:08.048295] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.410 [2024-12-06 17:54:08.048309] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.410 [2024-12-06 17:54:08.048320] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.410 [2024-12-06 17:54:08.060707] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.410 [2024-12-06 17:54:08.061118] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.410 [2024-12-06 17:54:08.061161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.410 [2024-12-06 17:54:08.061177] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.410 [2024-12-06 17:54:08.061429] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.410 [2024-12-06 17:54:08.061630] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.410 [2024-12-06 17:54:08.061677] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.410 [2024-12-06 17:54:08.061697] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.410 [2024-12-06 17:54:08.061724] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.410 [2024-12-06 17:54:08.073927] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.410 [2024-12-06 17:54:08.074310] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.410 [2024-12-06 17:54:08.074354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.410 [2024-12-06 17:54:08.074369] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.410 [2024-12-06 17:54:08.074611] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.410 [2024-12-06 17:54:08.074869] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.410 [2024-12-06 17:54:08.074896] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.411 [2024-12-06 17:54:08.074914] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.411 [2024-12-06 17:54:08.074927] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.411 [2024-12-06 17:54:08.087109] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.411 [2024-12-06 17:54:08.087544] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.411 [2024-12-06 17:54:08.087587] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.411 [2024-12-06 17:54:08.087604] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.411 [2024-12-06 17:54:08.087848] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.411 [2024-12-06 17:54:08.088095] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.411 [2024-12-06 17:54:08.088114] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.411 [2024-12-06 17:54:08.088126] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.411 [2024-12-06 17:54:08.088137] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.411 [2024-12-06 17:54:08.100273] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.411 [2024-12-06 17:54:08.100711] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.411 [2024-12-06 17:54:08.100739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.411 [2024-12-06 17:54:08.100755] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.411 [2024-12-06 17:54:08.101009] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.411 [2024-12-06 17:54:08.101220] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.411 [2024-12-06 17:54:08.101239] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.411 [2024-12-06 17:54:08.101251] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.411 [2024-12-06 17:54:08.101262] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.411 [2024-12-06 17:54:08.113594] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.411 [2024-12-06 17:54:08.114034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.411 [2024-12-06 17:54:08.114076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.411 [2024-12-06 17:54:08.114093] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.411 [2024-12-06 17:54:08.114337] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.411 [2024-12-06 17:54:08.114532] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.411 [2024-12-06 17:54:08.114550] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.411 [2024-12-06 17:54:08.114562] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.411 [2024-12-06 17:54:08.114578] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.411 [2024-12-06 17:54:08.126882] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.411 [2024-12-06 17:54:08.127340] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.411 [2024-12-06 17:54:08.127384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.411 [2024-12-06 17:54:08.127399] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.411 [2024-12-06 17:54:08.127649] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.411 [2024-12-06 17:54:08.127858] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.411 [2024-12-06 17:54:08.127878] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.411 [2024-12-06 17:54:08.127890] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.411 [2024-12-06 17:54:08.127902] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.411 [2024-12-06 17:54:08.140017] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.411 [2024-12-06 17:54:08.140451] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.411 [2024-12-06 17:54:08.140506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.411 [2024-12-06 17:54:08.140522] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.411 [2024-12-06 17:54:08.140793] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.411 [2024-12-06 17:54:08.141029] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.411 [2024-12-06 17:54:08.141050] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.411 [2024-12-06 17:54:08.141063] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.411 [2024-12-06 17:54:08.141075] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.411 [2024-12-06 17:54:08.153302] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.411 [2024-12-06 17:54:08.153635] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.411 [2024-12-06 17:54:08.153661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.411 [2024-12-06 17:54:08.153702] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.411 [2024-12-06 17:54:08.153945] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.411 [2024-12-06 17:54:08.154158] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.411 [2024-12-06 17:54:08.154176] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.411 [2024-12-06 17:54:08.154189] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.411 [2024-12-06 17:54:08.154200] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.411 [2024-12-06 17:54:08.166705] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.411 [2024-12-06 17:54:08.167176] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.411 [2024-12-06 17:54:08.167228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.411 [2024-12-06 17:54:08.167244] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.411 [2024-12-06 17:54:08.167508] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.411 [2024-12-06 17:54:08.167731] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.411 [2024-12-06 17:54:08.167751] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.411 [2024-12-06 17:54:08.167764] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.411 [2024-12-06 17:54:08.167776] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.411 [2024-12-06 17:54:08.179930] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.411 [2024-12-06 17:54:08.180356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.411 [2024-12-06 17:54:08.180406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.411 [2024-12-06 17:54:08.180421] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.412 [2024-12-06 17:54:08.180680] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.412 [2024-12-06 17:54:08.180895] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.412 [2024-12-06 17:54:08.180916] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.412 [2024-12-06 17:54:08.180929] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.412 [2024-12-06 17:54:08.180941] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.412 [2024-12-06 17:54:08.193140] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.412 [2024-12-06 17:54:08.193488] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.412 [2024-12-06 17:54:08.193514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.412 [2024-12-06 17:54:08.193529] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.412 [2024-12-06 17:54:08.193793] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.412 [2024-12-06 17:54:08.194035] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.412 [2024-12-06 17:54:08.194054] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.412 [2024-12-06 17:54:08.194067] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.412 [2024-12-06 17:54:08.194078] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.412 [2024-12-06 17:54:08.206373] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.412 [2024-12-06 17:54:08.206718] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.412 [2024-12-06 17:54:08.206747] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.412 [2024-12-06 17:54:08.206768] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.412 [2024-12-06 17:54:08.207013] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.412 [2024-12-06 17:54:08.207214] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.412 [2024-12-06 17:54:08.207234] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.412 [2024-12-06 17:54:08.207246] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.412 [2024-12-06 17:54:08.207258] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.412 [2024-12-06 17:54:08.219550] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.412 [2024-12-06 17:54:08.219912] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.412 [2024-12-06 17:54:08.219957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.412 [2024-12-06 17:54:08.219974] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.412 [2024-12-06 17:54:08.220248] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.412 [2024-12-06 17:54:08.220450] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.412 [2024-12-06 17:54:08.220469] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.412 [2024-12-06 17:54:08.220481] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.412 [2024-12-06 17:54:08.220493] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.412 [2024-12-06 17:54:08.232747] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.412 [2024-12-06 17:54:08.233107] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.412 [2024-12-06 17:54:08.233184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.412 [2024-12-06 17:54:08.233201] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.412 [2024-12-06 17:54:08.233449] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.412 [2024-12-06 17:54:08.233652] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.412 [2024-12-06 17:54:08.233698] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.412 [2024-12-06 17:54:08.233713] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.412 [2024-12-06 17:54:08.233725] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.412 [2024-12-06 17:54:08.246182] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.412 [2024-12-06 17:54:08.246562] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.412 [2024-12-06 17:54:08.246591] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.412 [2024-12-06 17:54:08.246607] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.670 [2024-12-06 17:54:08.246833] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.670 [2024-12-06 17:54:08.247084] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.670 [2024-12-06 17:54:08.247110] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.670 [2024-12-06 17:54:08.247124] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.670 [2024-12-06 17:54:08.247136] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.670 [2024-12-06 17:54:08.259326] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.670 [2024-12-06 17:54:08.259696] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.670 [2024-12-06 17:54:08.259739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.670 [2024-12-06 17:54:08.259754] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.670 [2024-12-06 17:54:08.260022] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.670 [2024-12-06 17:54:08.260217] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.670 [2024-12-06 17:54:08.260236] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.670 [2024-12-06 17:54:08.260248] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.670 [2024-12-06 17:54:08.260259] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.671 [2024-12-06 17:54:08.272492] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.671 [2024-12-06 17:54:08.272843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.671 [2024-12-06 17:54:08.272871] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.671 [2024-12-06 17:54:08.272888] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.671 [2024-12-06 17:54:08.273111] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.671 [2024-12-06 17:54:08.273321] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.671 [2024-12-06 17:54:08.273339] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.671 [2024-12-06 17:54:08.273352] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.671 [2024-12-06 17:54:08.273363] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.671 [2024-12-06 17:54:08.285643] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.671 [2024-12-06 17:54:08.286142] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.671 [2024-12-06 17:54:08.286184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.671 [2024-12-06 17:54:08.286201] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.671 [2024-12-06 17:54:08.286452] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.671 [2024-12-06 17:54:08.286688] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.671 [2024-12-06 17:54:08.286721] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.671 [2024-12-06 17:54:08.286735] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.671 [2024-12-06 17:54:08.286752] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.671 [2024-12-06 17:54:08.298713] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.671 [2024-12-06 17:54:08.299111] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.671 [2024-12-06 17:54:08.299138] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.671 [2024-12-06 17:54:08.299153] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.671 [2024-12-06 17:54:08.299378] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.671 [2024-12-06 17:54:08.299589] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.671 [2024-12-06 17:54:08.299608] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.671 [2024-12-06 17:54:08.299620] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.671 [2024-12-06 17:54:08.299631] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.671 [2024-12-06 17:54:08.311749] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.671 [2024-12-06 17:54:08.312177] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.671 [2024-12-06 17:54:08.312204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.671 [2024-12-06 17:54:08.312235] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.671 [2024-12-06 17:54:08.312476] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.671 [2024-12-06 17:54:08.312712] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.671 [2024-12-06 17:54:08.312732] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.671 [2024-12-06 17:54:08.312745] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.671 [2024-12-06 17:54:08.312756] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.671 [2024-12-06 17:54:08.324858] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.671 [2024-12-06 17:54:08.325349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.671 [2024-12-06 17:54:08.325390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.671 [2024-12-06 17:54:08.325407] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.671 [2024-12-06 17:54:08.325658] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.671 [2024-12-06 17:54:08.325910] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.671 [2024-12-06 17:54:08.325931] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.671 [2024-12-06 17:54:08.325944] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.671 [2024-12-06 17:54:08.325956] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.671 [2024-12-06 17:54:08.337937] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.671 [2024-12-06 17:54:08.338274] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.671 [2024-12-06 17:54:08.338302] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.671 [2024-12-06 17:54:08.338318] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.671 [2024-12-06 17:54:08.338541] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.671 [2024-12-06 17:54:08.338797] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.671 [2024-12-06 17:54:08.338818] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.671 [2024-12-06 17:54:08.338830] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.671 [2024-12-06 17:54:08.338843] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.671 [2024-12-06 17:54:08.351196] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.671 [2024-12-06 17:54:08.351571] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.671 [2024-12-06 17:54:08.351613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.671 [2024-12-06 17:54:08.351629] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.671 [2024-12-06 17:54:08.351916] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.671 [2024-12-06 17:54:08.352149] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.671 [2024-12-06 17:54:08.352169] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.671 [2024-12-06 17:54:08.352181] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.671 [2024-12-06 17:54:08.352192] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.671 [2024-12-06 17:54:08.364382] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.671 [2024-12-06 17:54:08.364756] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.671 [2024-12-06 17:54:08.364799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.671 [2024-12-06 17:54:08.364815] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.671 [2024-12-06 17:54:08.365084] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.671 [2024-12-06 17:54:08.365279] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.671 [2024-12-06 17:54:08.365298] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.671 [2024-12-06 17:54:08.365310] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.671 [2024-12-06 17:54:08.365321] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.671 [2024-12-06 17:54:08.377445] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.671 [2024-12-06 17:54:08.377845] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.671 [2024-12-06 17:54:08.377887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.671 [2024-12-06 17:54:08.377904] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.671 [2024-12-06 17:54:08.378132] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.671 [2024-12-06 17:54:08.378343] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.671 [2024-12-06 17:54:08.378362] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.671 [2024-12-06 17:54:08.378374] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.671 [2024-12-06 17:54:08.378385] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.671 [2024-12-06 17:54:08.390547] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.671 [2024-12-06 17:54:08.390916] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.671 [2024-12-06 17:54:08.390944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.671 [2024-12-06 17:54:08.390960] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.671 [2024-12-06 17:54:08.391191] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.671 [2024-12-06 17:54:08.391448] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.671 [2024-12-06 17:54:08.391469] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.672 [2024-12-06 17:54:08.391483] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.672 [2024-12-06 17:54:08.391496] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.672 [2024-12-06 17:54:08.403927] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.672 [2024-12-06 17:54:08.404438] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.672 [2024-12-06 17:54:08.404465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.672 [2024-12-06 17:54:08.404496] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.672 [2024-12-06 17:54:08.404778] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.672 [2024-12-06 17:54:08.404992] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.672 [2024-12-06 17:54:08.405013] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.672 [2024-12-06 17:54:08.405026] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.672 [2024-12-06 17:54:08.405039] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.672 [2024-12-06 17:54:08.416996] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.672 [2024-12-06 17:54:08.417332] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.672 [2024-12-06 17:54:08.417359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.672 [2024-12-06 17:54:08.417375] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.672 [2024-12-06 17:54:08.417598] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.672 [2024-12-06 17:54:08.417860] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.672 [2024-12-06 17:54:08.417888] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.672 [2024-12-06 17:54:08.417902] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.672 [2024-12-06 17:54:08.417915] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.672 [2024-12-06 17:54:08.430139] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.672 [2024-12-06 17:54:08.430470] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.672 [2024-12-06 17:54:08.430497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.672 [2024-12-06 17:54:08.430513] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.672 [2024-12-06 17:54:08.430764] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.672 [2024-12-06 17:54:08.430982] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.672 [2024-12-06 17:54:08.431002] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.672 [2024-12-06 17:54:08.431015] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.672 [2024-12-06 17:54:08.431026] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.672 [2024-12-06 17:54:08.443238] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.672 [2024-12-06 17:54:08.443668] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.672 [2024-12-06 17:54:08.443696] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.672 [2024-12-06 17:54:08.443711] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.672 [2024-12-06 17:54:08.443947] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.672 [2024-12-06 17:54:08.444159] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.672 [2024-12-06 17:54:08.444178] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.672 [2024-12-06 17:54:08.444190] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.672 [2024-12-06 17:54:08.444201] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.672 [2024-12-06 17:54:08.456385] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.672 [2024-12-06 17:54:08.456810] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.672 [2024-12-06 17:54:08.456838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.672 [2024-12-06 17:54:08.456853] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.672 [2024-12-06 17:54:08.457092] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.672 [2024-12-06 17:54:08.457303] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.672 [2024-12-06 17:54:08.457322] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.672 [2024-12-06 17:54:08.457334] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.672 [2024-12-06 17:54:08.457349] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.672 [2024-12-06 17:54:08.469479] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.672 [2024-12-06 17:54:08.469851] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.672 [2024-12-06 17:54:08.469895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.672 [2024-12-06 17:54:08.469911] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.672 [2024-12-06 17:54:08.470188] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.672 [2024-12-06 17:54:08.470389] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.672 [2024-12-06 17:54:08.470408] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.672 [2024-12-06 17:54:08.470420] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.672 [2024-12-06 17:54:08.470432] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.672 4479.00 IOPS, 17.50 MiB/s [2024-12-06T16:54:08.511Z] [2024-12-06 17:54:08.482758] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.672 [2024-12-06 17:54:08.483197] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.672 [2024-12-06 17:54:08.483240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.672 [2024-12-06 17:54:08.483255] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.672 [2024-12-06 17:54:08.483505] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.672 [2024-12-06 17:54:08.483729] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.672 [2024-12-06 17:54:08.483765] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.672 [2024-12-06 17:54:08.483778] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.672 [2024-12-06 17:54:08.483790] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.672 [2024-12-06 17:54:08.495945] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.672 [2024-12-06 17:54:08.498870] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.672 [2024-12-06 17:54:08.498924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.672 [2024-12-06 17:54:08.498953] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.672 [2024-12-06 17:54:08.499211] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.672 [2024-12-06 17:54:08.499408] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.672 [2024-12-06 17:54:08.499427] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.672 [2024-12-06 17:54:08.499439] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.672 [2024-12-06 17:54:08.499450] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.931 [2024-12-06 17:54:08.509492] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.931 [2024-12-06 17:54:08.509975] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.931 [2024-12-06 17:54:08.510041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.931 [2024-12-06 17:54:08.510058] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.931 [2024-12-06 17:54:08.510311] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.931 [2024-12-06 17:54:08.510507] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.931 [2024-12-06 17:54:08.510526] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.931 [2024-12-06 17:54:08.510538] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.931 [2024-12-06 17:54:08.510550] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.931 [2024-12-06 17:54:08.522589] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.931 [2024-12-06 17:54:08.523039] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.931 [2024-12-06 17:54:08.523095] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.931 [2024-12-06 17:54:08.523112] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.931 [2024-12-06 17:54:08.523373] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.931 [2024-12-06 17:54:08.523568] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.931 [2024-12-06 17:54:08.523587] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.931 [2024-12-06 17:54:08.523600] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.931 [2024-12-06 17:54:08.523611] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.931 [2024-12-06 17:54:08.535764] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.931 [2024-12-06 17:54:08.536251] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.931 [2024-12-06 17:54:08.536303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.931 [2024-12-06 17:54:08.536318] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.931 [2024-12-06 17:54:08.536561] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.931 [2024-12-06 17:54:08.536785] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.931 [2024-12-06 17:54:08.536806] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.931 [2024-12-06 17:54:08.536818] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.931 [2024-12-06 17:54:08.536830] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.931 [2024-12-06 17:54:08.549151] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.931 [2024-12-06 17:54:08.549495] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.931 [2024-12-06 17:54:08.549521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.931 [2024-12-06 17:54:08.549536] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.931 [2024-12-06 17:54:08.549778] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.932 [2024-12-06 17:54:08.549993] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.932 [2024-12-06 17:54:08.550012] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.932 [2024-12-06 17:54:08.550024] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.932 [2024-12-06 17:54:08.550035] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.932 [2024-12-06 17:54:08.562821] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.932 [2024-12-06 17:54:08.563295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.932 [2024-12-06 17:54:08.563341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.932 [2024-12-06 17:54:08.563357] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.932 [2024-12-06 17:54:08.563617] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.932 [2024-12-06 17:54:08.563878] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.932 [2024-12-06 17:54:08.563908] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.932 [2024-12-06 17:54:08.563929] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.932 [2024-12-06 17:54:08.563959] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.932 [2024-12-06 17:54:08.576117] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.932 [2024-12-06 17:54:08.576523] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.932 [2024-12-06 17:54:08.576566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.932 [2024-12-06 17:54:08.576582] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.932 [2024-12-06 17:54:08.576855] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.932 [2024-12-06 17:54:08.577075] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.932 [2024-12-06 17:54:08.577099] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.932 [2024-12-06 17:54:08.577114] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.932 [2024-12-06 17:54:08.577126] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.932 [2024-12-06 17:54:08.589266] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.932 [2024-12-06 17:54:08.589603] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.932 [2024-12-06 17:54:08.589631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.932 [2024-12-06 17:54:08.589647] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.932 [2024-12-06 17:54:08.589903] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.932 [2024-12-06 17:54:08.590132] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.932 [2024-12-06 17:54:08.590157] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.932 [2024-12-06 17:54:08.590170] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.932 [2024-12-06 17:54:08.590181] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.932 [2024-12-06 17:54:08.602908] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.932 [2024-12-06 17:54:08.603302] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.932 [2024-12-06 17:54:08.603330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.932 [2024-12-06 17:54:08.603347] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.932 [2024-12-06 17:54:08.603589] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.932 [2024-12-06 17:54:08.603847] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.932 [2024-12-06 17:54:08.603869] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.932 [2024-12-06 17:54:08.603882] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.932 [2024-12-06 17:54:08.603895] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.932 [2024-12-06 17:54:08.616248] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.932 [2024-12-06 17:54:08.616626] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.932 [2024-12-06 17:54:08.616655] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.932 [2024-12-06 17:54:08.616680] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.932 [2024-12-06 17:54:08.616898] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.932 [2024-12-06 17:54:08.617136] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.932 [2024-12-06 17:54:08.617155] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.932 [2024-12-06 17:54:08.617168] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.932 [2024-12-06 17:54:08.617179] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.932 [2024-12-06 17:54:08.629630] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.932 [2024-12-06 17:54:08.629993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.932 [2024-12-06 17:54:08.630021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.932 [2024-12-06 17:54:08.630052] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.932 [2024-12-06 17:54:08.630296] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.932 [2024-12-06 17:54:08.630514] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.932 [2024-12-06 17:54:08.630533] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.932 [2024-12-06 17:54:08.630545] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.932 [2024-12-06 17:54:08.630562] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.932 [2024-12-06 17:54:08.643052] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.932 [2024-12-06 17:54:08.643446] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.932 [2024-12-06 17:54:08.643474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.932 [2024-12-06 17:54:08.643499] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.932 [2024-12-06 17:54:08.643732] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.932 [2024-12-06 17:54:08.643959] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.932 [2024-12-06 17:54:08.643982] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.932 [2024-12-06 17:54:08.643996] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.932 [2024-12-06 17:54:08.644008] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.932 [2024-12-06 17:54:08.656440] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.932 [2024-12-06 17:54:08.656845] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.932 [2024-12-06 17:54:08.656875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.932 [2024-12-06 17:54:08.656902] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.932 [2024-12-06 17:54:08.657155] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.932 [2024-12-06 17:54:08.657357] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.932 [2024-12-06 17:54:08.657376] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.932 [2024-12-06 17:54:08.657389] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.932 [2024-12-06 17:54:08.657400] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.932 [2024-12-06 17:54:08.669920] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.933 [2024-12-06 17:54:08.670315] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.933 [2024-12-06 17:54:08.670358] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.933 [2024-12-06 17:54:08.670375] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.933 [2024-12-06 17:54:08.670630] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.933 [2024-12-06 17:54:08.670892] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.933 [2024-12-06 17:54:08.670915] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.933 [2024-12-06 17:54:08.670928] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.933 [2024-12-06 17:54:08.670941] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.933 [2024-12-06 17:54:08.683239] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.933 [2024-12-06 17:54:08.683604] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.933 [2024-12-06 17:54:08.683652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.933 [2024-12-06 17:54:08.683676] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.933 [2024-12-06 17:54:08.683922] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.933 [2024-12-06 17:54:08.684152] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.933 [2024-12-06 17:54:08.684171] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.933 [2024-12-06 17:54:08.684183] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.933 [2024-12-06 17:54:08.684194] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.933 [2024-12-06 17:54:08.696756] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.933 [2024-12-06 17:54:08.697209] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.933 [2024-12-06 17:54:08.697237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.933 [2024-12-06 17:54:08.697254] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.933 [2024-12-06 17:54:08.697497] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.933 [2024-12-06 17:54:08.697765] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.933 [2024-12-06 17:54:08.697787] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.933 [2024-12-06 17:54:08.697801] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.933 [2024-12-06 17:54:08.697814] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.933 [2024-12-06 17:54:08.710239] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.933 [2024-12-06 17:54:08.710610] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.933 [2024-12-06 17:54:08.710638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.933 [2024-12-06 17:54:08.710662] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.933 [2024-12-06 17:54:08.710887] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.933 [2024-12-06 17:54:08.711104] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.933 [2024-12-06 17:54:08.711124] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.933 [2024-12-06 17:54:08.711137] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.933 [2024-12-06 17:54:08.711148] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.933 [2024-12-06 17:54:08.723525] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.933 [2024-12-06 17:54:08.723866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.933 [2024-12-06 17:54:08.723896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.933 [2024-12-06 17:54:08.723913] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.933 [2024-12-06 17:54:08.724168] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.933 [2024-12-06 17:54:08.724363] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.933 [2024-12-06 17:54:08.724382] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.933 [2024-12-06 17:54:08.724394] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.933 [2024-12-06 17:54:08.724405] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.933 [2024-12-06 17:54:08.736830] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.933 [2024-12-06 17:54:08.737200] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.933 [2024-12-06 17:54:08.737237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.933 [2024-12-06 17:54:08.737253] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.933 [2024-12-06 17:54:08.737497] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.933 [2024-12-06 17:54:08.737729] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.933 [2024-12-06 17:54:08.737749] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.933 [2024-12-06 17:54:08.737762] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.933 [2024-12-06 17:54:08.737773] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.933 [2024-12-06 17:54:08.750082] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.933 [2024-12-06 17:54:08.750514] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.933 [2024-12-06 17:54:08.750558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.933 [2024-12-06 17:54:08.750575] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.933 [2024-12-06 17:54:08.750833] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.933 [2024-12-06 17:54:08.751073] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.933 [2024-12-06 17:54:08.751093] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.933 [2024-12-06 17:54:08.751113] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.933 [2024-12-06 17:54:08.751127] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:26.933 [2024-12-06 17:54:08.763399] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:26.933 [2024-12-06 17:54:08.763779] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:26.933 [2024-12-06 17:54:08.763823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:26.933 [2024-12-06 17:54:08.763839] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:26.933 [2024-12-06 17:54:08.764118] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:26.933 [2024-12-06 17:54:08.764374] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:26.933 [2024-12-06 17:54:08.764401] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:26.933 [2024-12-06 17:54:08.764422] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:26.933 [2024-12-06 17:54:08.764437] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.193 [2024-12-06 17:54:08.776820] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.193 [2024-12-06 17:54:08.777279] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.193 [2024-12-06 17:54:08.777325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.193 [2024-12-06 17:54:08.777342] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.193 [2024-12-06 17:54:08.777589] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.193 [2024-12-06 17:54:08.777856] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.193 [2024-12-06 17:54:08.777878] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.193 [2024-12-06 17:54:08.777898] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.194 [2024-12-06 17:54:08.777915] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.194 [2024-12-06 17:54:08.789912] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.194 [2024-12-06 17:54:08.790251] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.194 [2024-12-06 17:54:08.790294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.194 [2024-12-06 17:54:08.790309] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.194 [2024-12-06 17:54:08.790527] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.194 [2024-12-06 17:54:08.790785] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.194 [2024-12-06 17:54:08.790806] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.194 [2024-12-06 17:54:08.790819] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.194 [2024-12-06 17:54:08.790832] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.194 [2024-12-06 17:54:08.803619] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.194 [2024-12-06 17:54:08.803974] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.194 [2024-12-06 17:54:08.804003] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.194 [2024-12-06 17:54:08.804019] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.194 [2024-12-06 17:54:08.804264] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.194 [2024-12-06 17:54:08.804492] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.194 [2024-12-06 17:54:08.804512] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.194 [2024-12-06 17:54:08.804525] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.194 [2024-12-06 17:54:08.804538] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.194 [2024-12-06 17:54:08.817237] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.194 [2024-12-06 17:54:08.817676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.194 [2024-12-06 17:54:08.817705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.194 [2024-12-06 17:54:08.817721] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.194 [2024-12-06 17:54:08.817937] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.194 [2024-12-06 17:54:08.818195] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.194 [2024-12-06 17:54:08.818215] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.194 [2024-12-06 17:54:08.818243] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.194 [2024-12-06 17:54:08.818255] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.194 [2024-12-06 17:54:08.830510] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.194 [2024-12-06 17:54:08.830953] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.194 [2024-12-06 17:54:08.830998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.194 [2024-12-06 17:54:08.831013] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.194 [2024-12-06 17:54:08.831245] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.194 [2024-12-06 17:54:08.831439] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.194 [2024-12-06 17:54:08.831458] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.194 [2024-12-06 17:54:08.831471] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.194 [2024-12-06 17:54:08.831482] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.194 [2024-12-06 17:54:08.843867] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.194 [2024-12-06 17:54:08.844264] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.194 [2024-12-06 17:54:08.844306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.194 [2024-12-06 17:54:08.844322] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.194 [2024-12-06 17:54:08.844571] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.194 [2024-12-06 17:54:08.844807] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.194 [2024-12-06 17:54:08.844829] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.194 [2024-12-06 17:54:08.844843] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.194 [2024-12-06 17:54:08.844856] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.194 [2024-12-06 17:54:08.857133] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.194 [2024-12-06 17:54:08.857577] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.194 [2024-12-06 17:54:08.857629] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.194 [2024-12-06 17:54:08.857646] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.194 [2024-12-06 17:54:08.857919] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.194 [2024-12-06 17:54:08.858154] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.194 [2024-12-06 17:54:08.858173] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.194 [2024-12-06 17:54:08.858185] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.194 [2024-12-06 17:54:08.858197] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.194 [2024-12-06 17:54:08.870422] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.194 [2024-12-06 17:54:08.870813] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.194 [2024-12-06 17:54:08.870843] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.194 [2024-12-06 17:54:08.870859] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.194 [2024-12-06 17:54:08.871096] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.194 [2024-12-06 17:54:08.871318] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.194 [2024-12-06 17:54:08.871338] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.194 [2024-12-06 17:54:08.871351] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.194 [2024-12-06 17:54:08.871362] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.194 [2024-12-06 17:54:08.883696] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.194 [2024-12-06 17:54:08.884112] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.194 [2024-12-06 17:54:08.884170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.194 [2024-12-06 17:54:08.884186] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.194 [2024-12-06 17:54:08.884421] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.194 [2024-12-06 17:54:08.884622] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.194 [2024-12-06 17:54:08.884656] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.194 [2024-12-06 17:54:08.884687] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.194 [2024-12-06 17:54:08.884701] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.194 [2024-12-06 17:54:08.896966] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.194 [2024-12-06 17:54:08.897434] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.194 [2024-12-06 17:54:08.897463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.194 [2024-12-06 17:54:08.897480] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.194 [2024-12-06 17:54:08.897723] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.194 [2024-12-06 17:54:08.897956] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.194 [2024-12-06 17:54:08.897979] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.194 [2024-12-06 17:54:08.897993] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.194 [2024-12-06 17:54:08.898005] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.194 [2024-12-06 17:54:08.910230] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.194 [2024-12-06 17:54:08.910643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.194 [2024-12-06 17:54:08.910702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.194 [2024-12-06 17:54:08.910720] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.194 [2024-12-06 17:54:08.910950] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.195 [2024-12-06 17:54:08.911179] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.195 [2024-12-06 17:54:08.911198] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.195 [2024-12-06 17:54:08.911210] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.195 [2024-12-06 17:54:08.911221] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.195 [2024-12-06 17:54:08.923438] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.195 [2024-12-06 17:54:08.923860] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.195 [2024-12-06 17:54:08.923888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.195 [2024-12-06 17:54:08.923905] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.195 [2024-12-06 17:54:08.924148] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.195 [2024-12-06 17:54:08.924359] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.195 [2024-12-06 17:54:08.924377] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.195 [2024-12-06 17:54:08.924389] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.195 [2024-12-06 17:54:08.924400] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.195 [2024-12-06 17:54:08.936633] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.195 [2024-12-06 17:54:08.937021] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.195 [2024-12-06 17:54:08.937057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.195 [2024-12-06 17:54:08.937072] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.195 [2024-12-06 17:54:08.937309] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.195 [2024-12-06 17:54:08.937520] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.195 [2024-12-06 17:54:08.937538] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.195 [2024-12-06 17:54:08.937556] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.195 [2024-12-06 17:54:08.937567] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.195 [2024-12-06 17:54:08.949880] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.195 [2024-12-06 17:54:08.950342] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.195 [2024-12-06 17:54:08.950394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.195 [2024-12-06 17:54:08.950410] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.195 [2024-12-06 17:54:08.950709] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.195 [2024-12-06 17:54:08.950917] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.195 [2024-12-06 17:54:08.950937] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.195 [2024-12-06 17:54:08.950950] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.195 [2024-12-06 17:54:08.950961] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.195 [2024-12-06 17:54:08.963383] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.195 [2024-12-06 17:54:08.963796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.195 [2024-12-06 17:54:08.963830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.195 [2024-12-06 17:54:08.963861] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.195 [2024-12-06 17:54:08.964104] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.195 [2024-12-06 17:54:08.964304] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.195 [2024-12-06 17:54:08.964323] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.195 [2024-12-06 17:54:08.964335] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.195 [2024-12-06 17:54:08.964347] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.195 [2024-12-06 17:54:08.976888] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.195 [2024-12-06 17:54:08.977344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.195 [2024-12-06 17:54:08.977390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.195 [2024-12-06 17:54:08.977405] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.195 [2024-12-06 17:54:08.977627] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.195 [2024-12-06 17:54:08.977880] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.195 [2024-12-06 17:54:08.977902] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.195 [2024-12-06 17:54:08.977916] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.195 [2024-12-06 17:54:08.977928] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.195 [2024-12-06 17:54:08.990320] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.195 [2024-12-06 17:54:08.990807] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.195 [2024-12-06 17:54:08.990836] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.195 [2024-12-06 17:54:08.990852] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.195 [2024-12-06 17:54:08.991084] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.195 [2024-12-06 17:54:08.991302] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.195 [2024-12-06 17:54:08.991321] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.195 [2024-12-06 17:54:08.991333] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.195 [2024-12-06 17:54:08.991344] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.195 [2024-12-06 17:54:09.003812] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.195 [2024-12-06 17:54:09.004253] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.195 [2024-12-06 17:54:09.004301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.195 [2024-12-06 17:54:09.004317] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.195 [2024-12-06 17:54:09.004574] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.195 [2024-12-06 17:54:09.004818] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.195 [2024-12-06 17:54:09.004841] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.195 [2024-12-06 17:54:09.004854] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.195 [2024-12-06 17:54:09.004867] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.195 [2024-12-06 17:54:09.017367] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.195 [2024-12-06 17:54:09.017750] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.195 [2024-12-06 17:54:09.017779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.195 [2024-12-06 17:54:09.017795] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.195 [2024-12-06 17:54:09.018026] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.195 [2024-12-06 17:54:09.018258] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.195 [2024-12-06 17:54:09.018278] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.195 [2024-12-06 17:54:09.018292] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.195 [2024-12-06 17:54:09.018304] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.456 [2024-12-06 17:54:09.031088] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.456 [2024-12-06 17:54:09.031508] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.456 [2024-12-06 17:54:09.031536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.456 [2024-12-06 17:54:09.031558] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.456 [2024-12-06 17:54:09.031792] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.456 [2024-12-06 17:54:09.032040] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.456 [2024-12-06 17:54:09.032060] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.456 [2024-12-06 17:54:09.032073] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.456 [2024-12-06 17:54:09.032085] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.456 [2024-12-06 17:54:09.044549] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.456 [2024-12-06 17:54:09.044969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.456 [2024-12-06 17:54:09.044999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.456 [2024-12-06 17:54:09.045016] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.456 [2024-12-06 17:54:09.045267] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.456 [2024-12-06 17:54:09.045477] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.456 [2024-12-06 17:54:09.045497] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.456 [2024-12-06 17:54:09.045510] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.456 [2024-12-06 17:54:09.045521] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.456 [2024-12-06 17:54:09.058170] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.456 [2024-12-06 17:54:09.058563] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.456 [2024-12-06 17:54:09.058605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.456 [2024-12-06 17:54:09.058621] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.456 [2024-12-06 17:54:09.058876] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.456 [2024-12-06 17:54:09.059106] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.456 [2024-12-06 17:54:09.059125] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.456 [2024-12-06 17:54:09.059138] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.456 [2024-12-06 17:54:09.059149] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.456 [2024-12-06 17:54:09.071687] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.456 [2024-12-06 17:54:09.072093] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.456 [2024-12-06 17:54:09.072135] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.456 [2024-12-06 17:54:09.072151] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.456 [2024-12-06 17:54:09.072411] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.456 [2024-12-06 17:54:09.072619] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.456 [2024-12-06 17:54:09.072639] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.456 [2024-12-06 17:54:09.072689] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.456 [2024-12-06 17:54:09.072702] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.456 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/bdevperf.sh: line 35: 410070 Killed "${NVMF_APP[@]}" "$@" 00:38:27.456 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@36 -- # tgt_init 00:38:27.456 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@15 -- # nvmfappstart -m 0xE 00:38:27.456 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:38:27.456 [2024-12-06 17:54:09.085178] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.456 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@726 -- # xtrace_disable 00:38:27.456 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:38:27.456 [2024-12-06 17:54:09.085592] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.456 [2024-12-06 17:54:09.085635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.456 [2024-12-06 17:54:09.085651] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.456 [2024-12-06 17:54:09.085888] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.456 [2024-12-06 17:54:09.086129] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.456 [2024-12-06 17:54:09.086150] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.456 [2024-12-06 17:54:09.086162] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.456 [2024-12-06 17:54:09.086174] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.456 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@509 -- # nvmfpid=411564 00:38:27.456 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:38:27.456 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@510 -- # waitforlisten 411564 00:38:27.456 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@835 -- # '[' -z 411564 ']' 00:38:27.456 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:38:27.456 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@840 -- # local max_retries=100 00:38:27.456 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:38:27.456 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:38:27.456 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@844 -- # xtrace_disable 00:38:27.456 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:38:27.456 [2024-12-06 17:54:09.098707] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.456 [2024-12-06 17:54:09.099107] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.456 [2024-12-06 17:54:09.099136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.456 [2024-12-06 17:54:09.099152] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.456 [2024-12-06 17:54:09.099388] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.456 [2024-12-06 17:54:09.099610] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.456 [2024-12-06 17:54:09.099630] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.456 [2024-12-06 17:54:09.099643] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.456 [2024-12-06 17:54:09.099681] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.456 [2024-12-06 17:54:09.112321] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.456 [2024-12-06 17:54:09.112704] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.456 [2024-12-06 17:54:09.112734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.456 [2024-12-06 17:54:09.112750] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.456 [2024-12-06 17:54:09.112988] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.456 [2024-12-06 17:54:09.113217] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.456 [2024-12-06 17:54:09.113238] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.456 [2024-12-06 17:54:09.113251] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.456 [2024-12-06 17:54:09.113263] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.456 [2024-12-06 17:54:09.125980] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.457 [2024-12-06 17:54:09.126391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.457 [2024-12-06 17:54:09.126421] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.457 [2024-12-06 17:54:09.126438] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.457 [2024-12-06 17:54:09.126763] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.457 [2024-12-06 17:54:09.127004] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.457 [2024-12-06 17:54:09.127025] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.457 [2024-12-06 17:54:09.127054] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.457 [2024-12-06 17:54:09.127067] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.457 [2024-12-06 17:54:09.137081] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:38:27.457 [2024-12-06 17:54:09.137156] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:38:27.457 [2024-12-06 17:54:09.139492] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.457 [2024-12-06 17:54:09.139883] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.457 [2024-12-06 17:54:09.139912] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.457 [2024-12-06 17:54:09.139937] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.457 [2024-12-06 17:54:09.140172] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.457 [2024-12-06 17:54:09.140395] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.457 [2024-12-06 17:54:09.140416] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.457 [2024-12-06 17:54:09.140429] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.457 [2024-12-06 17:54:09.140441] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.457 [2024-12-06 17:54:09.153001] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.457 [2024-12-06 17:54:09.153380] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.457 [2024-12-06 17:54:09.153408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.457 [2024-12-06 17:54:09.153433] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.457 [2024-12-06 17:54:09.153656] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.457 [2024-12-06 17:54:09.153887] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.457 [2024-12-06 17:54:09.153909] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.457 [2024-12-06 17:54:09.153923] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.457 [2024-12-06 17:54:09.153935] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.457 [2024-12-06 17:54:09.166475] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.457 [2024-12-06 17:54:09.166845] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.457 [2024-12-06 17:54:09.166874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.457 [2024-12-06 17:54:09.166891] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.457 [2024-12-06 17:54:09.167107] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.457 [2024-12-06 17:54:09.167329] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.457 [2024-12-06 17:54:09.167349] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.457 [2024-12-06 17:54:09.167362] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.457 [2024-12-06 17:54:09.167374] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.457 [2024-12-06 17:54:09.180238] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.457 [2024-12-06 17:54:09.180629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.457 [2024-12-06 17:54:09.180657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.457 [2024-12-06 17:54:09.180683] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.457 [2024-12-06 17:54:09.180900] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.457 [2024-12-06 17:54:09.181141] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.457 [2024-12-06 17:54:09.181167] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.457 [2024-12-06 17:54:09.181181] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.457 [2024-12-06 17:54:09.181194] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.457 [2024-12-06 17:54:09.193767] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.457 [2024-12-06 17:54:09.194166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.457 [2024-12-06 17:54:09.194195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.457 [2024-12-06 17:54:09.194211] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.457 [2024-12-06 17:54:09.194427] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.457 [2024-12-06 17:54:09.194672] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.457 [2024-12-06 17:54:09.194694] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.457 [2024-12-06 17:54:09.194707] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.457 [2024-12-06 17:54:09.194720] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.457 [2024-12-06 17:54:09.207300] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.457 [2024-12-06 17:54:09.207632] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.457 [2024-12-06 17:54:09.207683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.457 [2024-12-06 17:54:09.207701] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.457 [2024-12-06 17:54:09.207916] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.457 [2024-12-06 17:54:09.208141] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.457 [2024-12-06 17:54:09.208162] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.457 [2024-12-06 17:54:09.208175] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.457 [2024-12-06 17:54:09.208187] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.457 [2024-12-06 17:54:09.215063] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:38:27.457 [2024-12-06 17:54:09.220984] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.457 [2024-12-06 17:54:09.221466] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.457 [2024-12-06 17:54:09.221497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.457 [2024-12-06 17:54:09.221515] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.457 [2024-12-06 17:54:09.221745] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.457 [2024-12-06 17:54:09.221986] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.457 [2024-12-06 17:54:09.222022] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.457 [2024-12-06 17:54:09.222044] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.457 [2024-12-06 17:54:09.222059] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.457 [2024-12-06 17:54:09.234573] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.457 [2024-12-06 17:54:09.235092] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.457 [2024-12-06 17:54:09.235127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.457 [2024-12-06 17:54:09.235147] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.457 [2024-12-06 17:54:09.235388] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.457 [2024-12-06 17:54:09.235614] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.457 [2024-12-06 17:54:09.235635] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.457 [2024-12-06 17:54:09.235672] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.457 [2024-12-06 17:54:09.235692] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.457 [2024-12-06 17:54:09.248006] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.457 [2024-12-06 17:54:09.248391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.457 [2024-12-06 17:54:09.248419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.457 [2024-12-06 17:54:09.248436] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.457 [2024-12-06 17:54:09.248652] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.458 [2024-12-06 17:54:09.248883] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.458 [2024-12-06 17:54:09.248905] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.458 [2024-12-06 17:54:09.248919] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.458 [2024-12-06 17:54:09.248932] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.458 [2024-12-06 17:54:09.261491] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.458 [2024-12-06 17:54:09.261854] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.458 [2024-12-06 17:54:09.261883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.458 [2024-12-06 17:54:09.261901] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.458 [2024-12-06 17:54:09.262119] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.458 [2024-12-06 17:54:09.262340] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.458 [2024-12-06 17:54:09.262362] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.458 [2024-12-06 17:54:09.262377] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.458 [2024-12-06 17:54:09.262390] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.458 [2024-12-06 17:54:09.262802] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:38:27.458 [2024-12-06 17:54:09.262839] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:38:27.458 [2024-12-06 17:54:09.262854] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:38:27.458 [2024-12-06 17:54:09.262866] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:38:27.458 [2024-12-06 17:54:09.262876] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:38:27.458 [2024-12-06 17:54:09.264245] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:38:27.458 [2024-12-06 17:54:09.264305] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:38:27.458 [2024-12-06 17:54:09.264308] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:38:27.458 [2024-12-06 17:54:09.275181] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.458 [2024-12-06 17:54:09.275671] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.458 [2024-12-06 17:54:09.275708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.458 [2024-12-06 17:54:09.275728] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.458 [2024-12-06 17:54:09.275951] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.458 [2024-12-06 17:54:09.276175] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.458 [2024-12-06 17:54:09.276197] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.458 [2024-12-06 17:54:09.276214] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.458 [2024-12-06 17:54:09.276229] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.458 [2024-12-06 17:54:09.288896] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.458 [2024-12-06 17:54:09.289386] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.458 [2024-12-06 17:54:09.289422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.458 [2024-12-06 17:54:09.289442] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.458 [2024-12-06 17:54:09.289674] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.458 [2024-12-06 17:54:09.289900] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.458 [2024-12-06 17:54:09.289922] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.458 [2024-12-06 17:54:09.289937] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.458 [2024-12-06 17:54:09.289953] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.717 [2024-12-06 17:54:09.302570] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.717 [2024-12-06 17:54:09.303088] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.717 [2024-12-06 17:54:09.303125] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.717 [2024-12-06 17:54:09.303144] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.717 [2024-12-06 17:54:09.303367] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.717 [2024-12-06 17:54:09.303590] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.717 [2024-12-06 17:54:09.303620] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.717 [2024-12-06 17:54:09.303637] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.717 [2024-12-06 17:54:09.303653] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.717 [2024-12-06 17:54:09.316225] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.717 [2024-12-06 17:54:09.316679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.717 [2024-12-06 17:54:09.316714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.717 [2024-12-06 17:54:09.316734] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.717 [2024-12-06 17:54:09.316955] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.717 [2024-12-06 17:54:09.317180] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.717 [2024-12-06 17:54:09.317211] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.717 [2024-12-06 17:54:09.317226] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.717 [2024-12-06 17:54:09.317242] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.717 [2024-12-06 17:54:09.329846] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.717 [2024-12-06 17:54:09.330329] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.717 [2024-12-06 17:54:09.330367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.717 [2024-12-06 17:54:09.330387] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.717 [2024-12-06 17:54:09.330611] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.717 [2024-12-06 17:54:09.330852] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.717 [2024-12-06 17:54:09.330875] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.717 [2024-12-06 17:54:09.330891] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.717 [2024-12-06 17:54:09.330907] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.717 [2024-12-06 17:54:09.343475] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.717 [2024-12-06 17:54:09.344023] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.717 [2024-12-06 17:54:09.344058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.717 [2024-12-06 17:54:09.344077] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.717 [2024-12-06 17:54:09.344300] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.717 [2024-12-06 17:54:09.344524] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.717 [2024-12-06 17:54:09.344545] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.717 [2024-12-06 17:54:09.344562] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.717 [2024-12-06 17:54:09.344586] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.717 [2024-12-06 17:54:09.357137] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.717 [2024-12-06 17:54:09.357513] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.717 [2024-12-06 17:54:09.357541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.717 [2024-12-06 17:54:09.357558] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.717 [2024-12-06 17:54:09.357784] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.717 [2024-12-06 17:54:09.358005] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.717 [2024-12-06 17:54:09.358026] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.717 [2024-12-06 17:54:09.358040] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.717 [2024-12-06 17:54:09.358053] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.717 [2024-12-06 17:54:09.370801] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.717 [2024-12-06 17:54:09.371141] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.717 [2024-12-06 17:54:09.371170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.718 [2024-12-06 17:54:09.371186] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.718 [2024-12-06 17:54:09.371401] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.718 [2024-12-06 17:54:09.371621] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.718 [2024-12-06 17:54:09.371642] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.718 [2024-12-06 17:54:09.371656] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.718 [2024-12-06 17:54:09.371679] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.718 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:38:27.718 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@868 -- # return 0 00:38:27.718 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:38:27.718 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@732 -- # xtrace_disable 00:38:27.718 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:38:27.718 [2024-12-06 17:54:09.384367] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.718 [2024-12-06 17:54:09.384698] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.718 [2024-12-06 17:54:09.384728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.718 [2024-12-06 17:54:09.384745] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.718 [2024-12-06 17:54:09.384976] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.718 [2024-12-06 17:54:09.385191] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.718 [2024-12-06 17:54:09.385212] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.718 [2024-12-06 17:54:09.385231] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.718 [2024-12-06 17:54:09.385246] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.718 [2024-12-06 17:54:09.398063] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.718 [2024-12-06 17:54:09.398443] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.718 [2024-12-06 17:54:09.398472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.718 [2024-12-06 17:54:09.398488] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.718 [2024-12-06 17:54:09.398715] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.718 [2024-12-06 17:54:09.398936] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.718 [2024-12-06 17:54:09.398957] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.718 [2024-12-06 17:54:09.398986] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.718 [2024-12-06 17:54:09.398999] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.718 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:38:27.718 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:38:27.718 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:27.718 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:38:27.718 [2024-12-06 17:54:09.409359] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:38:27.718 [2024-12-06 17:54:09.411703] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.718 [2024-12-06 17:54:09.412074] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.718 [2024-12-06 17:54:09.412101] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.718 [2024-12-06 17:54:09.412118] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.718 [2024-12-06 17:54:09.412334] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.718 [2024-12-06 17:54:09.412553] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.718 [2024-12-06 17:54:09.412575] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.718 [2024-12-06 17:54:09.412589] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.718 [2024-12-06 17:54:09.412602] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.718 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:27.718 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@18 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:38:27.718 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:27.718 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:38:27.718 [2024-12-06 17:54:09.425283] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.718 [2024-12-06 17:54:09.425682] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.718 [2024-12-06 17:54:09.425716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.718 [2024-12-06 17:54:09.425741] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.718 [2024-12-06 17:54:09.425975] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.718 [2024-12-06 17:54:09.426200] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.718 [2024-12-06 17:54:09.426220] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.718 [2024-12-06 17:54:09.426234] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.718 [2024-12-06 17:54:09.426247] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.718 [2024-12-06 17:54:09.438974] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.718 [2024-12-06 17:54:09.439412] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.718 [2024-12-06 17:54:09.439446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.718 [2024-12-06 17:54:09.439465] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.718 [2024-12-06 17:54:09.439696] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.718 [2024-12-06 17:54:09.439920] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.718 [2024-12-06 17:54:09.439942] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.718 [2024-12-06 17:54:09.439964] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.718 [2024-12-06 17:54:09.439979] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.718 Malloc0 00:38:27.718 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:27.718 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@19 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:38:27.718 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:27.718 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:38:27.718 [2024-12-06 17:54:09.452675] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.718 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:27.718 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@20 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:38:27.718 [2024-12-06 17:54:09.453026] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:27.718 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:27.718 [2024-12-06 17:54:09.453055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x938980 with addr=10.0.0.2, port=4420 00:38:27.718 [2024-12-06 17:54:09.453071] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x938980 is same with the state(6) to be set 00:38:27.718 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:38:27.718 [2024-12-06 17:54:09.453287] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x938980 (9): Bad file descriptor 00:38:27.718 [2024-12-06 17:54:09.453507] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:38:27.718 [2024-12-06 17:54:09.453528] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:38:27.718 [2024-12-06 17:54:09.453543] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:38:27.718 [2024-12-06 17:54:09.453564] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:38:27.718 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:27.718 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@21 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:38:27.718 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:27.718 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:38:27.719 [2024-12-06 17:54:09.464682] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:38:27.719 [2024-12-06 17:54:09.466371] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:38:27.719 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:27.719 17:54:09 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@38 -- # wait 410394 00:38:27.976 3732.50 IOPS, 14.58 MiB/s [2024-12-06T16:54:09.815Z] [2024-12-06 17:54:09.626055] bdev_nvme.c:2286:bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 1] Resetting controller successful. 00:38:29.842 4248.71 IOPS, 16.60 MiB/s [2024-12-06T16:54:12.614Z] 4814.25 IOPS, 18.81 MiB/s [2024-12-06T16:54:13.548Z] 5238.22 IOPS, 20.46 MiB/s [2024-12-06T16:54:14.921Z] 5581.70 IOPS, 21.80 MiB/s [2024-12-06T16:54:15.858Z] 5858.00 IOPS, 22.88 MiB/s [2024-12-06T16:54:16.790Z] 6095.33 IOPS, 23.81 MiB/s [2024-12-06T16:54:17.724Z] 6302.31 IOPS, 24.62 MiB/s [2024-12-06T16:54:18.656Z] 6479.64 IOPS, 25.31 MiB/s 00:38:36.817 Latency(us) 00:38:36.817 [2024-12-06T16:54:18.656Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:38:36.817 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:38:36.817 Verification LBA range: start 0x0 length 0x4000 00:38:36.817 Nvme1n1 : 15.01 6633.77 25.91 9937.20 0.00 7701.24 515.79 20777.34 00:38:36.817 [2024-12-06T16:54:18.656Z] =================================================================================================================== 00:38:36.817 [2024-12-06T16:54:18.656Z] Total : 6633.77 25.91 9937.20 0.00 7701.24 515.79 20777.34 00:38:37.074 17:54:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@39 -- # sync 00:38:37.074 17:54:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@40 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:38:37.074 17:54:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:37.074 17:54:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:38:37.075 17:54:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:37.075 17:54:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@42 -- # trap - SIGINT SIGTERM EXIT 00:38:37.075 17:54:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@44 -- # nvmftestfini 00:38:37.075 17:54:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@516 -- # nvmfcleanup 00:38:37.075 17:54:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@121 -- # sync 00:38:37.075 17:54:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:38:37.075 17:54:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@124 -- # set +e 00:38:37.075 17:54:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@125 -- # for i in {1..20} 00:38:37.075 17:54:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:38:37.075 rmmod nvme_tcp 00:38:37.075 rmmod nvme_fabrics 00:38:37.075 rmmod nvme_keyring 00:38:37.075 17:54:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:38:37.075 17:54:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@128 -- # set -e 00:38:37.075 17:54:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@129 -- # return 0 00:38:37.075 17:54:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@517 -- # '[' -n 411564 ']' 00:38:37.075 17:54:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@518 -- # killprocess 411564 00:38:37.075 17:54:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@954 -- # '[' -z 411564 ']' 00:38:37.075 17:54:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@958 -- # kill -0 411564 00:38:37.075 17:54:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@959 -- # uname 00:38:37.075 17:54:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:38:37.075 17:54:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 411564 00:38:37.075 17:54:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:38:37.075 17:54:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:38:37.075 17:54:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@972 -- # echo 'killing process with pid 411564' 00:38:37.075 killing process with pid 411564 00:38:37.075 17:54:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@973 -- # kill 411564 00:38:37.075 17:54:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@978 -- # wait 411564 00:38:37.334 17:54:19 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:38:37.334 17:54:19 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:38:37.334 17:54:19 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:38:37.334 17:54:19 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@297 -- # iptr 00:38:37.334 17:54:19 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@791 -- # iptables-save 00:38:37.334 17:54:19 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:38:37.334 17:54:19 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@791 -- # iptables-restore 00:38:37.334 17:54:19 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:38:37.334 17:54:19 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@302 -- # remove_spdk_ns 00:38:37.334 17:54:19 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:38:37.334 17:54:19 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:38:37.334 17:54:19 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:38:39.864 00:38:39.864 real 0m22.369s 00:38:39.864 user 0m59.815s 00:38:39.864 sys 0m4.189s 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1130 -- # xtrace_disable 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:38:39.864 ************************************ 00:38:39.864 END TEST nvmf_bdevperf 00:38:39.864 ************************************ 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@48 -- # run_test nvmf_target_disconnect /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/target_disconnect.sh --transport=tcp 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1111 -- # xtrace_disable 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:38:39.864 ************************************ 00:38:39.864 START TEST nvmf_target_disconnect 00:38:39.864 ************************************ 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/target_disconnect.sh --transport=tcp 00:38:39.864 * Looking for test storage... 00:38:39.864 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1711 -- # lcov --version 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@333 -- # local ver1 ver1_l 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@334 -- # local ver2 ver2_l 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@336 -- # IFS=.-: 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@336 -- # read -ra ver1 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@337 -- # IFS=.-: 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@337 -- # read -ra ver2 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@338 -- # local 'op=<' 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@340 -- # ver1_l=2 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@341 -- # ver2_l=1 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@344 -- # case "$op" in 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@345 -- # : 1 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@364 -- # (( v = 0 )) 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@365 -- # decimal 1 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@353 -- # local d=1 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@355 -- # echo 1 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@365 -- # ver1[v]=1 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@366 -- # decimal 2 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@353 -- # local d=2 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@355 -- # echo 2 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@366 -- # ver2[v]=2 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@368 -- # return 0 00:38:39.864 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:38:39.865 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:39.865 --rc genhtml_branch_coverage=1 00:38:39.865 --rc genhtml_function_coverage=1 00:38:39.865 --rc genhtml_legend=1 00:38:39.865 --rc geninfo_all_blocks=1 00:38:39.865 --rc geninfo_unexecuted_blocks=1 00:38:39.865 00:38:39.865 ' 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:38:39.865 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:39.865 --rc genhtml_branch_coverage=1 00:38:39.865 --rc genhtml_function_coverage=1 00:38:39.865 --rc genhtml_legend=1 00:38:39.865 --rc geninfo_all_blocks=1 00:38:39.865 --rc geninfo_unexecuted_blocks=1 00:38:39.865 00:38:39.865 ' 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:38:39.865 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:39.865 --rc genhtml_branch_coverage=1 00:38:39.865 --rc genhtml_function_coverage=1 00:38:39.865 --rc genhtml_legend=1 00:38:39.865 --rc geninfo_all_blocks=1 00:38:39.865 --rc geninfo_unexecuted_blocks=1 00:38:39.865 00:38:39.865 ' 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:38:39.865 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:39.865 --rc genhtml_branch_coverage=1 00:38:39.865 --rc genhtml_function_coverage=1 00:38:39.865 --rc genhtml_legend=1 00:38:39.865 --rc geninfo_all_blocks=1 00:38:39.865 --rc geninfo_unexecuted_blocks=1 00:38:39.865 00:38:39.865 ' 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@7 -- # uname -s 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@15 -- # shopt -s extglob 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- paths/export.sh@5 -- # export PATH 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@51 -- # : 0 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:38:39.865 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@55 -- # have_pci_nics=0 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@11 -- # PLUGIN_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@13 -- # MALLOC_BDEV_SIZE=64 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@14 -- # MALLOC_BLOCK_SIZE=512 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@69 -- # nvmftestinit 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@476 -- # prepare_net_devs 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@438 -- # local -g is_hw=no 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@440 -- # remove_spdk_ns 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@309 -- # xtrace_disable 00:38:39.865 17:54:21 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@10 -- # set +x 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@315 -- # pci_devs=() 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@315 -- # local -a pci_devs 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@316 -- # pci_net_devs=() 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@317 -- # pci_drivers=() 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@317 -- # local -A pci_drivers 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@319 -- # net_devs=() 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@319 -- # local -ga net_devs 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@320 -- # e810=() 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@320 -- # local -ga e810 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@321 -- # x722=() 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@321 -- # local -ga x722 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@322 -- # mlx=() 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@322 -- # local -ga mlx 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:38:41.766 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:38:41.766 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@418 -- # [[ up == up ]] 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:38:41.766 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:38:41.767 Found net devices under 0000:0a:00.0: cvl_0_0 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@418 -- # [[ up == up ]] 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:38:41.767 Found net devices under 0000:0a:00.1: cvl_0_1 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@442 -- # is_hw=yes 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:38:41.767 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:38:42.025 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:38:42.025 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:38:42.025 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:38:42.025 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:38:42.025 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:38:42.025 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:38:42.025 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:38:42.025 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:38:42.025 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:38:42.025 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.178 ms 00:38:42.025 00:38:42.025 --- 10.0.0.2 ping statistics --- 00:38:42.025 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:38:42.025 rtt min/avg/max/mdev = 0.178/0.178/0.178/0.000 ms 00:38:42.025 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:38:42.025 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:38:42.025 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.074 ms 00:38:42.025 00:38:42.025 --- 10.0.0.1 ping statistics --- 00:38:42.025 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:38:42.025 rtt min/avg/max/mdev = 0.074/0.074/0.074/0.000 ms 00:38:42.025 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:38:42.025 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@450 -- # return 0 00:38:42.025 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:38:42.025 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:38:42.025 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:38:42.025 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:38:42.025 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:38:42.025 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:38:42.025 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:38:42.026 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@70 -- # run_test nvmf_target_disconnect_tc1 nvmf_target_disconnect_tc1 00:38:42.026 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:38:42.026 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1111 -- # xtrace_disable 00:38:42.026 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@10 -- # set +x 00:38:42.026 ************************************ 00:38:42.026 START TEST nvmf_target_disconnect_tc1 00:38:42.026 ************************************ 00:38:42.026 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@1129 -- # nvmf_target_disconnect_tc1 00:38:42.026 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- host/target_disconnect.sh@32 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -q 32 -o 4096 -w randrw -M 50 -t 10 -c 0xF -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:38:42.026 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@652 -- # local es=0 00:38:42.026 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@654 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -q 32 -o 4096 -w randrw -M 50 -t 10 -c 0xF -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:38:42.026 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@640 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect 00:38:42.026 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:38:42.026 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@644 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect 00:38:42.026 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:38:42.026 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@646 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect 00:38:42.026 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:38:42.026 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@646 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect 00:38:42.026 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@646 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect ]] 00:38:42.026 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@655 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -q 32 -o 4096 -w randrw -M 50 -t 10 -c 0xF -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:38:42.026 [2024-12-06 17:54:23.837107] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:42.026 [2024-12-06 17:54:23.837169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1cdf620 with addr=10.0.0.2, port=4420 00:38:42.026 [2024-12-06 17:54:23.837199] nvme_tcp.c:2612:nvme_tcp_ctrlr_construct: *ERROR*: failed to create admin qpair 00:38:42.026 [2024-12-06 17:54:23.837235] nvme.c: 842:nvme_probe_internal: *ERROR*: NVMe ctrlr scan failed 00:38:42.026 [2024-12-06 17:54:23.837249] nvme.c: 951:spdk_nvme_probe_ext: *ERROR*: Create probe context failed 00:38:42.026 spdk_nvme_probe() failed for transport address '10.0.0.2' 00:38:42.026 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect: errors occurred 00:38:42.026 Initializing NVMe Controllers 00:38:42.026 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@655 -- # es=1 00:38:42.026 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:38:42.026 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:38:42.026 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:38:42.026 00:38:42.026 real 0m0.097s 00:38:42.026 user 0m0.047s 00:38:42.026 sys 0m0.050s 00:38:42.026 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@1130 -- # xtrace_disable 00:38:42.026 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@10 -- # set +x 00:38:42.026 ************************************ 00:38:42.026 END TEST nvmf_target_disconnect_tc1 00:38:42.026 ************************************ 00:38:42.284 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@71 -- # run_test nvmf_target_disconnect_tc2 nvmf_target_disconnect_tc2 00:38:42.284 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:38:42.284 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1111 -- # xtrace_disable 00:38:42.284 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@10 -- # set +x 00:38:42.284 ************************************ 00:38:42.284 START TEST nvmf_target_disconnect_tc2 00:38:42.284 ************************************ 00:38:42.284 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@1129 -- # nvmf_target_disconnect_tc2 00:38:42.284 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@37 -- # disconnect_init 10.0.0.2 00:38:42.284 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@17 -- # nvmfappstart -m 0xF0 00:38:42.284 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:38:42.284 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@726 -- # xtrace_disable 00:38:42.284 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:38:42.284 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@509 -- # nvmfpid=414723 00:38:42.284 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF0 00:38:42.284 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@510 -- # waitforlisten 414723 00:38:42.284 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@835 -- # '[' -z 414723 ']' 00:38:42.284 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:38:42.284 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@840 -- # local max_retries=100 00:38:42.284 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:38:42.284 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:38:42.284 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@844 -- # xtrace_disable 00:38:42.284 17:54:23 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:38:42.284 [2024-12-06 17:54:23.950814] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:38:42.284 [2024-12-06 17:54:23.950909] [ DPDK EAL parameters: nvmf -c 0xF0 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:38:42.284 [2024-12-06 17:54:24.024865] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:38:42.284 [2024-12-06 17:54:24.072091] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:38:42.284 [2024-12-06 17:54:24.072150] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:38:42.284 [2024-12-06 17:54:24.072180] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:38:42.284 [2024-12-06 17:54:24.072196] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:38:42.284 [2024-12-06 17:54:24.072206] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:38:42.284 [2024-12-06 17:54:24.073700] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 5 00:38:42.284 [2024-12-06 17:54:24.073764] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 6 00:38:42.284 [2024-12-06 17:54:24.073827] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 7 00:38:42.284 [2024-12-06 17:54:24.073830] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@868 -- # return 0 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@732 -- # xtrace_disable 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:38:42.544 Malloc0 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@21 -- # rpc_cmd nvmf_create_transport -t tcp -o 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:38:42.544 [2024-12-06 17:54:24.247403] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:38:42.544 [2024-12-06 17:54:24.275681] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@26 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@42 -- # reconnectpid=414754 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -q 32 -o 4096 -w randrw -M 50 -t 10 -c 0xF -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:38:42.544 17:54:24 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@44 -- # sleep 2 00:38:45.105 17:54:26 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@45 -- # kill -9 414723 00:38:45.105 17:54:26 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@47 -- # sleep 2 00:38:45.105 Read completed with error (sct=0, sc=8) 00:38:45.105 starting I/O failed 00:38:45.105 Read completed with error (sct=0, sc=8) 00:38:45.105 starting I/O failed 00:38:45.105 Read completed with error (sct=0, sc=8) 00:38:45.105 starting I/O failed 00:38:45.105 Read completed with error (sct=0, sc=8) 00:38:45.105 starting I/O failed 00:38:45.105 Read completed with error (sct=0, sc=8) 00:38:45.105 starting I/O failed 00:38:45.105 Read completed with error (sct=0, sc=8) 00:38:45.105 starting I/O failed 00:38:45.105 Write completed with error (sct=0, sc=8) 00:38:45.105 starting I/O failed 00:38:45.105 Read completed with error (sct=0, sc=8) 00:38:45.105 starting I/O failed 00:38:45.105 Read completed with error (sct=0, sc=8) 00:38:45.105 starting I/O failed 00:38:45.105 Write completed with error (sct=0, sc=8) 00:38:45.105 starting I/O failed 00:38:45.105 Read completed with error (sct=0, sc=8) 00:38:45.105 starting I/O failed 00:38:45.105 Write completed with error (sct=0, sc=8) 00:38:45.105 starting I/O failed 00:38:45.105 Read completed with error (sct=0, sc=8) 00:38:45.105 starting I/O failed 00:38:45.105 Read completed with error (sct=0, sc=8) 00:38:45.105 starting I/O failed 00:38:45.105 Write completed with error (sct=0, sc=8) 00:38:45.105 starting I/O failed 00:38:45.105 Write completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Write completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Write completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Write completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Write completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Write completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Write completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 [2024-12-06 17:54:26.299928] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Write completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Write completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Write completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Write completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Write completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Write completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Write completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Write completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Write completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Write completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Read completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 Write completed with error (sct=0, sc=8) 00:38:45.106 starting I/O failed 00:38:45.106 [2024-12-06 17:54:26.300229] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:38:45.106 [2024-12-06 17:54:26.300384] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.106 [2024-12-06 17:54:26.300422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.106 qpair failed and we were unable to recover it. 00:38:45.106 [2024-12-06 17:54:26.300519] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.106 [2024-12-06 17:54:26.300544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.106 qpair failed and we were unable to recover it. 00:38:45.106 [2024-12-06 17:54:26.300633] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.106 [2024-12-06 17:54:26.300675] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.106 qpair failed and we were unable to recover it. 00:38:45.106 [2024-12-06 17:54:26.300780] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.106 [2024-12-06 17:54:26.300804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.106 qpair failed and we were unable to recover it. 00:38:45.106 [2024-12-06 17:54:26.300899] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.106 [2024-12-06 17:54:26.300926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.106 qpair failed and we were unable to recover it. 00:38:45.106 [2024-12-06 17:54:26.301064] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.106 [2024-12-06 17:54:26.301089] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.106 qpair failed and we were unable to recover it. 00:38:45.106 [2024-12-06 17:54:26.301189] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.106 [2024-12-06 17:54:26.301214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.106 qpair failed and we were unable to recover it. 00:38:45.106 [2024-12-06 17:54:26.301299] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.106 [2024-12-06 17:54:26.301333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.106 qpair failed and we were unable to recover it. 00:38:45.106 [2024-12-06 17:54:26.301427] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.106 [2024-12-06 17:54:26.301452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.106 qpair failed and we were unable to recover it. 00:38:45.106 [2024-12-06 17:54:26.301532] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.106 [2024-12-06 17:54:26.301557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.106 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.301679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.107 [2024-12-06 17:54:26.301706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.107 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.301824] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.107 [2024-12-06 17:54:26.301849] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.107 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.301945] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.107 [2024-12-06 17:54:26.301971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.107 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.302066] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.107 [2024-12-06 17:54:26.302090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.107 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.302186] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.107 [2024-12-06 17:54:26.302212] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.107 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.302315] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.107 [2024-12-06 17:54:26.302341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.107 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.302464] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.107 [2024-12-06 17:54:26.302489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.107 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.302606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.107 [2024-12-06 17:54:26.302631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.107 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.302725] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.107 [2024-12-06 17:54:26.302751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.107 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.302843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.107 [2024-12-06 17:54:26.302868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.107 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.302979] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.107 [2024-12-06 17:54:26.303004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.107 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.303129] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.107 [2024-12-06 17:54:26.303154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.107 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.303293] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.107 [2024-12-06 17:54:26.303318] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.107 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.303428] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.107 [2024-12-06 17:54:26.303454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.107 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.303577] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.107 [2024-12-06 17:54:26.303601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.107 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.303692] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.107 [2024-12-06 17:54:26.303718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.107 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.303801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.107 [2024-12-06 17:54:26.303826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.107 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.303919] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.107 [2024-12-06 17:54:26.303944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.107 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.304059] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.107 [2024-12-06 17:54:26.304084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.107 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.304190] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.107 [2024-12-06 17:54:26.304215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.107 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.304332] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.107 [2024-12-06 17:54:26.304357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.107 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.304445] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.107 [2024-12-06 17:54:26.304472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.107 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.304583] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.107 [2024-12-06 17:54:26.304607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.107 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.304709] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.107 [2024-12-06 17:54:26.304735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.107 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.304822] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.107 [2024-12-06 17:54:26.304847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.107 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.304931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.107 [2024-12-06 17:54:26.304956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.107 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.305037] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.107 [2024-12-06 17:54:26.305064] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.107 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.305181] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.107 [2024-12-06 17:54:26.305207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.107 qpair failed and we were unable to recover it. 00:38:45.107 [2024-12-06 17:54:26.305317] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.305343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.305420] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.305446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.305537] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.305563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.305642] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.305676] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.305775] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.305801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.305885] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.305910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.306006] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.306031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.306150] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.306176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.306262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.306286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.306398] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.306423] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.306564] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.306594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.306731] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.306758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.306839] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.306865] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.306993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.307019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.307145] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.307170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.307262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.307288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.307390] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.307416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.307507] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.307533] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.307634] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.307660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.307768] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.307794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.307877] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.307902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.307991] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.308016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.308108] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.308134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.308221] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.308246] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.308386] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.308412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.308489] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.308514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.308599] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.308624] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.308718] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.308743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.308839] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.308864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.308986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.309011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.108 qpair failed and we were unable to recover it. 00:38:45.108 [2024-12-06 17:54:26.309101] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.108 [2024-12-06 17:54:26.309126] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.309211] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.309236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.309320] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.309345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.309436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.309461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.309573] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.309598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.309725] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.309751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.309862] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.309888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.310068] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.310098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.310217] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.310243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.310359] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.310384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.310467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.310493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.310610] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.310636] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.310754] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.310780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.310864] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.310890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.310972] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.310998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.311109] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.311135] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.311249] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.311275] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.311353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.311379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.311460] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.311486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.311600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.311626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.311711] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.311738] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.311875] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.311924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.312048] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.312075] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.312155] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.312182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.312274] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.312300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.312386] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.312412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.312533] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.312560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.312679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.312706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.312819] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.312845] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.109 [2024-12-06 17:54:26.312934] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.109 [2024-12-06 17:54:26.312960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.109 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.313101] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.313127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.313244] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.313269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.313342] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.313368] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.313457] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.313482] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.313588] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.313617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.313743] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.313769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.313874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.313900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.314030] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.314056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.314145] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.314170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.314248] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.314272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.314349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.314374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.314484] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.314509] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.314627] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.314653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.314797] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.314823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.314937] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.314961] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.315078] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.315104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.315218] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.315243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.315717] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.315744] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.315880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.315920] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.316043] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.316071] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.316187] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.316214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.316329] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.316357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.316483] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.316509] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.316649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.316685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.316800] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.316827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.316907] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.316933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.317022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.317048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.317141] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.317167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.317273] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.317298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.110 [2024-12-06 17:54:26.317376] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.110 [2024-12-06 17:54:26.317402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.110 qpair failed and we were unable to recover it. 00:38:45.111 [2024-12-06 17:54:26.317546] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.111 [2024-12-06 17:54:26.317571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.111 qpair failed and we were unable to recover it. 00:38:45.111 [2024-12-06 17:54:26.317649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.111 [2024-12-06 17:54:26.317686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.111 qpair failed and we were unable to recover it. 00:38:45.111 [2024-12-06 17:54:26.317769] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.111 [2024-12-06 17:54:26.317794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.111 qpair failed and we were unable to recover it. 00:38:45.111 [2024-12-06 17:54:26.317912] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.111 [2024-12-06 17:54:26.317937] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.111 qpair failed and we were unable to recover it. 00:38:45.111 [2024-12-06 17:54:26.318040] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.111 [2024-12-06 17:54:26.318065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.111 qpair failed and we were unable to recover it. 00:38:45.111 [2024-12-06 17:54:26.318149] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.111 [2024-12-06 17:54:26.318175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.111 qpair failed and we were unable to recover it. 00:38:45.111 [2024-12-06 17:54:26.318295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.111 [2024-12-06 17:54:26.318323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.111 qpair failed and we were unable to recover it. 00:38:45.111 [2024-12-06 17:54:26.318435] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.111 [2024-12-06 17:54:26.318461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.111 qpair failed and we were unable to recover it. 00:38:45.111 [2024-12-06 17:54:26.318548] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.111 [2024-12-06 17:54:26.318574] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.111 qpair failed and we were unable to recover it. 00:38:45.111 [2024-12-06 17:54:26.318682] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.111 [2024-12-06 17:54:26.318709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.111 qpair failed and we were unable to recover it. 00:38:45.111 [2024-12-06 17:54:26.318828] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.111 [2024-12-06 17:54:26.318854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.111 qpair failed and we were unable to recover it. 00:38:45.111 [2024-12-06 17:54:26.318937] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.111 [2024-12-06 17:54:26.318963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.111 qpair failed and we were unable to recover it. 00:38:45.111 [2024-12-06 17:54:26.319074] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.111 [2024-12-06 17:54:26.319101] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.111 qpair failed and we were unable to recover it. 00:38:45.111 [2024-12-06 17:54:26.319212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.111 [2024-12-06 17:54:26.319239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.111 qpair failed and we were unable to recover it. 00:38:45.111 [2024-12-06 17:54:26.319350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.111 [2024-12-06 17:54:26.319377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.111 qpair failed and we were unable to recover it. 00:38:45.111 [2024-12-06 17:54:26.319532] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.111 [2024-12-06 17:54:26.319558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.111 qpair failed and we were unable to recover it. 00:38:45.111 [2024-12-06 17:54:26.319678] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.111 [2024-12-06 17:54:26.319705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.111 qpair failed and we were unable to recover it. 00:38:45.111 [2024-12-06 17:54:26.319788] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.111 [2024-12-06 17:54:26.319813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.111 qpair failed and we were unable to recover it. 00:38:45.111 [2024-12-06 17:54:26.319925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.111 [2024-12-06 17:54:26.319949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.111 qpair failed and we were unable to recover it. 00:38:45.111 [2024-12-06 17:54:26.320035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.111 [2024-12-06 17:54:26.320060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.111 qpair failed and we were unable to recover it. 00:38:45.111 [2024-12-06 17:54:26.320201] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.111 [2024-12-06 17:54:26.320227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.111 qpair failed and we were unable to recover it. 00:38:45.111 [2024-12-06 17:54:26.320315] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.111 [2024-12-06 17:54:26.320340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.111 qpair failed and we were unable to recover it. 00:38:45.111 [2024-12-06 17:54:26.320452] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.111 [2024-12-06 17:54:26.320478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.111 qpair failed and we were unable to recover it. 00:38:45.111 [2024-12-06 17:54:26.320586] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.111 [2024-12-06 17:54:26.320611] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.111 qpair failed and we were unable to recover it. 00:38:45.111 [2024-12-06 17:54:26.320734] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.320761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.320866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.320891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.320984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.321010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.321099] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.321123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.321232] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.321262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.321368] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.321393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.321481] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.321507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.321589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.321614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.321722] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.321747] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.321853] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.321878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.322000] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.322026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.322106] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.322131] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.322212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.322238] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.322348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.322374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.322461] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.322486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.322607] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.322647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.322778] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.322806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.322949] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.322976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.323097] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.323124] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.323211] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.323236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.323347] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.323373] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.323456] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.323481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.323564] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.323589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.323701] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.323727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.323844] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.323869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.323956] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.323980] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.324058] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.324084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.324200] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.324225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.324337] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.112 [2024-12-06 17:54:26.324363] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.112 qpair failed and we were unable to recover it. 00:38:45.112 [2024-12-06 17:54:26.324470] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.324495] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.113 qpair failed and we were unable to recover it. 00:38:45.113 [2024-12-06 17:54:26.324634] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.324659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.113 qpair failed and we were unable to recover it. 00:38:45.113 [2024-12-06 17:54:26.324751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.324776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.113 qpair failed and we were unable to recover it. 00:38:45.113 [2024-12-06 17:54:26.324864] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.324889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.113 qpair failed and we were unable to recover it. 00:38:45.113 [2024-12-06 17:54:26.325012] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.325037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.113 qpair failed and we were unable to recover it. 00:38:45.113 [2024-12-06 17:54:26.325140] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.325166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.113 qpair failed and we were unable to recover it. 00:38:45.113 [2024-12-06 17:54:26.325278] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.325302] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.113 qpair failed and we were unable to recover it. 00:38:45.113 [2024-12-06 17:54:26.325382] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.325408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.113 qpair failed and we were unable to recover it. 00:38:45.113 [2024-12-06 17:54:26.325503] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.325527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.113 qpair failed and we were unable to recover it. 00:38:45.113 [2024-12-06 17:54:26.325606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.325632] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.113 qpair failed and we were unable to recover it. 00:38:45.113 [2024-12-06 17:54:26.325717] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.325742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.113 qpair failed and we were unable to recover it. 00:38:45.113 [2024-12-06 17:54:26.325817] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.325842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.113 qpair failed and we were unable to recover it. 00:38:45.113 [2024-12-06 17:54:26.325948] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.325973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.113 qpair failed and we were unable to recover it. 00:38:45.113 [2024-12-06 17:54:26.326047] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.326072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.113 qpair failed and we were unable to recover it. 00:38:45.113 [2024-12-06 17:54:26.326210] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.326235] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.113 qpair failed and we were unable to recover it. 00:38:45.113 [2024-12-06 17:54:26.326318] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.326343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.113 qpair failed and we were unable to recover it. 00:38:45.113 [2024-12-06 17:54:26.326432] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.326461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.113 qpair failed and we were unable to recover it. 00:38:45.113 [2024-12-06 17:54:26.326572] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.326598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.113 qpair failed and we were unable to recover it. 00:38:45.113 [2024-12-06 17:54:26.326688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.326714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.113 qpair failed and we were unable to recover it. 00:38:45.113 [2024-12-06 17:54:26.326853] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.326879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.113 qpair failed and we were unable to recover it. 00:38:45.113 [2024-12-06 17:54:26.326988] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.327012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.113 qpair failed and we were unable to recover it. 00:38:45.113 [2024-12-06 17:54:26.327133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.327159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.113 qpair failed and we were unable to recover it. 00:38:45.113 [2024-12-06 17:54:26.327251] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.327276] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.113 qpair failed and we were unable to recover it. 00:38:45.113 [2024-12-06 17:54:26.327386] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.327412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.113 qpair failed and we were unable to recover it. 00:38:45.113 [2024-12-06 17:54:26.327536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.327561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.113 qpair failed and we were unable to recover it. 00:38:45.113 [2024-12-06 17:54:26.327643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.327679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.113 qpair failed and we were unable to recover it. 00:38:45.113 [2024-12-06 17:54:26.327759] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.327784] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.113 qpair failed and we were unable to recover it. 00:38:45.113 [2024-12-06 17:54:26.327886] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.327912] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.113 qpair failed and we were unable to recover it. 00:38:45.113 [2024-12-06 17:54:26.328002] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.113 [2024-12-06 17:54:26.328026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 [2024-12-06 17:54:26.328114] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.114 [2024-12-06 17:54:26.328140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 [2024-12-06 17:54:26.328226] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.114 [2024-12-06 17:54:26.328251] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 [2024-12-06 17:54:26.328367] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.114 [2024-12-06 17:54:26.328393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 [2024-12-06 17:54:26.328482] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.114 [2024-12-06 17:54:26.328508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 [2024-12-06 17:54:26.328626] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.114 [2024-12-06 17:54:26.328651] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 [2024-12-06 17:54:26.328781] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.114 [2024-12-06 17:54:26.328806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 [2024-12-06 17:54:26.328895] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.114 [2024-12-06 17:54:26.328920] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 [2024-12-06 17:54:26.329006] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.114 [2024-12-06 17:54:26.329031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 [2024-12-06 17:54:26.329138] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.114 [2024-12-06 17:54:26.329163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 [2024-12-06 17:54:26.329256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.114 [2024-12-06 17:54:26.329281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 [2024-12-06 17:54:26.329389] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.114 [2024-12-06 17:54:26.329414] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 [2024-12-06 17:54:26.329520] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.114 [2024-12-06 17:54:26.329545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 [2024-12-06 17:54:26.329678] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.114 [2024-12-06 17:54:26.329703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 [2024-12-06 17:54:26.329796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.114 [2024-12-06 17:54:26.329822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 [2024-12-06 17:54:26.329932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.114 [2024-12-06 17:54:26.329970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 [2024-12-06 17:54:26.330079] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.114 [2024-12-06 17:54:26.330103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 [2024-12-06 17:54:26.330212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.114 [2024-12-06 17:54:26.330237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 [2024-12-06 17:54:26.330330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.114 [2024-12-06 17:54:26.330355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 [2024-12-06 17:54:26.330440] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.114 [2024-12-06 17:54:26.330464] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 [2024-12-06 17:54:26.330605] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.114 [2024-12-06 17:54:26.330630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 [2024-12-06 17:54:26.330803] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.114 [2024-12-06 17:54:26.330830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 [2024-12-06 17:54:26.330945] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.114 [2024-12-06 17:54:26.330970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 [2024-12-06 17:54:26.331089] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.114 [2024-12-06 17:54:26.331114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 [2024-12-06 17:54:26.331236] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.114 [2024-12-06 17:54:26.331261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 [2024-12-06 17:54:26.331348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.114 [2024-12-06 17:54:26.331373] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 [2024-12-06 17:54:26.331455] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.114 [2024-12-06 17:54:26.331480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 [2024-12-06 17:54:26.331571] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.114 [2024-12-06 17:54:26.331597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 [2024-12-06 17:54:26.331714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.114 [2024-12-06 17:54:26.331740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.114 qpair failed and we were unable to recover it. 00:38:45.114 Read completed with error (sct=0, sc=8) 00:38:45.114 starting I/O failed 00:38:45.114 Read completed with error (sct=0, sc=8) 00:38:45.114 starting I/O failed 00:38:45.114 Read completed with error (sct=0, sc=8) 00:38:45.114 starting I/O failed 00:38:45.114 Read completed with error (sct=0, sc=8) 00:38:45.114 starting I/O failed 00:38:45.114 Read completed with error (sct=0, sc=8) 00:38:45.114 starting I/O failed 00:38:45.114 Read completed with error (sct=0, sc=8) 00:38:45.114 starting I/O failed 00:38:45.114 Read completed with error (sct=0, sc=8) 00:38:45.114 starting I/O failed 00:38:45.115 Read completed with error (sct=0, sc=8) 00:38:45.115 starting I/O failed 00:38:45.115 Read completed with error (sct=0, sc=8) 00:38:45.115 starting I/O failed 00:38:45.115 Read completed with error (sct=0, sc=8) 00:38:45.115 starting I/O failed 00:38:45.115 Read completed with error (sct=0, sc=8) 00:38:45.115 starting I/O failed 00:38:45.115 Read completed with error (sct=0, sc=8) 00:38:45.115 starting I/O failed 00:38:45.115 Read completed with error (sct=0, sc=8) 00:38:45.115 starting I/O failed 00:38:45.115 Read completed with error (sct=0, sc=8) 00:38:45.115 starting I/O failed 00:38:45.115 Read completed with error (sct=0, sc=8) 00:38:45.115 starting I/O failed 00:38:45.115 Write completed with error (sct=0, sc=8) 00:38:45.115 starting I/O failed 00:38:45.115 Read completed with error (sct=0, sc=8) 00:38:45.115 starting I/O failed 00:38:45.115 Write completed with error (sct=0, sc=8) 00:38:45.115 starting I/O failed 00:38:45.115 Write completed with error (sct=0, sc=8) 00:38:45.115 starting I/O failed 00:38:45.115 Read completed with error (sct=0, sc=8) 00:38:45.115 starting I/O failed 00:38:45.115 Write completed with error (sct=0, sc=8) 00:38:45.115 starting I/O failed 00:38:45.115 Read completed with error (sct=0, sc=8) 00:38:45.115 starting I/O failed 00:38:45.115 Write completed with error (sct=0, sc=8) 00:38:45.115 starting I/O failed 00:38:45.115 Write completed with error (sct=0, sc=8) 00:38:45.115 starting I/O failed 00:38:45.115 Write completed with error (sct=0, sc=8) 00:38:45.115 starting I/O failed 00:38:45.115 Write completed with error (sct=0, sc=8) 00:38:45.115 starting I/O failed 00:38:45.115 Write completed with error (sct=0, sc=8) 00:38:45.115 starting I/O failed 00:38:45.115 Read completed with error (sct=0, sc=8) 00:38:45.115 starting I/O failed 00:38:45.115 Write completed with error (sct=0, sc=8) 00:38:45.115 starting I/O failed 00:38:45.115 Write completed with error (sct=0, sc=8) 00:38:45.115 starting I/O failed 00:38:45.115 Write completed with error (sct=0, sc=8) 00:38:45.115 starting I/O failed 00:38:45.115 Write completed with error (sct=0, sc=8) 00:38:45.115 starting I/O failed 00:38:45.115 [2024-12-06 17:54:26.332045] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:45.115 [2024-12-06 17:54:26.332183] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.115 [2024-12-06 17:54:26.332223] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.115 qpair failed and we were unable to recover it. 00:38:45.115 [2024-12-06 17:54:26.332344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.115 [2024-12-06 17:54:26.332372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.115 qpair failed and we were unable to recover it. 00:38:45.115 [2024-12-06 17:54:26.332501] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.115 [2024-12-06 17:54:26.332528] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.115 qpair failed and we were unable to recover it. 00:38:45.115 [2024-12-06 17:54:26.332608] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.115 [2024-12-06 17:54:26.332634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.115 qpair failed and we were unable to recover it. 00:38:45.115 [2024-12-06 17:54:26.332732] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.115 [2024-12-06 17:54:26.332760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.115 qpair failed and we were unable to recover it. 00:38:45.115 [2024-12-06 17:54:26.332849] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.115 [2024-12-06 17:54:26.332876] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.115 qpair failed and we were unable to recover it. 00:38:45.115 [2024-12-06 17:54:26.332995] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.115 [2024-12-06 17:54:26.333022] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.115 qpair failed and we were unable to recover it. 00:38:45.115 [2024-12-06 17:54:26.333147] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.115 [2024-12-06 17:54:26.333173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.115 qpair failed and we were unable to recover it. 00:38:45.115 [2024-12-06 17:54:26.333262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.115 [2024-12-06 17:54:26.333289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.115 qpair failed and we were unable to recover it. 00:38:45.115 [2024-12-06 17:54:26.333386] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.115 [2024-12-06 17:54:26.333426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.115 qpair failed and we were unable to recover it. 00:38:45.115 [2024-12-06 17:54:26.333524] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.115 [2024-12-06 17:54:26.333551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.115 qpair failed and we were unable to recover it. 00:38:45.115 [2024-12-06 17:54:26.333681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.115 [2024-12-06 17:54:26.333708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.115 qpair failed and we were unable to recover it. 00:38:45.115 [2024-12-06 17:54:26.333796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.115 [2024-12-06 17:54:26.333823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.115 qpair failed and we were unable to recover it. 00:38:45.115 [2024-12-06 17:54:26.333905] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.115 [2024-12-06 17:54:26.333930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.115 qpair failed and we were unable to recover it. 00:38:45.115 [2024-12-06 17:54:26.334041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.115 [2024-12-06 17:54:26.334066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.115 qpair failed and we were unable to recover it. 00:38:45.115 [2024-12-06 17:54:26.334151] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.115 [2024-12-06 17:54:26.334176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.115 qpair failed and we were unable to recover it. 00:38:45.115 [2024-12-06 17:54:26.334261] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.115 [2024-12-06 17:54:26.334286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.115 qpair failed and we were unable to recover it. 00:38:45.115 [2024-12-06 17:54:26.334377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.115 [2024-12-06 17:54:26.334406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.115 qpair failed and we were unable to recover it. 00:38:45.115 [2024-12-06 17:54:26.334489] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.115 [2024-12-06 17:54:26.334516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.334653] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.334684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.334801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.334827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.334970] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.334996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.335084] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.335110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.335227] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.335256] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.335373] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.335400] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.335486] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.335511] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.335621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.335647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.335748] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.335776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.335866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.335893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.336013] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.336040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.336152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.336179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.336288] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.336315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.336426] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.336453] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.336596] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.336632] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.336738] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.336765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.336857] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.336883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.337022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.337048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.337128] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.337154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.337267] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.337292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.337399] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.337426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.337515] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.337541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.337628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.337652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.337743] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.337769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.337850] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.337876] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.337974] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.338000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.338110] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.338135] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.338269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.338294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.338401] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.116 [2024-12-06 17:54:26.338426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.116 qpair failed and we were unable to recover it. 00:38:45.116 [2024-12-06 17:54:26.338536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.117 [2024-12-06 17:54:26.338562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.117 qpair failed and we were unable to recover it. 00:38:45.117 [2024-12-06 17:54:26.338707] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.117 [2024-12-06 17:54:26.338733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.117 qpair failed and we were unable to recover it. 00:38:45.117 [2024-12-06 17:54:26.338807] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.117 [2024-12-06 17:54:26.338833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.117 qpair failed and we were unable to recover it. 00:38:45.117 [2024-12-06 17:54:26.338921] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.117 [2024-12-06 17:54:26.338946] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.117 qpair failed and we were unable to recover it. 00:38:45.117 [2024-12-06 17:54:26.339033] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.117 [2024-12-06 17:54:26.339059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.117 qpair failed and we were unable to recover it. 00:38:45.117 [2024-12-06 17:54:26.339200] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.117 [2024-12-06 17:54:26.339226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.117 qpair failed and we were unable to recover it. 00:38:45.117 [2024-12-06 17:54:26.339319] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.117 [2024-12-06 17:54:26.339344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.117 qpair failed and we were unable to recover it. 00:38:45.117 [2024-12-06 17:54:26.339468] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.117 [2024-12-06 17:54:26.339493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.117 qpair failed and we were unable to recover it. 00:38:45.117 [2024-12-06 17:54:26.339577] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.117 [2024-12-06 17:54:26.339602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.117 qpair failed and we were unable to recover it. 00:38:45.117 [2024-12-06 17:54:26.339725] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.117 [2024-12-06 17:54:26.339753] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.117 qpair failed and we were unable to recover it. 00:38:45.117 [2024-12-06 17:54:26.339860] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.117 [2024-12-06 17:54:26.339886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.117 qpair failed and we were unable to recover it. 00:38:45.117 [2024-12-06 17:54:26.339996] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.117 [2024-12-06 17:54:26.340041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.117 qpair failed and we were unable to recover it. 00:38:45.117 [2024-12-06 17:54:26.340167] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.117 [2024-12-06 17:54:26.340215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.117 qpair failed and we were unable to recover it. 00:38:45.117 [2024-12-06 17:54:26.340346] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.117 [2024-12-06 17:54:26.340390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.117 qpair failed and we were unable to recover it. 00:38:45.117 [2024-12-06 17:54:26.340468] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.117 [2024-12-06 17:54:26.340494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.117 qpair failed and we were unable to recover it. 00:38:45.117 [2024-12-06 17:54:26.340587] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.117 [2024-12-06 17:54:26.340614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.117 qpair failed and we were unable to recover it. 00:38:45.117 [2024-12-06 17:54:26.340744] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.117 [2024-12-06 17:54:26.340770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.117 qpair failed and we were unable to recover it. 00:38:45.117 [2024-12-06 17:54:26.340870] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.117 [2024-12-06 17:54:26.340910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.117 qpair failed and we were unable to recover it. 00:38:45.117 [2024-12-06 17:54:26.341034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.117 [2024-12-06 17:54:26.341060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.117 qpair failed and we were unable to recover it. 00:38:45.117 [2024-12-06 17:54:26.341168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.117 [2024-12-06 17:54:26.341194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.117 qpair failed and we were unable to recover it. 00:38:45.117 [2024-12-06 17:54:26.341276] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.117 [2024-12-06 17:54:26.341301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.117 qpair failed and we were unable to recover it. 00:38:45.117 [2024-12-06 17:54:26.341415] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.117 [2024-12-06 17:54:26.341440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.117 qpair failed and we were unable to recover it. 00:38:45.117 [2024-12-06 17:54:26.341529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.117 [2024-12-06 17:54:26.341555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.117 qpair failed and we were unable to recover it. 00:38:45.117 [2024-12-06 17:54:26.341703] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.117 [2024-12-06 17:54:26.341730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.117 qpair failed and we were unable to recover it. 00:38:45.117 [2024-12-06 17:54:26.341841] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.117 [2024-12-06 17:54:26.341866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.117 qpair failed and we were unable to recover it. 00:38:45.117 [2024-12-06 17:54:26.341954] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.117 [2024-12-06 17:54:26.341981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.117 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.342125] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.118 [2024-12-06 17:54:26.342152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.118 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.342290] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.118 [2024-12-06 17:54:26.342315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.118 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.342427] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.118 [2024-12-06 17:54:26.342452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.118 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.342573] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.118 [2024-12-06 17:54:26.342598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.118 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.342681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.118 [2024-12-06 17:54:26.342707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.118 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.342802] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.118 [2024-12-06 17:54:26.342828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.118 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.342913] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.118 [2024-12-06 17:54:26.342939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.118 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.343048] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.118 [2024-12-06 17:54:26.343084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.118 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.343200] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.118 [2024-12-06 17:54:26.343225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.118 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.343311] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.118 [2024-12-06 17:54:26.343337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.118 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.343440] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.118 [2024-12-06 17:54:26.343479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.118 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.343626] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.118 [2024-12-06 17:54:26.343653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.118 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.343755] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.118 [2024-12-06 17:54:26.343794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.118 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.343891] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.118 [2024-12-06 17:54:26.343921] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.118 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.344034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.118 [2024-12-06 17:54:26.344059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.118 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.344176] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.118 [2024-12-06 17:54:26.344201] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.118 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.344288] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.118 [2024-12-06 17:54:26.344313] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.118 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.344395] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.118 [2024-12-06 17:54:26.344420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.118 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.344533] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.118 [2024-12-06 17:54:26.344558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.118 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.344681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.118 [2024-12-06 17:54:26.344708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.118 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.344794] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.118 [2024-12-06 17:54:26.344820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.118 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.344932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.118 [2024-12-06 17:54:26.344958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.118 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.345080] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.118 [2024-12-06 17:54:26.345105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.118 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.345225] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.118 [2024-12-06 17:54:26.345251] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.118 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.345375] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.118 [2024-12-06 17:54:26.345401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.118 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.345536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.118 [2024-12-06 17:54:26.345561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.118 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.345678] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.118 [2024-12-06 17:54:26.345705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.118 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.345796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.118 [2024-12-06 17:54:26.345822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.118 qpair failed and we were unable to recover it. 00:38:45.118 [2024-12-06 17:54:26.345928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.345957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.346116] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.346171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.346390] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.346450] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.346585] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.346611] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.346732] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.346759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.346852] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.346878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.346994] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.347021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.347159] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.347203] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.347358] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.347383] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.347491] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.347516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.347594] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.347620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.347730] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.347757] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.347866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.347896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.348015] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.348040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.348144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.348168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.348279] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.348306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.348382] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.348408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.348502] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.348541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.348637] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.348672] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.348762] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.348789] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.348877] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.348904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.349028] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.349054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.349169] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.349196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.349287] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.349314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.349395] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.349422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.349560] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.349585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.349726] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.349751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.349872] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.349897] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.350016] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.350040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.119 [2024-12-06 17:54:26.350126] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.119 [2024-12-06 17:54:26.350151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.119 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.350235] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.350260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.120 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.350341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.350365] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.120 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.350467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.350492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.120 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.350603] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.350628] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.120 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.350739] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.350764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.120 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.350874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.350899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.120 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.350993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.351018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.120 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.351097] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.351121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.120 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.351223] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.351249] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.120 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.351361] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.351390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.120 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.351470] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.351495] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.120 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.351579] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.351605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.120 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.351747] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.351773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.120 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.351863] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.351888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.120 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.352008] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.352033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.120 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.352176] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.352201] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.120 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.352294] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.352319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.120 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.352435] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.352464] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.120 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.352593] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.352631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.120 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.352737] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.352765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.120 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.352874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.352900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.120 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.352995] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.353036] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.120 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.353204] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.353255] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.120 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.353367] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.353394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.120 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.353473] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.353498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.120 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.353609] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.353634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.120 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.353716] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.353742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.120 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.353824] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.353849] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.120 qpair failed and we were unable to recover it. 00:38:45.120 [2024-12-06 17:54:26.353973] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.120 [2024-12-06 17:54:26.353997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.121 [2024-12-06 17:54:26.354082] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.121 [2024-12-06 17:54:26.354107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.121 [2024-12-06 17:54:26.354188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.121 [2024-12-06 17:54:26.354213] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.121 [2024-12-06 17:54:26.354299] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.121 [2024-12-06 17:54:26.354324] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.121 [2024-12-06 17:54:26.354449] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.121 [2024-12-06 17:54:26.354489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.121 [2024-12-06 17:54:26.354574] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.121 [2024-12-06 17:54:26.354603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.121 [2024-12-06 17:54:26.354703] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.121 [2024-12-06 17:54:26.354730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.121 [2024-12-06 17:54:26.354850] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.121 [2024-12-06 17:54:26.354877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.121 [2024-12-06 17:54:26.355005] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.121 [2024-12-06 17:54:26.355031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.121 [2024-12-06 17:54:26.355124] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.121 [2024-12-06 17:54:26.355151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.121 [2024-12-06 17:54:26.355302] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.121 [2024-12-06 17:54:26.355329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.121 [2024-12-06 17:54:26.355423] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.121 [2024-12-06 17:54:26.355452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.121 [2024-12-06 17:54:26.355576] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.121 [2024-12-06 17:54:26.355602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.121 [2024-12-06 17:54:26.355691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.121 [2024-12-06 17:54:26.355718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.121 [2024-12-06 17:54:26.355827] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.121 [2024-12-06 17:54:26.355873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.121 [2024-12-06 17:54:26.356050] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.121 [2024-12-06 17:54:26.356076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.121 [2024-12-06 17:54:26.356182] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.121 [2024-12-06 17:54:26.356208] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.121 [2024-12-06 17:54:26.356313] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.121 [2024-12-06 17:54:26.356342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.121 [2024-12-06 17:54:26.356429] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.121 [2024-12-06 17:54:26.356458] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.121 [2024-12-06 17:54:26.356567] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.121 [2024-12-06 17:54:26.356593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.121 [2024-12-06 17:54:26.356704] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.121 [2024-12-06 17:54:26.356730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.121 [2024-12-06 17:54:26.356820] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.121 [2024-12-06 17:54:26.356845] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.121 [2024-12-06 17:54:26.356933] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.121 [2024-12-06 17:54:26.356959] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.121 [2024-12-06 17:54:26.357034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.121 [2024-12-06 17:54:26.357060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.121 [2024-12-06 17:54:26.357168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.121 [2024-12-06 17:54:26.357193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.121 [2024-12-06 17:54:26.357279] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.121 [2024-12-06 17:54:26.357305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.121 [2024-12-06 17:54:26.357432] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.121 [2024-12-06 17:54:26.357470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.121 [2024-12-06 17:54:26.357568] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.121 [2024-12-06 17:54:26.357595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.121 [2024-12-06 17:54:26.357684] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.121 [2024-12-06 17:54:26.357713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.121 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.357826] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.357853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.357938] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.357965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.358084] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.358110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.358230] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.358257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.358374] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.358400] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.358552] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.358592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.358744] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.358776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.358865] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.358891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.359006] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.359032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.359144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.359170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.359254] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.359279] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.359373] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.359401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.359517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.359546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.359686] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.359712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.359798] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.359823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.359915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.359940] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.360028] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.360053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.360139] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.360163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.360271] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.360297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.360445] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.360470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.360584] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.360610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.360701] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.360727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.360808] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.360833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.360911] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.360937] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.361024] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.361049] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.361162] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.361187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.361296] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.361321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.361437] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.361470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.122 [2024-12-06 17:54:26.361624] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.122 [2024-12-06 17:54:26.361655] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.122 qpair failed and we were unable to recover it. 00:38:45.123 [2024-12-06 17:54:26.361799] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.123 [2024-12-06 17:54:26.361825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.123 qpair failed and we were unable to recover it. 00:38:45.123 [2024-12-06 17:54:26.361942] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.123 [2024-12-06 17:54:26.361969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.123 qpair failed and we were unable to recover it. 00:38:45.123 [2024-12-06 17:54:26.362054] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.123 [2024-12-06 17:54:26.362081] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.123 qpair failed and we were unable to recover it. 00:38:45.123 [2024-12-06 17:54:26.362166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.123 [2024-12-06 17:54:26.362193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.123 qpair failed and we were unable to recover it. 00:38:45.123 [2024-12-06 17:54:26.362318] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.123 [2024-12-06 17:54:26.362345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.123 qpair failed and we were unable to recover it. 00:38:45.123 Read completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Read completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Read completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Read completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Read completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Read completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Read completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Read completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Read completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Read completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Read completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Read completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Write completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Write completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Read completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Write completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Read completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Read completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Read completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Write completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Read completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Read completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Read completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Write completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Write completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Read completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Read completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Read completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Write completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Write completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Write completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 Read completed with error (sct=0, sc=8) 00:38:45.123 starting I/O failed 00:38:45.123 [2024-12-06 17:54:26.362640] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:38:45.123 [2024-12-06 17:54:26.362755] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.123 [2024-12-06 17:54:26.362794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.123 qpair failed and we were unable to recover it. 00:38:45.123 [2024-12-06 17:54:26.362912] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.123 [2024-12-06 17:54:26.362938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.123 qpair failed and we were unable to recover it. 00:38:45.123 [2024-12-06 17:54:26.363049] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.123 [2024-12-06 17:54:26.363074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.123 qpair failed and we were unable to recover it. 00:38:45.123 [2024-12-06 17:54:26.363183] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.123 [2024-12-06 17:54:26.363207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.123 qpair failed and we were unable to recover it. 00:38:45.123 [2024-12-06 17:54:26.363325] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.123 [2024-12-06 17:54:26.363352] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.123 qpair failed and we were unable to recover it. 00:38:45.123 [2024-12-06 17:54:26.363483] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.123 [2024-12-06 17:54:26.363510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.123 qpair failed and we were unable to recover it. 00:38:45.123 [2024-12-06 17:54:26.363628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.123 [2024-12-06 17:54:26.363655] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.123 qpair failed and we were unable to recover it. 00:38:45.123 [2024-12-06 17:54:26.363780] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.123 [2024-12-06 17:54:26.363806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.123 qpair failed and we were unable to recover it. 00:38:45.123 [2024-12-06 17:54:26.363913] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.123 [2024-12-06 17:54:26.363939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.123 qpair failed and we were unable to recover it. 00:38:45.123 [2024-12-06 17:54:26.364015] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.123 [2024-12-06 17:54:26.364042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.124 qpair failed and we were unable to recover it. 00:38:45.124 [2024-12-06 17:54:26.364244] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.124 [2024-12-06 17:54:26.364308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.124 qpair failed and we were unable to recover it. 00:38:45.124 [2024-12-06 17:54:26.364430] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.124 [2024-12-06 17:54:26.364456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.124 qpair failed and we were unable to recover it. 00:38:45.124 [2024-12-06 17:54:26.364573] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.124 [2024-12-06 17:54:26.364599] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.124 qpair failed and we were unable to recover it. 00:38:45.124 [2024-12-06 17:54:26.364681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.124 [2024-12-06 17:54:26.364708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.124 qpair failed and we were unable to recover it. 00:38:45.124 [2024-12-06 17:54:26.364803] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.124 [2024-12-06 17:54:26.364831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.124 qpair failed and we were unable to recover it. 00:38:45.124 [2024-12-06 17:54:26.364917] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.124 [2024-12-06 17:54:26.364944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.124 qpair failed and we were unable to recover it. 00:38:45.124 [2024-12-06 17:54:26.365049] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.124 [2024-12-06 17:54:26.365108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.124 qpair failed and we were unable to recover it. 00:38:45.124 [2024-12-06 17:54:26.365308] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.124 [2024-12-06 17:54:26.365358] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.124 qpair failed and we were unable to recover it. 00:38:45.124 [2024-12-06 17:54:26.365460] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.124 [2024-12-06 17:54:26.365489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.124 qpair failed and we were unable to recover it. 00:38:45.124 [2024-12-06 17:54:26.365625] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.124 [2024-12-06 17:54:26.365651] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.124 qpair failed and we were unable to recover it. 00:38:45.124 [2024-12-06 17:54:26.365759] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.124 [2024-12-06 17:54:26.365787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.124 qpair failed and we were unable to recover it. 00:38:45.124 [2024-12-06 17:54:26.365877] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.124 [2024-12-06 17:54:26.365903] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.124 qpair failed and we were unable to recover it. 00:38:45.124 [2024-12-06 17:54:26.365982] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.124 [2024-12-06 17:54:26.366009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.124 qpair failed and we were unable to recover it. 00:38:45.124 [2024-12-06 17:54:26.366130] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.124 [2024-12-06 17:54:26.366156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.124 qpair failed and we were unable to recover it. 00:38:45.124 [2024-12-06 17:54:26.366240] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.124 [2024-12-06 17:54:26.366266] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.124 qpair failed and we were unable to recover it. 00:38:45.124 [2024-12-06 17:54:26.366372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.124 [2024-12-06 17:54:26.366398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.124 qpair failed and we were unable to recover it. 00:38:45.124 [2024-12-06 17:54:26.366477] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.124 [2024-12-06 17:54:26.366502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.124 qpair failed and we were unable to recover it. 00:38:45.124 [2024-12-06 17:54:26.366615] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.124 [2024-12-06 17:54:26.366641] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.124 qpair failed and we were unable to recover it. 00:38:45.124 [2024-12-06 17:54:26.366779] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.124 [2024-12-06 17:54:26.366818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.124 qpair failed and we were unable to recover it. 00:38:45.124 [2024-12-06 17:54:26.366914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.124 [2024-12-06 17:54:26.366941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.124 qpair failed and we were unable to recover it. 00:38:45.124 [2024-12-06 17:54:26.367048] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.124 [2024-12-06 17:54:26.367074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.124 qpair failed and we were unable to recover it. 00:38:45.124 [2024-12-06 17:54:26.367186] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.124 [2024-12-06 17:54:26.367211] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.124 qpair failed and we were unable to recover it. 00:38:45.124 [2024-12-06 17:54:26.367301] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.124 [2024-12-06 17:54:26.367333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.124 qpair failed and we were unable to recover it. 00:38:45.124 [2024-12-06 17:54:26.367427] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.124 [2024-12-06 17:54:26.367474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.124 qpair failed and we were unable to recover it. 00:38:45.124 [2024-12-06 17:54:26.367569] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.124 [2024-12-06 17:54:26.367597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.124 qpair failed and we were unable to recover it. 00:38:45.124 [2024-12-06 17:54:26.367688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.124 [2024-12-06 17:54:26.367715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.124 qpair failed and we were unable to recover it. 00:38:45.124 [2024-12-06 17:54:26.367807] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.124 [2024-12-06 17:54:26.367833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.124 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.368009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.368060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.368194] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.368254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.368363] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.368390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.368540] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.368565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.368643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.368676] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.368760] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.368786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.368910] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.368939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.369090] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.369142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.369303] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.369354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.369523] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.369572] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.369724] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.369751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.369863] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.369889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.370003] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.370030] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.370242] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.370291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.370456] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.370483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.370626] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.370652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.370821] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.370859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.370952] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.370979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.371083] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.371139] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.371258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.371310] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.371420] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.371445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.371529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.371553] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.371691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.371723] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.371836] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.371861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.371939] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.371963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.372054] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.372080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.372156] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.372181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.372269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.372295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.372384] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.372423] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.372540] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.125 [2024-12-06 17:54:26.372567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.125 qpair failed and we were unable to recover it. 00:38:45.125 [2024-12-06 17:54:26.372686] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.372725] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.372870] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.372898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.373015] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.373042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.373158] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.373184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.373351] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.373406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.373484] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.373509] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.373633] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.373661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.373762] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.373787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.373875] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.373901] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.374046] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.374072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.374268] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.374324] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.374436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.374463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.374597] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.374636] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.374736] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.374763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.374872] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.374898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.375002] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.375058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.375264] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.375293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.375446] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.375475] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.375603] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.375628] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.375740] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.375779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.375874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.375901] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.376043] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.376094] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.376205] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.376261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.376454] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.376507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.376621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.376646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.376736] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.376762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.376844] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.376869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.376954] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.376979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.377094] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.377119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.126 [2024-12-06 17:54:26.377241] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.126 [2024-12-06 17:54:26.377280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.126 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.377371] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.377400] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.377517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.377545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.377662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.377694] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.377806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.377832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.377981] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.378042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.378176] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.378226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.378354] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.378387] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.378526] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.378554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.378674] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.378702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.378793] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.378818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.378926] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.378970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.379065] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.379095] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.379241] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.379285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.379391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.379417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.379502] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.379529] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.379646] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.379683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.379804] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.379833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.379923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.379949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.380044] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.380070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.380204] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.380230] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.380424] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.380470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.380627] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.380654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.380756] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.380782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.380893] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.380918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.381030] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.381055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.381167] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.381193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.381376] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.381401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.381538] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.381563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.381687] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.381714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.381805] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.381838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.127 [2024-12-06 17:54:26.381945] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.127 [2024-12-06 17:54:26.381971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.127 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.382096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.382141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.382225] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.382250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.382393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.382419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.382525] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.382551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.382643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.382677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.382773] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.382812] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.382928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.382966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.383061] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.383087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.383195] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.383246] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.383336] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.383361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.383451] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.383476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.383588] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.383616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.383717] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.383747] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.383885] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.383931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.384065] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.384109] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.384241] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.384286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.384421] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.384488] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.384606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.384633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.384753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.384779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.384862] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.384888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.384976] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.385001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.385082] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.385108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.385193] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.385218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.385330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.385356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.385441] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.385470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.385571] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.385601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.385719] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.385746] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.385836] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.385861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.385944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.128 [2024-12-06 17:54:26.385970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.128 qpair failed and we were unable to recover it. 00:38:45.128 [2024-12-06 17:54:26.386052] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.129 [2024-12-06 17:54:26.386078] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.129 qpair failed and we were unable to recover it. 00:38:45.129 [2024-12-06 17:54:26.386221] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.129 [2024-12-06 17:54:26.386274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.129 qpair failed and we were unable to recover it. 00:38:45.129 [2024-12-06 17:54:26.386463] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.129 [2024-12-06 17:54:26.386515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.129 qpair failed and we were unable to recover it. 00:38:45.129 [2024-12-06 17:54:26.386651] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.129 [2024-12-06 17:54:26.386698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.129 qpair failed and we were unable to recover it. 00:38:45.129 [2024-12-06 17:54:26.386824] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.129 [2024-12-06 17:54:26.386852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.129 qpair failed and we were unable to recover it. 00:38:45.129 [2024-12-06 17:54:26.386940] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.129 [2024-12-06 17:54:26.386969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.129 qpair failed and we were unable to recover it. 00:38:45.129 [2024-12-06 17:54:26.387073] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.129 [2024-12-06 17:54:26.387104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.129 qpair failed and we were unable to recover it. 00:38:45.129 [2024-12-06 17:54:26.387284] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.129 [2024-12-06 17:54:26.387328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.129 qpair failed and we were unable to recover it. 00:38:45.129 [2024-12-06 17:54:26.387447] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.129 [2024-12-06 17:54:26.387473] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.129 qpair failed and we were unable to recover it. 00:38:45.129 [2024-12-06 17:54:26.387619] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.129 [2024-12-06 17:54:26.387645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.129 qpair failed and we were unable to recover it. 00:38:45.129 [2024-12-06 17:54:26.387752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.129 [2024-12-06 17:54:26.387779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.129 qpair failed and we were unable to recover it. 00:38:45.129 [2024-12-06 17:54:26.387870] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.129 [2024-12-06 17:54:26.387896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.129 qpair failed and we were unable to recover it. 00:38:45.129 [2024-12-06 17:54:26.388004] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.129 [2024-12-06 17:54:26.388029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.129 qpair failed and we were unable to recover it. 00:38:45.129 [2024-12-06 17:54:26.388142] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.129 [2024-12-06 17:54:26.388168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.129 qpair failed and we were unable to recover it. 00:38:45.129 [2024-12-06 17:54:26.388251] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.129 [2024-12-06 17:54:26.388277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.129 qpair failed and we were unable to recover it. 00:38:45.129 [2024-12-06 17:54:26.388422] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.129 [2024-12-06 17:54:26.388469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.129 qpair failed and we were unable to recover it. 00:38:45.129 [2024-12-06 17:54:26.388630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.129 [2024-12-06 17:54:26.388656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.129 qpair failed and we were unable to recover it. 00:38:45.129 [2024-12-06 17:54:26.388771] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.129 [2024-12-06 17:54:26.388797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.129 qpair failed and we were unable to recover it. 00:38:45.129 [2024-12-06 17:54:26.388886] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.129 [2024-12-06 17:54:26.388912] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.129 qpair failed and we were unable to recover it. 00:38:45.129 [2024-12-06 17:54:26.388998] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.129 [2024-12-06 17:54:26.389024] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.129 qpair failed and we were unable to recover it. 00:38:45.129 [2024-12-06 17:54:26.389166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.129 [2024-12-06 17:54:26.389192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.129 qpair failed and we were unable to recover it. 00:38:45.129 [2024-12-06 17:54:26.389306] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.129 [2024-12-06 17:54:26.389360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.129 qpair failed and we were unable to recover it. 00:38:45.129 [2024-12-06 17:54:26.389481] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.129 [2024-12-06 17:54:26.389510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.129 qpair failed and we were unable to recover it. 00:38:45.129 [2024-12-06 17:54:26.389676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.129 [2024-12-06 17:54:26.389716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.129 qpair failed and we were unable to recover it. 00:38:45.129 [2024-12-06 17:54:26.389840] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.129 [2024-12-06 17:54:26.389866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.129 qpair failed and we were unable to recover it. 00:38:45.129 [2024-12-06 17:54:26.389955] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.389982] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.390091] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.390117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.390312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.390359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.390490] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.390555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.390700] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.390727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.390845] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.390871] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.390987] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.391012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.391096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.391122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.391195] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.391220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.391339] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.391411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.391512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.391538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.391654] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.391742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.391870] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.391896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.392035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.392084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.392177] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.392203] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.392295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.392322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.392475] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.392515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.392634] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.392662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.392766] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.392793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.392911] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.392938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.393051] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.393078] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.393203] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.393229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.393354] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.393381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.393484] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.393510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.393620] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.393647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.393771] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.393797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.393878] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.393904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.393982] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.394009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.394117] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.394179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.130 [2024-12-06 17:54:26.394372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.130 [2024-12-06 17:54:26.394417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.130 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.394575] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.394604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.394745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.394773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.394909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.394955] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.395081] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.395125] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.395271] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.395322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.395461] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.395511] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.395623] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.395650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.395751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.395777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.395860] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.395890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.396028] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.396058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.396245] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.396288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.396420] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.396471] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.396590] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.396617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.396714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.396743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.396868] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.396894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.396977] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.397004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.397172] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.397218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.397358] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.397405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.397548] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.397575] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.397723] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.397762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.397900] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.397938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.398040] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.398084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.398255] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.398307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.398413] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.398438] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.398551] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.398579] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.398675] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.398702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.398782] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.398809] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.398888] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.398914] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.399034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.131 [2024-12-06 17:54:26.399059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.131 qpair failed and we were unable to recover it. 00:38:45.131 [2024-12-06 17:54:26.399212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.399255] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.399421] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.399465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.399621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.399647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.399740] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.399766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.399853] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.399880] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.400007] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.400035] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.400188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.400245] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.400448] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.400503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.400594] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.400620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.400763] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.400789] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.400868] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.400894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.401072] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.401121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.401262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.401307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.401458] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.401510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.401598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.401625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.401719] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.401746] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.401863] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.401890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.402030] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.402076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.402206] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.402237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.402425] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.402491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.402640] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.402673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.402781] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.402807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.402931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.402977] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.403113] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.403159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.403303] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.403359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.403506] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.403532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.403639] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.403670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.403758] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.403785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.403880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.132 [2024-12-06 17:54:26.403906] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.132 qpair failed and we were unable to recover it. 00:38:45.132 [2024-12-06 17:54:26.403989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.404015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.404161] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.404206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.404353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.404400] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.404521] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.404547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.404629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.404655] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.404769] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.404802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.404988] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.405033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.405136] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.405167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.405262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.405288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.405378] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.405417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.405513] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.405540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.405649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.405687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.405801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.405827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.405963] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.405989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.406102] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.406128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.406300] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.406343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.406489] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.406539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.406702] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.406744] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.406842] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.406868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.406983] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.407009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.407125] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.407156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.407438] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.407481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.407615] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.407659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.407781] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.407808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.407924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.407950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.408089] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.408115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.408258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.408309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.408460] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.408517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.408635] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.408661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.408806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.408833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.133 qpair failed and we were unable to recover it. 00:38:45.133 [2024-12-06 17:54:26.408950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.133 [2024-12-06 17:54:26.408976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.409124] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.409151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.409269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.409295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.409384] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.409410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.409536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.409574] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.409674] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.409702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.409786] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.409812] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.409901] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.409927] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.410013] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.410041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.410156] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.410182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.410272] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.410299] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.410384] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.410409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.410518] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.410544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.410687] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.410729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.410820] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.410847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.410989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.411015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.411129] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.411172] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.411342] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.411372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.411485] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.411511] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.411629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.411655] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.411751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.411777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.411861] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.411887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.412001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.412027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.412182] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.412235] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.412341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.412366] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.412536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.412565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.412696] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.412742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.412885] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.412911] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.413052] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.413078] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.413225] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.413250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.134 qpair failed and we were unable to recover it. 00:38:45.134 [2024-12-06 17:54:26.413379] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.134 [2024-12-06 17:54:26.413422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.135 qpair failed and we were unable to recover it. 00:38:45.135 [2024-12-06 17:54:26.413585] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.135 [2024-12-06 17:54:26.413625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.135 qpair failed and we were unable to recover it. 00:38:45.135 [2024-12-06 17:54:26.413726] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.135 [2024-12-06 17:54:26.413755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.135 qpair failed and we were unable to recover it. 00:38:45.135 [2024-12-06 17:54:26.413847] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.135 [2024-12-06 17:54:26.413875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.135 qpair failed and we were unable to recover it. 00:38:45.135 [2024-12-06 17:54:26.413993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.135 [2024-12-06 17:54:26.414020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.135 qpair failed and we were unable to recover it. 00:38:45.135 [2024-12-06 17:54:26.414103] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.135 [2024-12-06 17:54:26.414130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.135 qpair failed and we were unable to recover it. 00:38:45.135 [2024-12-06 17:54:26.414323] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.135 [2024-12-06 17:54:26.414375] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.135 qpair failed and we were unable to recover it. 00:38:45.135 [2024-12-06 17:54:26.414489] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.135 [2024-12-06 17:54:26.414516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.135 qpair failed and we were unable to recover it. 00:38:45.135 [2024-12-06 17:54:26.414604] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.135 [2024-12-06 17:54:26.414632] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.135 qpair failed and we were unable to recover it. 00:38:45.135 [2024-12-06 17:54:26.414770] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.135 [2024-12-06 17:54:26.414808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.135 qpair failed and we were unable to recover it. 00:38:45.135 [2024-12-06 17:54:26.414894] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.135 [2024-12-06 17:54:26.414920] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.135 qpair failed and we were unable to recover it. 00:38:45.135 [2024-12-06 17:54:26.415056] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.135 [2024-12-06 17:54:26.415101] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.135 qpair failed and we were unable to recover it. 00:38:45.135 [2024-12-06 17:54:26.415250] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.135 [2024-12-06 17:54:26.415301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.135 qpair failed and we were unable to recover it. 00:38:45.135 [2024-12-06 17:54:26.415490] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.135 [2024-12-06 17:54:26.415546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.135 qpair failed and we were unable to recover it. 00:38:45.135 [2024-12-06 17:54:26.415673] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.135 [2024-12-06 17:54:26.415734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.135 qpair failed and we were unable to recover it. 00:38:45.135 [2024-12-06 17:54:26.415843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.135 [2024-12-06 17:54:26.415869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.135 qpair failed and we were unable to recover it. 00:38:45.135 [2024-12-06 17:54:26.415955] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.135 [2024-12-06 17:54:26.415981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.135 qpair failed and we were unable to recover it. 00:38:45.135 [2024-12-06 17:54:26.416084] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.135 [2024-12-06 17:54:26.416126] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.135 qpair failed and we were unable to recover it. 00:38:45.135 [2024-12-06 17:54:26.416295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.135 [2024-12-06 17:54:26.416338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.135 qpair failed and we were unable to recover it. 00:38:45.135 [2024-12-06 17:54:26.416456] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.135 [2024-12-06 17:54:26.416498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.135 qpair failed and we were unable to recover it. 00:38:45.135 [2024-12-06 17:54:26.416649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.135 [2024-12-06 17:54:26.416680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.135 qpair failed and we were unable to recover it. 00:38:45.135 [2024-12-06 17:54:26.416766] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.135 [2024-12-06 17:54:26.416793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.135 qpair failed and we were unable to recover it. 00:38:45.135 [2024-12-06 17:54:26.416906] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.135 [2024-12-06 17:54:26.416934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.135 qpair failed and we were unable to recover it. 00:38:45.135 [2024-12-06 17:54:26.417023] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.135 [2024-12-06 17:54:26.417049] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.135 qpair failed and we were unable to recover it. 00:38:45.135 [2024-12-06 17:54:26.417157] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.417189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.417276] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.417302] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.417445] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.417472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.417583] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.417610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.417729] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.417756] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.417837] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.417863] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.417961] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.417989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.418103] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.418128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.418215] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.418240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.418345] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.418371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.418480] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.418505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.418608] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.418647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.418747] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.418774] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.418904] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.418943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.419071] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.419099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.419285] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.419329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.419486] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.419513] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.419633] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.419659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.419778] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.419804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.419916] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.419942] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.420032] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.420058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.420148] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.420176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.420296] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.420356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.420520] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.420559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.420680] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.420708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.420796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.420822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.420913] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.420939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.421091] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.421123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.421354] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.136 [2024-12-06 17:54:26.421396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.136 qpair failed and we were unable to recover it. 00:38:45.136 [2024-12-06 17:54:26.421548] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.421574] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.137 [2024-12-06 17:54:26.421717] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.421744] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.137 [2024-12-06 17:54:26.421840] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.421865] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.137 [2024-12-06 17:54:26.421957] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.421985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.137 [2024-12-06 17:54:26.422080] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.422111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.137 [2024-12-06 17:54:26.422244] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.422296] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.137 [2024-12-06 17:54:26.422436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.422463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.137 [2024-12-06 17:54:26.422569] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.422596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.137 [2024-12-06 17:54:26.422739] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.422766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.137 [2024-12-06 17:54:26.422884] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.422911] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.137 [2024-12-06 17:54:26.423002] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.423028] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.137 [2024-12-06 17:54:26.423141] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.423173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.137 [2024-12-06 17:54:26.423363] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.423407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.137 [2024-12-06 17:54:26.423590] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.423652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.137 [2024-12-06 17:54:26.423816] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.423843] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.137 [2024-12-06 17:54:26.424003] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.424045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.137 [2024-12-06 17:54:26.424252] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.424304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.137 [2024-12-06 17:54:26.424410] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.424436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.137 [2024-12-06 17:54:26.424577] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.424604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.137 [2024-12-06 17:54:26.424745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.424772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.137 [2024-12-06 17:54:26.424859] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.424886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.137 [2024-12-06 17:54:26.425000] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.425027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.137 [2024-12-06 17:54:26.425169] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.425220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.137 [2024-12-06 17:54:26.425365] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.425413] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.137 [2024-12-06 17:54:26.425547] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.425574] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.137 [2024-12-06 17:54:26.425719] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.425747] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.137 [2024-12-06 17:54:26.425832] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.425858] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.137 [2024-12-06 17:54:26.425952] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.425979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.137 [2024-12-06 17:54:26.426093] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.137 [2024-12-06 17:54:26.426119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.137 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.426226] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.426252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.138 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.426342] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.426369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.138 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.426468] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.426507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.138 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.426623] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.426651] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.138 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.426791] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.426820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.138 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.426911] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.426937] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.138 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.427071] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.427096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.138 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.427205] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.427231] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.138 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.427336] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.427367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.138 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.427488] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.427532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.138 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.427675] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.427702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.138 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.427818] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.427844] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.138 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.427929] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.427954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.138 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.428069] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.428094] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.138 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.428184] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.428212] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.138 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.428301] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.428327] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.138 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.428456] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.428495] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.138 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.428582] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.428627] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.138 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.428753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.428779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.138 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.428858] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.428884] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.138 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.429002] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.429030] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.138 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.429141] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.429181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.138 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.429279] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.429312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.138 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.429452] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.429478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.138 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.429557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.429584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.138 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.429725] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.429752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.138 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.429864] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.429890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.138 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.430000] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.430027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.138 qpair failed and we were unable to recover it. 00:38:45.138 [2024-12-06 17:54:26.430114] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.138 [2024-12-06 17:54:26.430141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.430254] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.139 [2024-12-06 17:54:26.430280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.430370] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.139 [2024-12-06 17:54:26.430397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.430529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.139 [2024-12-06 17:54:26.430567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.430671] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.139 [2024-12-06 17:54:26.430700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.430793] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.139 [2024-12-06 17:54:26.430820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.430913] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.139 [2024-12-06 17:54:26.430939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.431021] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.139 [2024-12-06 17:54:26.431047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.431133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.139 [2024-12-06 17:54:26.431159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.431273] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.139 [2024-12-06 17:54:26.431331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.431516] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.139 [2024-12-06 17:54:26.431574] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.431710] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.139 [2024-12-06 17:54:26.431749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.431847] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.139 [2024-12-06 17:54:26.431874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.432013] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.139 [2024-12-06 17:54:26.432056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.432184] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.139 [2024-12-06 17:54:26.432228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.432363] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.139 [2024-12-06 17:54:26.432408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.432494] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.139 [2024-12-06 17:54:26.432520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.432645] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.139 [2024-12-06 17:54:26.432691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.432813] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.139 [2024-12-06 17:54:26.432840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.432923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.139 [2024-12-06 17:54:26.432948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.433056] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.139 [2024-12-06 17:54:26.433082] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.433195] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.139 [2024-12-06 17:54:26.433226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.433310] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.139 [2024-12-06 17:54:26.433335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.433452] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.139 [2024-12-06 17:54:26.433476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.433566] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.139 [2024-12-06 17:54:26.433596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.433715] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.139 [2024-12-06 17:54:26.433743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.433828] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.139 [2024-12-06 17:54:26.433854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.433926] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.139 [2024-12-06 17:54:26.433952] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.434030] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.139 [2024-12-06 17:54:26.434055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.139 qpair failed and we were unable to recover it. 00:38:45.139 [2024-12-06 17:54:26.434146] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.434172] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.140 [2024-12-06 17:54:26.434274] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.434304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.140 [2024-12-06 17:54:26.434400] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.434424] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.140 [2024-12-06 17:54:26.434503] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.434528] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.140 [2024-12-06 17:54:26.434618] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.434644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.140 [2024-12-06 17:54:26.434740] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.434768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.140 [2024-12-06 17:54:26.434860] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.434889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.140 [2024-12-06 17:54:26.434972] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.434998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.140 [2024-12-06 17:54:26.435111] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.435165] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.140 [2024-12-06 17:54:26.435340] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.435387] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.140 [2024-12-06 17:54:26.435502] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.435529] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.140 [2024-12-06 17:54:26.435625] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.435652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.140 [2024-12-06 17:54:26.435747] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.435772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.140 [2024-12-06 17:54:26.435860] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.435886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.140 [2024-12-06 17:54:26.435964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.435989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.140 [2024-12-06 17:54:26.436099] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.436125] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.140 [2024-12-06 17:54:26.436216] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.436244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.140 [2024-12-06 17:54:26.436355] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.436381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.140 [2024-12-06 17:54:26.436466] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.436492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.140 [2024-12-06 17:54:26.436580] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.436606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.140 [2024-12-06 17:54:26.436686] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.436711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.140 [2024-12-06 17:54:26.436800] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.436829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.140 [2024-12-06 17:54:26.436905] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.436931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.140 [2024-12-06 17:54:26.437040] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.437065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.140 [2024-12-06 17:54:26.437145] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.437170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.140 [2024-12-06 17:54:26.437287] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.437314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.140 [2024-12-06 17:54:26.437402] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.437427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.140 [2024-12-06 17:54:26.437533] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.437559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.140 [2024-12-06 17:54:26.437674] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.140 [2024-12-06 17:54:26.437700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.140 qpair failed and we were unable to recover it. 00:38:45.141 [2024-12-06 17:54:26.437819] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.141 [2024-12-06 17:54:26.437844] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.141 qpair failed and we were unable to recover it. 00:38:45.141 [2024-12-06 17:54:26.437916] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.141 [2024-12-06 17:54:26.437942] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.141 qpair failed and we were unable to recover it. 00:38:45.141 [2024-12-06 17:54:26.438075] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.141 [2024-12-06 17:54:26.438105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.141 qpair failed and we were unable to recover it. 00:38:45.141 [2024-12-06 17:54:26.438231] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.141 [2024-12-06 17:54:26.438260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.141 qpair failed and we were unable to recover it. 00:38:45.141 [2024-12-06 17:54:26.438408] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.141 [2024-12-06 17:54:26.438452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.141 qpair failed and we were unable to recover it. 00:38:45.141 [2024-12-06 17:54:26.438583] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.141 [2024-12-06 17:54:26.438609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.141 qpair failed and we were unable to recover it. 00:38:45.141 [2024-12-06 17:54:26.438694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.141 [2024-12-06 17:54:26.438720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.141 qpair failed and we were unable to recover it. 00:38:45.141 [2024-12-06 17:54:26.438808] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.141 [2024-12-06 17:54:26.438834] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.141 qpair failed and we were unable to recover it. 00:38:45.141 [2024-12-06 17:54:26.438971] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.141 [2024-12-06 17:54:26.439012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.141 qpair failed and we were unable to recover it. 00:38:45.141 [2024-12-06 17:54:26.439186] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.141 [2024-12-06 17:54:26.439226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.141 qpair failed and we were unable to recover it. 00:38:45.141 [2024-12-06 17:54:26.439337] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.141 [2024-12-06 17:54:26.439367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.141 qpair failed and we were unable to recover it. 00:38:45.141 [2024-12-06 17:54:26.439486] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.141 [2024-12-06 17:54:26.439515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.141 qpair failed and we were unable to recover it. 00:38:45.141 [2024-12-06 17:54:26.439630] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7835f0 is same with the state(6) to be set 00:38:45.141 [2024-12-06 17:54:26.439782] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.141 [2024-12-06 17:54:26.439822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.141 qpair failed and we were unable to recover it. 00:38:45.141 [2024-12-06 17:54:26.439954] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.141 [2024-12-06 17:54:26.439993] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.141 qpair failed and we were unable to recover it. 00:38:45.141 [2024-12-06 17:54:26.440184] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.141 [2024-12-06 17:54:26.440228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.141 qpair failed and we were unable to recover it. 00:38:45.141 [2024-12-06 17:54:26.440435] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.141 [2024-12-06 17:54:26.440476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.141 qpair failed and we were unable to recover it. 00:38:45.141 [2024-12-06 17:54:26.440627] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.141 [2024-12-06 17:54:26.440658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.141 qpair failed and we were unable to recover it. 00:38:45.141 [2024-12-06 17:54:26.440753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.141 [2024-12-06 17:54:26.440779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.141 qpair failed and we were unable to recover it. 00:38:45.141 [2024-12-06 17:54:26.440890] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.141 [2024-12-06 17:54:26.440915] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.141 qpair failed and we were unable to recover it. 00:38:45.141 [2024-12-06 17:54:26.441009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.141 [2024-12-06 17:54:26.441036] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.141 qpair failed and we were unable to recover it. 00:38:45.141 [2024-12-06 17:54:26.441130] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.141 [2024-12-06 17:54:26.441158] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.141 qpair failed and we were unable to recover it. 00:38:45.141 [2024-12-06 17:54:26.441249] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.141 [2024-12-06 17:54:26.441296] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.141 qpair failed and we were unable to recover it. 00:38:45.141 [2024-12-06 17:54:26.441445] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.141 [2024-12-06 17:54:26.441494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.141 qpair failed and we were unable to recover it. 00:38:45.141 [2024-12-06 17:54:26.441649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.141 [2024-12-06 17:54:26.441701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.141 qpair failed and we were unable to recover it. 00:38:45.142 [2024-12-06 17:54:26.441789] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.142 [2024-12-06 17:54:26.441815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.142 qpair failed and we were unable to recover it. 00:38:45.142 [2024-12-06 17:54:26.441957] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.142 [2024-12-06 17:54:26.441983] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.142 qpair failed and we were unable to recover it. 00:38:45.142 [2024-12-06 17:54:26.442131] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.142 [2024-12-06 17:54:26.442186] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.142 qpair failed and we were unable to recover it. 00:38:45.142 [2024-12-06 17:54:26.442334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.142 [2024-12-06 17:54:26.442375] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.142 qpair failed and we were unable to recover it. 00:38:45.142 [2024-12-06 17:54:26.442516] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.142 [2024-12-06 17:54:26.442546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.142 qpair failed and we were unable to recover it. 00:38:45.142 [2024-12-06 17:54:26.442686] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.142 [2024-12-06 17:54:26.442734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.142 qpair failed and we were unable to recover it. 00:38:45.142 [2024-12-06 17:54:26.442852] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.142 [2024-12-06 17:54:26.442878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.142 qpair failed and we were unable to recover it. 00:38:45.142 [2024-12-06 17:54:26.442964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.142 [2024-12-06 17:54:26.442990] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.142 qpair failed and we were unable to recover it. 00:38:45.142 [2024-12-06 17:54:26.443110] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.142 [2024-12-06 17:54:26.443136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.142 qpair failed and we were unable to recover it. 00:38:45.142 [2024-12-06 17:54:26.443291] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.142 [2024-12-06 17:54:26.443331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.142 qpair failed and we were unable to recover it. 00:38:45.142 [2024-12-06 17:54:26.443460] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.142 [2024-12-06 17:54:26.443503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.142 qpair failed and we were unable to recover it. 00:38:45.142 [2024-12-06 17:54:26.443688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.142 [2024-12-06 17:54:26.443731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.142 qpair failed and we were unable to recover it. 00:38:45.142 [2024-12-06 17:54:26.443845] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.142 [2024-12-06 17:54:26.443871] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.142 qpair failed and we were unable to recover it. 00:38:45.142 [2024-12-06 17:54:26.444013] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.142 [2024-12-06 17:54:26.444039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.142 qpair failed and we were unable to recover it. 00:38:45.142 [2024-12-06 17:54:26.444122] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.142 [2024-12-06 17:54:26.444147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.142 qpair failed and we were unable to recover it. 00:38:45.142 [2024-12-06 17:54:26.444263] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.142 [2024-12-06 17:54:26.444288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.142 qpair failed and we were unable to recover it. 00:38:45.142 [2024-12-06 17:54:26.444394] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.142 [2024-12-06 17:54:26.444423] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.142 qpair failed and we were unable to recover it. 00:38:45.142 [2024-12-06 17:54:26.444553] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.142 [2024-12-06 17:54:26.444584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.142 qpair failed and we were unable to recover it. 00:38:45.142 [2024-12-06 17:54:26.444726] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.142 [2024-12-06 17:54:26.444753] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.142 qpair failed and we were unable to recover it. 00:38:45.142 [2024-12-06 17:54:26.444855] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.142 [2024-12-06 17:54:26.444894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.142 qpair failed and we were unable to recover it. 00:38:45.142 [2024-12-06 17:54:26.445045] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.142 [2024-12-06 17:54:26.445096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.142 qpair failed and we were unable to recover it. 00:38:45.142 [2024-12-06 17:54:26.445203] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.142 [2024-12-06 17:54:26.445255] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.142 qpair failed and we were unable to recover it. 00:38:45.142 [2024-12-06 17:54:26.445364] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.142 [2024-12-06 17:54:26.445414] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.142 qpair failed and we were unable to recover it. 00:38:45.142 [2024-12-06 17:54:26.445553] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.142 [2024-12-06 17:54:26.445580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.142 qpair failed and we were unable to recover it. 00:38:45.142 [2024-12-06 17:54:26.445694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.142 [2024-12-06 17:54:26.445721] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.142 qpair failed and we were unable to recover it. 00:38:45.142 [2024-12-06 17:54:26.445814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.142 [2024-12-06 17:54:26.445841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.142 qpair failed and we were unable to recover it. 00:38:45.142 [2024-12-06 17:54:26.445922] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.142 [2024-12-06 17:54:26.445949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.142 qpair failed and we were unable to recover it. 00:38:45.142 [2024-12-06 17:54:26.446030] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.142 [2024-12-06 17:54:26.446057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.142 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.446143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.446169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.446283] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.446310] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.446418] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.446445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.446534] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.446561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.446678] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.446714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.446830] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.446856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.446966] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.446992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.447107] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.447135] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.447244] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.447270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.447378] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.447403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.447494] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.447520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.447631] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.447657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.447753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.447778] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.447868] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.447895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.448008] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.448034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.448128] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.448154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.448261] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.448307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.448433] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.448460] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.448587] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.448613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.448713] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.448740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.448886] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.448912] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.449003] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.449029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.449139] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.449164] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.449240] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.449265] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.449352] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.449378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.449495] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.449522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.449657] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.449705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.449849] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.449888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.450004] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.450031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.450115] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.450142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.143 qpair failed and we were unable to recover it. 00:38:45.143 [2024-12-06 17:54:26.450312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.143 [2024-12-06 17:54:26.450353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.450496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.450528] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.450680] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.450719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.450849] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.450878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.451006] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.451031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.451128] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.451167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.451300] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.451349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.451432] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.451457] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.451546] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.451572] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.451687] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.451712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.451826] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.451851] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.451964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.451989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.452067] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.452092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.452204] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.452229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.452339] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.452364] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.452482] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.452507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.452588] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.452614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.452707] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.452735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.452815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.452841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.452922] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.452948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.453032] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.453056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.453135] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.453159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.453232] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.453257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.453346] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.453373] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.453461] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.453486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.453624] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.453650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.453743] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.453768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.453901] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.453949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.454067] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.454115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.454226] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.454276] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.454385] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.144 [2024-12-06 17:54:26.454411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.144 qpair failed and we were unable to recover it. 00:38:45.144 [2024-12-06 17:54:26.454491] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.454516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.454624] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.454650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.454769] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.454808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.454930] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.454957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.455073] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.455099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.455237] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.455264] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.455391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.455430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.455517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.455544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.455631] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.455658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.455757] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.455782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.455896] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.455922] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.456005] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.456030] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.456116] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.456144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.456237] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.456266] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.456382] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.456410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.456496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.456522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.456660] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.456692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.456808] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.456836] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.457004] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.457051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.457189] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.457233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.457412] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.457455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.457612] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.457638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.457762] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.457788] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.457869] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.457895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.457999] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.458031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.458191] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.458240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.458371] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.458420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.458566] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.458591] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.458676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.458701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.458825] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.458863] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.458992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.459042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.145 [2024-12-06 17:54:26.459234] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.145 [2024-12-06 17:54:26.459282] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.145 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.459412] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.459462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.459590] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.459623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.459740] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.459767] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.459852] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.459878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.459963] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.459989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.460109] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.460135] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.460300] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.460342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.460481] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.460508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.460622] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.460648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.460796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.460823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.460918] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.460965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.461144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.461199] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.461318] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.461371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.461511] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.461540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.461671] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.461710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.461796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.461823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.461951] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.461978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.462093] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.462142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.462282] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.462333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.462462] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.462488] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.462594] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.462621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.462736] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.462764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.462846] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.462871] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.462999] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.463054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.463157] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.463181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.463314] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.463340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.463509] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.463541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.463643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.463675] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.463796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.463822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.463908] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.463933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.146 [2024-12-06 17:54:26.464080] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.146 [2024-12-06 17:54:26.464131] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.146 qpair failed and we were unable to recover it. 00:38:45.147 [2024-12-06 17:54:26.464274] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.147 [2024-12-06 17:54:26.464325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.147 qpair failed and we were unable to recover it. 00:38:45.147 [2024-12-06 17:54:26.464440] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.147 [2024-12-06 17:54:26.464470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.147 qpair failed and we were unable to recover it. 00:38:45.147 [2024-12-06 17:54:26.464560] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.147 [2024-12-06 17:54:26.464600] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.147 qpair failed and we were unable to recover it. 00:38:45.147 [2024-12-06 17:54:26.464747] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.147 [2024-12-06 17:54:26.464776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.147 qpair failed and we were unable to recover it. 00:38:45.147 [2024-12-06 17:54:26.464889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.147 [2024-12-06 17:54:26.464916] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.147 qpair failed and we were unable to recover it. 00:38:45.147 [2024-12-06 17:54:26.465057] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.147 [2024-12-06 17:54:26.465084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.147 qpair failed and we were unable to recover it. 00:38:45.147 [2024-12-06 17:54:26.465166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.147 [2024-12-06 17:54:26.465194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.147 qpair failed and we were unable to recover it. 00:38:45.147 [2024-12-06 17:54:26.465307] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.147 [2024-12-06 17:54:26.465333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.147 qpair failed and we were unable to recover it. 00:38:45.147 [2024-12-06 17:54:26.465463] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.147 [2024-12-06 17:54:26.465502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.147 qpair failed and we were unable to recover it. 00:38:45.147 [2024-12-06 17:54:26.465610] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.147 [2024-12-06 17:54:26.465649] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.147 qpair failed and we were unable to recover it. 00:38:45.147 [2024-12-06 17:54:26.465758] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.147 [2024-12-06 17:54:26.465786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.147 qpair failed and we were unable to recover it. 00:38:45.147 [2024-12-06 17:54:26.465870] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.147 [2024-12-06 17:54:26.465895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.147 qpair failed and we were unable to recover it. 00:38:45.147 [2024-12-06 17:54:26.466011] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.147 [2024-12-06 17:54:26.466038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.147 qpair failed and we were unable to recover it. 00:38:45.147 [2024-12-06 17:54:26.466149] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.147 [2024-12-06 17:54:26.466173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.147 qpair failed and we were unable to recover it. 00:38:45.147 [2024-12-06 17:54:26.466316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.147 [2024-12-06 17:54:26.466341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.147 qpair failed and we were unable to recover it. 00:38:45.147 [2024-12-06 17:54:26.466436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.147 [2024-12-06 17:54:26.466466] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.147 qpair failed and we were unable to recover it. 00:38:45.147 [2024-12-06 17:54:26.466561] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.147 [2024-12-06 17:54:26.466590] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.147 qpair failed and we were unable to recover it. 00:38:45.147 [2024-12-06 17:54:26.466709] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.147 [2024-12-06 17:54:26.466738] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.147 qpair failed and we were unable to recover it. 00:38:45.147 [2024-12-06 17:54:26.466829] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.147 [2024-12-06 17:54:26.466855] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.147 qpair failed and we were unable to recover it. 00:38:45.147 [2024-12-06 17:54:26.467001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.147 [2024-12-06 17:54:26.467028] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.147 qpair failed and we were unable to recover it. 00:38:45.147 [2024-12-06 17:54:26.467117] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.147 [2024-12-06 17:54:26.467143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.147 qpair failed and we were unable to recover it. 00:38:45.147 [2024-12-06 17:54:26.467292] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.147 [2024-12-06 17:54:26.467345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.147 qpair failed and we were unable to recover it. 00:38:45.147 [2024-12-06 17:54:26.467454] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.147 [2024-12-06 17:54:26.467480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.147 qpair failed and we were unable to recover it. 00:38:45.147 [2024-12-06 17:54:26.467588] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.147 [2024-12-06 17:54:26.467613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.147 qpair failed and we were unable to recover it. 00:38:45.147 [2024-12-06 17:54:26.467731] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.147 [2024-12-06 17:54:26.467757] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.147 qpair failed and we were unable to recover it. 00:38:45.147 [2024-12-06 17:54:26.467849] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.147 [2024-12-06 17:54:26.467874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.147 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.467956] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.467985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.468102] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.468129] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.468311] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.468368] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.468478] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.468509] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.468676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.468704] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.468810] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.468836] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.468931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.468957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.469055] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.469096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.469255] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.469300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.469431] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.469456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.469543] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.469569] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.469651] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.469684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.469767] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.469793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.469887] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.469926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.470086] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.470125] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.470246] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.470273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.470392] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.470417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.470553] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.470578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.470692] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.470717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.470809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.470834] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.470946] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.470971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.471083] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.471107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.471280] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.471326] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.471457] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.471487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.471621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.471646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.471745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.471771] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.471857] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.471882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.472024] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.472068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.472230] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.472285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.472383] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.472418] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.472539] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.472567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.472662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.148 [2024-12-06 17:54:26.472696] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.148 qpair failed and we were unable to recover it. 00:38:45.148 [2024-12-06 17:54:26.472791] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.472815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.472903] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.472928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.473060] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.473088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.473253] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.473282] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.473384] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.473412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.473515] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.473539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.473679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.473705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.473793] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.473817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.473898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.473923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.474034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.474059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.474154] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.474183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.474374] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.474402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.474528] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.474567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.474661] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.474700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.474818] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.474846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.474962] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.474989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.475078] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.475105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.475194] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.475220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.475346] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.475373] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.475479] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.475505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.475658] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.475692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.475804] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.475830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.475945] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.475970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.476082] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.476107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.476230] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.476257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.476396] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.476423] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.476550] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.476588] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.476707] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.476734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.476832] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.476870] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.476965] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.476993] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.477132] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.477163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.477295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.477339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.149 [2024-12-06 17:54:26.477467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.149 [2024-12-06 17:54:26.477495] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.149 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.477615] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.477644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.477732] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.477759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.477855] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.477881] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.478048] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.478093] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.478228] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.478287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.478429] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.478477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.478559] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.478585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.478671] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.478698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.478809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.478836] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.478914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.478940] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.479049] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.479075] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.479160] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.479186] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.479264] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.479289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.479427] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.479452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.479534] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.479559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.479645] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.479679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.479772] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.479797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.479900] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.479924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.480069] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.480095] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.480182] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.480207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.480316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.480342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.480440] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.480479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.480631] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.480659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.480783] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.480810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.480901] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.480929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.481021] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.481048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.481191] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.481217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.481337] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.481390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.481506] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.481532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.481646] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.481678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.481843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.481888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.481965] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.481995] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.482078] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.482103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.150 qpair failed and we were unable to recover it. 00:38:45.150 [2024-12-06 17:54:26.482253] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.150 [2024-12-06 17:54:26.482304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.482411] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.482437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.482522] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.482547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.482657] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.482690] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.482809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.482836] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.483003] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.483049] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.483156] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.483186] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.483344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.483401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.483480] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.483506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.483592] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.483619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.483728] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.483767] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.483908] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.483947] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.484076] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.484103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.484193] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.484220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.484309] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.484335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.484475] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.484501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.484586] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.484612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.484741] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.484787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.484876] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.484901] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.484988] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.485013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.485124] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.485150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.485243] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.485273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.485359] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.485388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.485511] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.485537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.485654] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.485686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.485768] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.485800] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.485886] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.485913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.486043] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.486087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.486229] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.486281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.486365] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.486391] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.486537] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.486562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.486650] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.486681] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.486768] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.486794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.486938] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.486981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.487065] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.487090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.487262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.487310] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.487425] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.151 [2024-12-06 17:54:26.487451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.151 qpair failed and we were unable to recover it. 00:38:45.151 [2024-12-06 17:54:26.487526] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.487551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.487628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.487654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.487812] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.487838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.487917] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.487943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.488116] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.488141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.488248] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.488273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.488383] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.488408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.488548] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.488587] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.488684] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.488711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.488825] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.488852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.488951] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.488995] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.489196] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.489244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.489417] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.489463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.489599] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.489626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.489755] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.489781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.489871] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.489901] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.490024] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.490067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.490197] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.490227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.490356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.490382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.490464] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.490492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.490649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.490697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.490796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.490824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.490911] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.490938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.491055] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.491082] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.491199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.491226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.491373] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.491411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.491530] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.491575] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.491683] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.491710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.491794] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.491819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.491904] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.491930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.492013] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.152 [2024-12-06 17:54:26.492039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.152 qpair failed and we were unable to recover it. 00:38:45.152 [2024-12-06 17:54:26.492160] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.492207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.492342] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.492374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.492512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.492538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.492618] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.492645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.492794] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.492820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.492899] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.492924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.493099] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.493147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.493226] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.493252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.493388] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.493437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.493579] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.493605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.493722] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.493750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.493837] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.493874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.493991] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.494018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.494101] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.494128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.494243] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.494269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.494377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.494404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.494527] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.494554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.494698] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.494725] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.494829] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.494853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.494974] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.495000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.495099] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.495128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.495248] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.495277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.495369] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.495398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.495515] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.495560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.495683] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.495711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.495831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.495857] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.495986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.496030] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.496111] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.496137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.496269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.496314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.496399] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.496425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.496505] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.496532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.496638] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.496674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.496805] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.496831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.496953] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.496992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.497081] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.153 [2024-12-06 17:54:26.497109] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.153 qpair failed and we were unable to recover it. 00:38:45.153 [2024-12-06 17:54:26.497188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.497215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.497293] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.497319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.497469] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.497495] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.497584] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.497612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.497728] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.497753] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.497867] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.497892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.498006] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.498031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.498195] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.498227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.498396] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.498446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.498562] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.498588] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.498682] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.498710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.498822] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.498848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.498969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.498997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.499081] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.499108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.499212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.499251] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.499467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.499522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.499631] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.499662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.499755] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.499781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.499883] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.499937] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.500048] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.500099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.500218] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.500245] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.500341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.500369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.500454] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.500480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.500562] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.500588] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.500695] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.500722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.500836] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.500863] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.500955] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.500981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.501066] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.501093] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.501236] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.501262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.501378] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.501404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.501519] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.501545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.501627] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.501652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.501778] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.501805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.501945] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.501997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.502135] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.502186] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.502299] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.502326] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.502469] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.502497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.502620] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.502660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.502795] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.502823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.154 qpair failed and we were unable to recover it. 00:38:45.154 [2024-12-06 17:54:26.502901] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.154 [2024-12-06 17:54:26.502928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.503015] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.503042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.503218] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.503270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.503415] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.503467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.503553] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.503585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.503734] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.503761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.503866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.503896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.504049] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.504095] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.504200] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.504245] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.504367] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.504394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.504505] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.504531] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.504627] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.504653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.504800] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.504826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.504914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.504940] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.505017] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.505042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.505117] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.505143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.505246] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.505271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.505352] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.505379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.505470] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.505498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.505595] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.505634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.505744] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.505772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.505866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.505892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.505987] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.506012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.506094] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.506121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.506277] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.506305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.506420] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.506447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.506529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.506556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.506640] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.506673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.506784] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.506810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.506893] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.506918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.507008] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.507037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.507211] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.507258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.507440] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.507479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.507607] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.507634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.507759] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.507785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.507876] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.507901] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.507991] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.508018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.155 [2024-12-06 17:54:26.508209] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.155 [2024-12-06 17:54:26.508259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.155 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.508393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.508453] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.508595] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.508621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.508761] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.508804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.508904] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.508935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.509073] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.509119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.509254] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.509298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.509409] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.509436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.509575] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.509614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.509729] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.509757] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.509883] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.509921] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.510087] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.510149] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.510335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.510385] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.510496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.510521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.510633] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.510659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.510780] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.510805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.510887] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.510912] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.511045] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.511075] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.511262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.511291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.511417] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.511447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.511573] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.511598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.511701] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.511740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.511878] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.511917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.512016] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.512044] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.512131] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.512159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.512263] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.512293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.512423] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.512453] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.512551] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.512581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.512714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.512743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.512850] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.512889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.513012] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.513040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.513180] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.513229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.513371] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.513417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.513545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.513571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.513714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.513747] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.513859] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.513885] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.514023] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.514049] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.514135] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.514163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.514314] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.156 [2024-12-06 17:54:26.514340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.156 qpair failed and we were unable to recover it. 00:38:45.156 [2024-12-06 17:54:26.514467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.514493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.514578] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.514605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.514714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.514753] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.514860] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.514899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.515042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.515069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.515192] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.515220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.515353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.515392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.515543] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.515570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.515685] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.515713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.515804] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.515831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.515919] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.515945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.516055] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.516082] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.516159] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.516185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.516302] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.516330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.516415] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.516442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.516530] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.516559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.516646] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.516677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.516766] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.516792] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.516902] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.516928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.517098] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.517147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.517294] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.517335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.517460] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.517490] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.517648] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.517689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.517772] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.517799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.517884] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.517910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.517987] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.518013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.518168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.518217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.518352] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.518398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.518535] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.518561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.518637] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.518673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.518760] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.518787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.518914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.518958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.519141] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.157 [2024-12-06 17:54:26.519172] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.157 qpair failed and we were unable to recover it. 00:38:45.157 [2024-12-06 17:54:26.519431] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.519470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.519600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.519628] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.519723] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.519750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.519870] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.519896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.520040] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.520086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.520170] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.520196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.520363] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.520428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.520570] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.520601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.520725] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.520752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.520844] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.520870] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.520981] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.521007] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.521117] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.521144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.521230] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.521258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.521364] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.521393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.521498] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.521537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.521656] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.521689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.521781] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.521808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.521920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.521946] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.522061] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.522088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.522229] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.522255] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.522405] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.522433] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.522512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.522538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.522626] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.522652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.522753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.522778] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.522893] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.522918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.523086] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.523115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.523226] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.523278] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.523390] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.523437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.523601] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.523627] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.523730] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.523764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.523907] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.523934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.524043] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.524069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.524214] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.524240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.524331] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.524359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.524536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.524566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.524702] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.524730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.524841] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.524867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.524956] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.524982] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.525099] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.525125] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.525243] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.525270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.158 [2024-12-06 17:54:26.525407] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.158 [2024-12-06 17:54:26.525447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.158 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.525604] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.525647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.525768] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.525793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.525909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.525935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.526015] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.526041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.526145] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.526174] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.526293] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.526322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.526441] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.526472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.526592] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.526620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.526729] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.526758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.526871] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.526897] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.526998] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.527028] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.527171] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.527217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.527354] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.527406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.527498] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.527526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.527619] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.527646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.527773] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.527801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.527899] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.527926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.528035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.528061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.528209] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.528234] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.528346] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.528398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.528552] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.528595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.528740] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.528768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.528854] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.528880] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.528962] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.528988] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.529102] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.529128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.529235] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.529261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.529401] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.529459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.529589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.529619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.529734] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.529767] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.529863] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.529891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.529990] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.530015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.530126] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.530152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.530288] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.530336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.530426] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.530452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.530525] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.530552] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.530670] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.530697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.530811] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.530838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.530920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.530946] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.531030] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.159 [2024-12-06 17:54:26.531055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.159 qpair failed and we were unable to recover it. 00:38:45.159 [2024-12-06 17:54:26.531203] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.531241] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.531364] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.531414] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.531541] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.531566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.531697] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.531736] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.531856] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.531884] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.532024] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.532054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.532208] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.532247] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.532363] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.532416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.532563] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.532590] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.532676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.532704] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.532818] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.532844] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.532990] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.533015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.533125] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.533151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.533265] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.533292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.533395] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.533445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.533527] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.533553] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.533692] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.533723] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.533844] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.533870] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.534019] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.534068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.534217] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.534265] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.534407] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.534460] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.534545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.534571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.534698] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.534724] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.534806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.534832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.534925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.534953] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.535039] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.535066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.535154] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.535181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.535291] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.535317] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.535429] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.535455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.535567] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.535592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.535740] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.535768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.535874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.535913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.536063] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.536091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.536204] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.536229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.536392] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.536444] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.536567] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.536607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.536722] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.536751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.536895] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.536923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.537039] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.537065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.537194] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.537235] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.537484] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.537523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.537704] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.537731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.160 [2024-12-06 17:54:26.537868] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.160 [2024-12-06 17:54:26.537894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.160 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.538009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.538074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.538192] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.538238] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.538353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.538402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.538514] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.538540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.538629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.538654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.538780] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.538805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.538891] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.538917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.539032] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.539057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.539145] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.539170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.539313] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.539338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.539414] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.539440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.539574] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.539613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.539724] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.539752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.539866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.539892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.540006] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.540035] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.540136] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.540166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.540347] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.540386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.540534] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.540560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.540686] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.540716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.540834] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.540861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.540961] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.540992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.541100] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.541126] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.541245] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.541283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.541412] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.541439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.541525] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.541552] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.541646] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.541678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.541792] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.541818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.541899] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.541927] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.542007] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.542032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.542132] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.542176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.542345] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.542398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.542488] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.542514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.542652] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.542687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.542778] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.542824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.542985] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.543025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.543199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.543237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.543349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.543388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.161 [2024-12-06 17:54:26.543557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.161 [2024-12-06 17:54:26.543583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.161 qpair failed and we were unable to recover it. 00:38:45.162 [2024-12-06 17:54:26.543719] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.162 [2024-12-06 17:54:26.543758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.162 qpair failed and we were unable to recover it. 00:38:45.162 [2024-12-06 17:54:26.543877] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.162 [2024-12-06 17:54:26.543904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.162 qpair failed and we were unable to recover it. 00:38:45.162 [2024-12-06 17:54:26.544016] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.162 [2024-12-06 17:54:26.544046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.162 qpair failed and we were unable to recover it. 00:38:45.162 [2024-12-06 17:54:26.544160] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.162 [2024-12-06 17:54:26.544186] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.162 qpair failed and we were unable to recover it. 00:38:45.162 [2024-12-06 17:54:26.544278] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.162 [2024-12-06 17:54:26.544303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.162 qpair failed and we were unable to recover it. 00:38:45.162 [2024-12-06 17:54:26.544384] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.162 [2024-12-06 17:54:26.544409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.162 qpair failed and we were unable to recover it. 00:38:45.162 [2024-12-06 17:54:26.544526] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.162 [2024-12-06 17:54:26.544554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.162 qpair failed and we were unable to recover it. 00:38:45.162 [2024-12-06 17:54:26.544648] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.162 [2024-12-06 17:54:26.544683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.162 qpair failed and we were unable to recover it. 00:38:45.162 [2024-12-06 17:54:26.544808] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.162 [2024-12-06 17:54:26.544847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.162 qpair failed and we were unable to recover it. 00:38:45.162 [2024-12-06 17:54:26.544944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.162 [2024-12-06 17:54:26.544973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.162 qpair failed and we were unable to recover it. 00:38:45.162 [2024-12-06 17:54:26.545087] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.162 [2024-12-06 17:54:26.545114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.162 qpair failed and we were unable to recover it. 00:38:45.162 [2024-12-06 17:54:26.545256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.162 [2024-12-06 17:54:26.545283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.162 qpair failed and we were unable to recover it. 00:38:45.162 [2024-12-06 17:54:26.545371] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.162 [2024-12-06 17:54:26.545398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.162 qpair failed and we were unable to recover it. 00:38:45.162 [2024-12-06 17:54:26.545476] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.162 [2024-12-06 17:54:26.545503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.162 qpair failed and we were unable to recover it. 00:38:45.162 [2024-12-06 17:54:26.545655] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.162 [2024-12-06 17:54:26.545689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.162 qpair failed and we were unable to recover it. 00:38:45.162 [2024-12-06 17:54:26.545782] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.162 [2024-12-06 17:54:26.545808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.162 qpair failed and we were unable to recover it. 00:38:45.162 [2024-12-06 17:54:26.545919] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.162 [2024-12-06 17:54:26.545949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.162 qpair failed and we were unable to recover it. 00:38:45.162 [2024-12-06 17:54:26.546099] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.162 [2024-12-06 17:54:26.546147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.162 qpair failed and we were unable to recover it. 00:38:45.162 [2024-12-06 17:54:26.546329] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.162 [2024-12-06 17:54:26.546379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.162 qpair failed and we were unable to recover it. 00:38:45.162 [2024-12-06 17:54:26.546489] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.162 [2024-12-06 17:54:26.546516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.162 qpair failed and we were unable to recover it. 00:38:45.162 [2024-12-06 17:54:26.546630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.162 [2024-12-06 17:54:26.546658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.162 qpair failed and we were unable to recover it. 00:38:45.162 [2024-12-06 17:54:26.546813] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.162 [2024-12-06 17:54:26.546841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.162 qpair failed and we were unable to recover it. 00:38:45.162 [2024-12-06 17:54:26.546932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.162 [2024-12-06 17:54:26.546959] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.162 qpair failed and we were unable to recover it. 00:38:45.162 [2024-12-06 17:54:26.547096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.162 [2024-12-06 17:54:26.547142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.162 qpair failed and we were unable to recover it. 00:38:45.162 [2024-12-06 17:54:26.547271] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.162 [2024-12-06 17:54:26.547315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.162 qpair failed and we were unable to recover it. 00:38:45.162 [2024-12-06 17:54:26.547456] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.162 [2024-12-06 17:54:26.547482] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.162 qpair failed and we were unable to recover it. 00:38:45.162 [2024-12-06 17:54:26.547590] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.162 [2024-12-06 17:54:26.547616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.162 qpair failed and we were unable to recover it. 00:38:45.162 [2024-12-06 17:54:26.547737] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.162 [2024-12-06 17:54:26.547765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.162 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.547854] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.547882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.548024] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.548057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.548169] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.548195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.548309] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.548336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.548423] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.548450] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.548591] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.548617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.548714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.548742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.548858] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.548885] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.548992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.549018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.549111] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.549139] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.549258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.549282] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.549391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.549416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.549524] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.549549] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.549678] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.549718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.549831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.549869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.550001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.550029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.550178] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.550226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.550381] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.550431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.550546] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.550572] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.550660] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.550692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.550780] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.550807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.550948] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.550974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.551113] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.551161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.551296] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.551347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.551439] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.551465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.551550] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.551576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.551690] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.551717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.551797] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.551823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.551914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.551941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.552022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.552048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.552165] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.552193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.552286] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.552313] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.552409] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.552436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.552554] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.552580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.552660] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.552691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.552771] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.552797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.552890] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.163 [2024-12-06 17:54:26.552919] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.163 qpair failed and we were unable to recover it. 00:38:45.163 [2024-12-06 17:54:26.552997] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.553023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.553106] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.553132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.553244] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.553271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.553385] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.553412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.553538] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.553583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.553716] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.553755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.553843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.553871] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.553953] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.553979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.554095] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.554120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.554223] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.554263] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.554415] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.554443] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.554573] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.554602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.554692] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.554719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.554829] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.554855] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.554939] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.554965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.555106] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.555156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.555277] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.555318] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.555441] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.555484] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.555600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.555627] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.555724] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.555751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.555839] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.555866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.555986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.556012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.556112] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.556143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.556245] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.556274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.556438] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.556468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.556701] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.556729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.556845] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.556873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.556950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.556976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.557063] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.557089] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.164 qpair failed and we were unable to recover it. 00:38:45.164 [2024-12-06 17:54:26.557214] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.164 [2024-12-06 17:54:26.557269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.557357] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.557382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.557469] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.557501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.557619] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.557645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.557760] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.557786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.557867] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.557895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.557982] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.558009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.558116] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.558141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.558249] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.558275] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.558396] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.558434] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.558561] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.558605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.558727] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.558756] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.558871] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.558898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.558978] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.559004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.559140] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.559171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.559301] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.559328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.559421] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.559448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.559543] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.559569] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.559716] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.559743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.559829] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.559857] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.560000] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.560027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.560111] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.560137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.560248] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.560274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.560349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.560375] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.560487] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.560514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.560610] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.560652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.560801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.560829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.560929] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.560967] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.561103] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.561147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.561298] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.561350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.561493] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.561518] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.561600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.561626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.561722] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.561749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.561876] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.561921] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.562052] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.562096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.562214] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.562262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.562370] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.562395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.165 [2024-12-06 17:54:26.562477] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.165 [2024-12-06 17:54:26.562502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.165 qpair failed and we were unable to recover it. 00:38:45.166 [2024-12-06 17:54:26.562615] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.166 [2024-12-06 17:54:26.562641] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.166 qpair failed and we were unable to recover it. 00:38:45.166 [2024-12-06 17:54:26.562725] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.166 [2024-12-06 17:54:26.562753] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.166 qpair failed and we were unable to recover it. 00:38:45.166 [2024-12-06 17:54:26.562867] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.166 [2024-12-06 17:54:26.562893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.166 qpair failed and we were unable to recover it. 00:38:45.166 [2024-12-06 17:54:26.563036] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.166 [2024-12-06 17:54:26.563063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.166 qpair failed and we were unable to recover it. 00:38:45.166 [2024-12-06 17:54:26.563147] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.166 [2024-12-06 17:54:26.563177] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.166 qpair failed and we were unable to recover it. 00:38:45.166 [2024-12-06 17:54:26.563263] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.166 [2024-12-06 17:54:26.563289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.166 qpair failed and we were unable to recover it. 00:38:45.166 [2024-12-06 17:54:26.563372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.166 [2024-12-06 17:54:26.563399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.166 qpair failed and we were unable to recover it. 00:38:45.166 [2024-12-06 17:54:26.563563] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.166 [2024-12-06 17:54:26.563589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.166 qpair failed and we were unable to recover it. 00:38:45.166 [2024-12-06 17:54:26.563688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.166 [2024-12-06 17:54:26.563727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.166 qpair failed and we were unable to recover it. 00:38:45.166 [2024-12-06 17:54:26.563848] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.166 [2024-12-06 17:54:26.563875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.166 qpair failed and we were unable to recover it. 00:38:45.166 [2024-12-06 17:54:26.563962] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.166 [2024-12-06 17:54:26.563989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.166 qpair failed and we were unable to recover it. 00:38:45.166 [2024-12-06 17:54:26.564086] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.166 [2024-12-06 17:54:26.564116] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.166 qpair failed and we were unable to recover it. 00:38:45.166 [2024-12-06 17:54:26.564269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.166 [2024-12-06 17:54:26.564311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.166 qpair failed and we were unable to recover it. 00:38:45.166 [2024-12-06 17:54:26.564396] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.166 [2024-12-06 17:54:26.564423] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.166 qpair failed and we were unable to recover it. 00:38:45.166 [2024-12-06 17:54:26.564540] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.166 [2024-12-06 17:54:26.564568] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.166 qpair failed and we were unable to recover it. 00:38:45.166 [2024-12-06 17:54:26.564660] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.166 [2024-12-06 17:54:26.564691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.166 qpair failed and we were unable to recover it. 00:38:45.166 [2024-12-06 17:54:26.564783] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.166 [2024-12-06 17:54:26.564810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.166 qpair failed and we were unable to recover it. 00:38:45.166 [2024-12-06 17:54:26.564924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.166 [2024-12-06 17:54:26.564951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.166 qpair failed and we were unable to recover it. 00:38:45.166 [2024-12-06 17:54:26.565097] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.166 [2024-12-06 17:54:26.565145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.166 qpair failed and we were unable to recover it. 00:38:45.166 [2024-12-06 17:54:26.565259] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.166 [2024-12-06 17:54:26.565307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.166 qpair failed and we were unable to recover it. 00:38:45.166 [2024-12-06 17:54:26.565425] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.166 [2024-12-06 17:54:26.565452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.166 qpair failed and we were unable to recover it. 00:38:45.166 [2024-12-06 17:54:26.565561] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.166 [2024-12-06 17:54:26.565587] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.166 qpair failed and we were unable to recover it. 00:38:45.166 [2024-12-06 17:54:26.565674] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.166 [2024-12-06 17:54:26.565701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.166 qpair failed and we were unable to recover it. 00:38:45.166 [2024-12-06 17:54:26.565789] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.166 [2024-12-06 17:54:26.565815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.166 qpair failed and we were unable to recover it. 00:38:45.166 [2024-12-06 17:54:26.565896] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.565923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.566042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.566068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.566174] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.566206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.566360] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.566416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.566542] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.566580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.566679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.566717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.566813] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.566839] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.567039] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.567089] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.567194] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.567224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.567377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.567427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.567514] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.567542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.567630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.567658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.567779] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.567805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.567938] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.567968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.568072] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.568102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.568211] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.568251] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.568375] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.568403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.568490] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.568517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.568639] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.568671] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.568765] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.568791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.568908] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.568934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.569028] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.569055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.569170] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.569197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.569314] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.569341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.569431] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.569457] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.569545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.569570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.569659] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.569690] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.569771] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.569797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.569880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.569908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.570022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.570048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.570128] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.570155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.570296] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.570323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.570447] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.570486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.570608] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.570637] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.570837] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.570864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.570949] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.570975] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.571119] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.571145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.167 qpair failed and we were unable to recover it. 00:38:45.167 [2024-12-06 17:54:26.571252] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.167 [2024-12-06 17:54:26.571281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.571451] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.168 [2024-12-06 17:54:26.571503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.571616] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.168 [2024-12-06 17:54:26.571642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.571759] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.168 [2024-12-06 17:54:26.571786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.571874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.168 [2024-12-06 17:54:26.571900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.572028] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.168 [2024-12-06 17:54:26.572070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.572200] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.168 [2024-12-06 17:54:26.572230] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.572367] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.168 [2024-12-06 17:54:26.572394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.572521] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.168 [2024-12-06 17:54:26.572563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.572678] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.168 [2024-12-06 17:54:26.572716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.572814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.168 [2024-12-06 17:54:26.572843] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.572941] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.168 [2024-12-06 17:54:26.572968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.573054] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.168 [2024-12-06 17:54:26.573081] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.573223] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.168 [2024-12-06 17:54:26.573262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.573375] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.168 [2024-12-06 17:54:26.573428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.573567] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.168 [2024-12-06 17:54:26.573607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.573762] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.168 [2024-12-06 17:54:26.573790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.573873] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.168 [2024-12-06 17:54:26.573901] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.574005] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.168 [2024-12-06 17:54:26.574034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.574241] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.168 [2024-12-06 17:54:26.574285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.574421] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.168 [2024-12-06 17:54:26.574466] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.574654] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.168 [2024-12-06 17:54:26.574685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.574831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.168 [2024-12-06 17:54:26.574856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.574937] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.168 [2024-12-06 17:54:26.574963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.575159] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.168 [2024-12-06 17:54:26.575204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.575301] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.168 [2024-12-06 17:54:26.575330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.575439] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.168 [2024-12-06 17:54:26.575465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.575541] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.168 [2024-12-06 17:54:26.575566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.575692] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.168 [2024-12-06 17:54:26.575718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.168 qpair failed and we were unable to recover it. 00:38:45.168 [2024-12-06 17:54:26.575800] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.575828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.575919] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.575945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.576084] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.576110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.576222] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.576248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.576335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.576361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.576445] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.576474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.576563] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.576590] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.576781] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.576807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.576934] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.576984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.577063] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.577088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.577173] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.577198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.577282] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.577308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.577391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.577416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.577497] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.577523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.577601] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.577628] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.577798] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.577838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.577957] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.577985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.578078] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.578107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.578222] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.578248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.578363] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.578390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.578510] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.578536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.578649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.578683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.578783] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.578809] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.578890] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.578916] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.579125] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.579154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.579308] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.579347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.579535] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.579573] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.579694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.579721] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.579834] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.579860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.579942] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.579970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.580070] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.580099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.580200] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.580227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.580359] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.580410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.580502] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.580528] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.580613] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.580638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.580759] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.580790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.580876] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.169 [2024-12-06 17:54:26.580903] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.169 qpair failed and we were unable to recover it. 00:38:45.169 [2024-12-06 17:54:26.581009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.581035] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.581149] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.581176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.581306] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.581345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.581449] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.581476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.581564] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.581591] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.581675] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.581704] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.581794] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.581820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.581933] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.581960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.582051] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.582077] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.582189] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.582214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.582338] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.582364] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.582467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.582506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.582611] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.582639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.582767] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.582793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.582907] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.582933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.583057] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.583087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.583174] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.583203] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.583358] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.583389] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.583524] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.583549] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.583638] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.583674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.583767] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.583795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.583874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.583900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.583999] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.584030] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.584202] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.584233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.584357] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.584401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.584529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.584560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.584640] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.584678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.584794] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.584819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.584906] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.584932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.585032] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.585077] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.585212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.585261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.585379] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.585407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.585500] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.585528] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.585648] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.585682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.585793] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.585820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.585903] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.585929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.586021] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.586047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.586159] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.170 [2024-12-06 17:54:26.586186] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.170 qpair failed and we were unable to recover it. 00:38:45.170 [2024-12-06 17:54:26.586280] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.586307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.586393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.586419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.586500] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.586526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.586671] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.586699] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.586815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.586841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.586952] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.586978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.587087] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.587117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.587285] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.587323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.587448] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.587475] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.587589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.587616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.587732] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.587758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.587847] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.587873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.587998] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.588042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.588178] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.588225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.588360] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.588407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.588585] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.588613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.588738] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.588765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.588847] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.588873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.588958] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.588985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.589076] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.589103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.589195] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.589222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.589342] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.589369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.589462] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.589489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.589568] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.589594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.589696] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.589723] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.589807] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.589833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.589948] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.589974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.590085] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.590117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.590270] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.590309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.590465] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.590504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.590622] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.590649] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.590744] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.590770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.590965] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.590991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.171 qpair failed and we were unable to recover it. 00:38:45.171 [2024-12-06 17:54:26.591090] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.171 [2024-12-06 17:54:26.591128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.591326] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.591375] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.591460] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.591486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.591568] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.591596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.591679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.591706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.591796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.591824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.591937] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.591978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.592144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.592194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.592349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.592401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.592490] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.592517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.592609] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.592634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.592726] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.592753] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.592833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.592858] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.593042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.593080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.593213] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.593261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.593373] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.593398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.593510] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.593536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.593626] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.593651] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.593772] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.593800] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.593912] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.593939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.594053] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.594079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.594166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.594192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.594286] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.594313] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.594425] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.594451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.594542] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.594569] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.594648] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.594681] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.594795] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.594821] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.594906] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.594932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.595050] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.595076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.595175] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.595214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.595307] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.595335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.595428] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.595455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.595537] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.595563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.595683] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.595710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.595827] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.595854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.595946] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.595973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.596064] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.596090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.596203] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.596229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.172 [2024-12-06 17:54:26.596320] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.172 [2024-12-06 17:54:26.596347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.172 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.596437] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.596463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.596571] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.596597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.596708] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.596734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.596858] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.596884] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.597005] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.597031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.597125] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.597150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.597235] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.597261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.597351] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.597376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.597513] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.597538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.597618] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.597643] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.597791] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.597817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.597897] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.597922] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.598002] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.598028] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.598133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.598159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.598242] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.598268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.598353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.598379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.598453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.598478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.598565] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.598590] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.598698] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.598724] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.598846] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.598872] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.598973] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.599001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.599109] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.599136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.599226] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.599252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.599334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.599361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.599452] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.599477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.599589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.599615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.599709] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.599735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.599846] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.599871] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.599952] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.599979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.600095] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.600121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.600245] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.600270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.600353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.600379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.600488] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.600514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.600593] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.600618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.600711] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.600737] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.600818] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.600844] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.600928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.600953] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.601047] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.173 [2024-12-06 17:54:26.601072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.173 qpair failed and we were unable to recover it. 00:38:45.173 [2024-12-06 17:54:26.601205] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.601231] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.601312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.601338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.601451] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.601477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.601554] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.601580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.601662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.601693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.601801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.601826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.601917] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.601944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.602057] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.602083] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.602203] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.602228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.602310] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.602335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.602419] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.602445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.602523] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.602549] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.602643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.602695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.602791] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.602820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.602912] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.602939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.603027] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.603054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.603191] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.603218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.603327] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.603353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.603464] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.603490] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.603571] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.603597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.603687] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.603714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.603858] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.603906] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.604081] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.604132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.604209] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.604234] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.604316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.604344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.604433] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.604459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.604557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.604585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.604677] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.604705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.604781] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.604807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.604892] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.604918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.605008] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.605035] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.605143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.605169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.605243] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.605269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.605386] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.605412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.605490] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.605516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.605599] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.605625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.605742] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.605770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.174 qpair failed and we were unable to recover it. 00:38:45.174 [2024-12-06 17:54:26.605880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.174 [2024-12-06 17:54:26.605907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.605992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.606019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.606128] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.606160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.606244] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.606271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.606375] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.606401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.606513] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.606539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.606690] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.606716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.606790] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.606816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.607003] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.607028] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.607214] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.607239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.607343] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.607368] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.607446] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.607472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.607575] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.607621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.607751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.607790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.607927] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.607982] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.608156] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.608205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.608350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.608400] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.608515] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.608541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.608667] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.608695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.608780] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.608805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.608913] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.608951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.609102] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.609150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.609250] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.609300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.609411] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.609437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.609520] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.609546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.609630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.609656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.609748] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.609774] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.609912] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.609937] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.610013] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.610038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.610144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.610175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.610298] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.610326] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.610418] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.610444] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.610540] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.610570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.610670] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.610699] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.610794] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.610821] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.610931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.610957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.611073] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.611112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.611260] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.611298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.175 [2024-12-06 17:54:26.611416] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.175 [2024-12-06 17:54:26.611452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.175 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.611562] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.611588] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.611692] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.611718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.611803] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.611829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.611908] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.611933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.612024] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.612050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.612137] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.612163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.612278] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.612304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.612384] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.612410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.612491] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.612517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.612603] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.612629] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.612751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.612777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.612860] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.612886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.612967] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.612993] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.613181] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.613206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.613314] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.613340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.613458] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.613484] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.613591] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.613616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.613701] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.613731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.613920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.613946] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.614029] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.614054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.614140] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.614165] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.614246] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.614272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.614349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.614375] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.614494] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.614533] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.614655] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.614693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.614779] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.614805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.614915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.614967] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.615128] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.615165] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.615349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.615386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.615513] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.615538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.615619] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.615644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.615750] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.615775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.615947] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.176 [2024-12-06 17:54:26.615982] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.176 qpair failed and we were unable to recover it. 00:38:45.176 [2024-12-06 17:54:26.616158] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.616209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.616356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.616406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.616492] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.616519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.616607] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.616633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.616724] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.616751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.616839] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.616865] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.616953] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.616980] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.617066] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.617092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.617210] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.617236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.617322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.617348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.617437] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.617464] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.617558] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.617596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.617683] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.617711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.617827] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.617852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.617939] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.617965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.618044] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.618088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.618240] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.618276] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.618395] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.618430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.618578] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.618603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.618689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.618715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.618788] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.618812] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.618890] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.618917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.619008] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.619033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.619145] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.619197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.619337] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.619386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.619498] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.619524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.619613] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.619639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.619767] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.619795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.619906] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.619933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.620048] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.620074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.620151] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.620177] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.620265] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.620291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.620382] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.620409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.620524] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.620551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.620671] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.620698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.620817] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.620842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.620923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.620949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.621064] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.177 [2024-12-06 17:54:26.621089] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.177 qpair failed and we were unable to recover it. 00:38:45.177 [2024-12-06 17:54:26.621220] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.621257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.621466] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.621508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.621627] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.621652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.621802] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.621830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.621915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.621941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.622061] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.622088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.622233] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.622281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.622395] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.622422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.622513] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.622540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.622626] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.622652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.622778] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.622804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.622920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.622946] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.623061] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.623088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.623197] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.623236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.623371] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.623398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.623490] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.623518] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.623600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.623625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.623706] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.623732] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.623874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.623900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.624012] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.624058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.624165] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.624200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.624323] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.624365] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.624514] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.624540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.624627] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.624653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.624752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.624777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.624888] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.624913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.624996] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.625043] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.625209] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.625246] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.625403] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.625439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.625596] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.625633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.625782] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.625808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.625897] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.625925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.626071] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.626122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.626262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.626312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.626422] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.626449] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.626536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.626561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.626671] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.626710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.626805] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.178 [2024-12-06 17:54:26.626833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.178 qpair failed and we were unable to recover it. 00:38:45.178 [2024-12-06 17:54:26.626920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.626947] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.627090] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.627116] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.627209] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.627234] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.627329] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.627355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.627474] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.627500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.627618] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.627643] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.627783] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.627829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.627961] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.627988] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.628069] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.628096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.628180] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.628206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.628299] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.628325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.628441] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.628467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.628547] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.628574] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.628685] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.628712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.628822] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.628849] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.628939] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.628970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.629046] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.629073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.629149] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.629175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.629262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.629289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.629377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.629404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.629492] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.629518] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.629635] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.629669] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.629757] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.629783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.629898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.629924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.630043] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.630069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.630179] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.630206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.630284] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.630309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.630401] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.630428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.630540] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.630566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.630704] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.630742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.630865] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.630894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.630981] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.631007] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.631121] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.631171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.631268] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.631294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.631372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.631398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.631508] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.631533] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.631724] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.631751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.631836] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.179 [2024-12-06 17:54:26.631862] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.179 qpair failed and we were unable to recover it. 00:38:45.179 [2024-12-06 17:54:26.631989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.632014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.632132] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.632157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.632260] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.632286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.632403] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.632431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.632531] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.632570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.632660] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.632705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.632791] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.632818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.632907] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.632933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.633046] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.633072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.633162] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.633189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.633282] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.633309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.633393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.633419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.633511] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.633538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.633620] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.633646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.633793] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.633819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.633925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.633951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.634030] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.634057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.634168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.634199] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.634308] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.634334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.634420] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.634447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.634536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.634562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.634646] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.634679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.634793] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.634820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.634906] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.634932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.635015] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.635041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.635157] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.635183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.635273] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.635300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.635391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.635420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.635533] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.635561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.635658] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.635689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.635776] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.635802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.635907] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.635945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.636063] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.636091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.180 [2024-12-06 17:54:26.636177] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.180 [2024-12-06 17:54:26.636204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.180 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.636340] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.636365] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.636450] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.636476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.636563] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.636591] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.636676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.636703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.636794] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.636820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.636932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.636958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.637034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.637060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.637150] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.637176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.637288] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.637315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.637403] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.637433] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.637514] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.637544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.637630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.637658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.637755] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.637780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.637886] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.637912] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.637991] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.638020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.638128] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.638153] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.638238] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.638263] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.638368] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.638406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.638504] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.638532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.638621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.638647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.638774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.638801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.638882] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.638908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.638987] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.639013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.639097] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.639125] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.639209] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.639236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.639326] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.639353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.639469] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.639497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.639572] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.639598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.639693] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.639722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.639806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.639831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.639941] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.639967] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.640078] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.640103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.640182] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.640207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.640323] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.640350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.640451] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.640477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.640556] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.640582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.640723] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.181 [2024-12-06 17:54:26.640749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.181 qpair failed and we were unable to recover it. 00:38:45.181 [2024-12-06 17:54:26.640838] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.640864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.640966] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.641007] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.641168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.641206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.641359] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.641396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.641536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.641562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.641648] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.641682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.641778] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.641803] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.641913] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.641939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.642070] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.642096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.642231] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.642268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.642461] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.642508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.642622] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.642649] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.642770] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.642797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.642885] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.642917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.643010] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.643037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.643146] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.643172] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.643252] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.643278] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.643401] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.643428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.643531] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.643557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.643648] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.643682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.643799] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.643825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.643945] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.643971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.644061] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.644087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.644193] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.644219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.644332] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.644358] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.644450] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.644479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.644592] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.644620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.644735] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.644761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.644843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.644869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.644981] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.645006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.645118] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.645144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.645258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.645298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.645457] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.645496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.645640] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.645690] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.645803] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.645828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.645915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.645941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.646058] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.646084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.646234] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.182 [2024-12-06 17:54:26.646271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.182 qpair failed and we were unable to recover it. 00:38:45.182 [2024-12-06 17:54:26.646395] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.646439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.646619] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.646657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.646795] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.646822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.646964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.646989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.647077] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.647102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.647236] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.647264] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.647345] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.647371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.647456] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.647482] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.647568] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.647595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.647703] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.647742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.647839] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.647867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.647955] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.647981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.648090] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.648116] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.648207] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.648233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.648348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.648374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.648458] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.648489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.648601] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.648627] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.648745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.648772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.648882] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.648908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.648981] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.649007] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.649085] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.649111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.649221] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.649247] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.649322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.649347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.649536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.649562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.649678] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.649706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.649815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.649841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.649929] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.649955] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.650056] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.650093] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.650223] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.650249] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.650344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.650371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.650483] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.650509] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.650620] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.650645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.650764] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.650794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.650873] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.650898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.651012] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.651038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.651137] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.651176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.651295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.651331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.651460] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.183 [2024-12-06 17:54:26.651485] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.183 qpair failed and we were unable to recover it. 00:38:45.183 [2024-12-06 17:54:26.651568] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.651594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.651713] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.651741] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.651828] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.651854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.651933] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.651958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.652043] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.652075] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.652155] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.652181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.652262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.652288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.652384] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.652411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.652525] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.652550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.652642] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.652676] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.652758] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.652784] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.652874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.652900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.652986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.653012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.653102] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.653128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.653207] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.653233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.653310] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.653340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.653590] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.653617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.653704] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.653731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.653828] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.653855] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.653944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.653971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.654086] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.654112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.654201] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.654228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.654347] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.654377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.654468] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.654493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.654574] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.654600] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.654687] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.654713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.654805] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.654831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.654964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.655004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.655165] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.655205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.655394] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.655431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.655578] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.655604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.655691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.655722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.655821] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.655848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.656022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.656061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.656226] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.656262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.656409] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.656446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.656587] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.656615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.656748] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.656786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.184 qpair failed and we were unable to recover it. 00:38:45.184 [2024-12-06 17:54:26.656901] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.184 [2024-12-06 17:54:26.656929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.657075] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.657120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.657232] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.657280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.657415] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.657462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.657548] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.657574] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.657660] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.657694] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.657777] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.657804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.657892] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.657919] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.658033] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.658061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.658176] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.658202] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.658286] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.658312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.658453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.658480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.658557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.658583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.658668] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.658694] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.658814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.658840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.658926] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.658952] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.659063] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.659089] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.659174] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.659200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.659313] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.659339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.659423] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.659449] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.659534] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.659565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.659676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.659702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.659786] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.659812] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.659892] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.659917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.659999] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.660025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.660130] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.660156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.660269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.660294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.660377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.660403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.660479] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.660505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.660593] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.660622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.660717] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.660743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.660824] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.660850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.660963] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.660990] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.661110] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.661140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.661252] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.185 [2024-12-06 17:54:26.661278] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.185 qpair failed and we were unable to recover it. 00:38:45.185 [2024-12-06 17:54:26.661425] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.661474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.661565] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.661590] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.661704] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.661730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.661805] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.661830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.661926] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.661963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.662144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.662195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.662306] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.662331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.662418] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.662445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.662557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.662582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.662681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.662708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.662798] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.662824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.662903] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.662929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.663080] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.663110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.663213] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.663239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.663342] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.663380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.663474] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.663501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.663608] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.663634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.663727] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.663754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.663833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.663859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.663969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.663995] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.664108] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.664134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.664258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.664287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.664373] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.664400] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.664494] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.664520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.664604] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.664630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.664718] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.664745] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.664832] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.664859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.664946] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.664972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.665077] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.665102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.665195] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.665222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.665308] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.665334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.665414] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.665440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.665516] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.665541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.665653] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.665687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.665772] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.665801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.665916] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.665944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.666082] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.666133] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.666269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.186 [2024-12-06 17:54:26.666321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.186 qpair failed and we were unable to recover it. 00:38:45.186 [2024-12-06 17:54:26.666412] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.666439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.666571] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.666609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.666693] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.666721] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.666871] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.666919] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.667071] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.667120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.667233] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.667282] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.667410] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.667451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.667590] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.667618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.667726] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.667754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.667840] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.667867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.668025] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.668063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.668218] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.668269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.668383] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.668409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.668495] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.668523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.668609] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.668635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.668773] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.668801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.668918] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.668968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.669102] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.669151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.669282] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.669331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.669477] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.669522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.669604] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.669631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.669782] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.669809] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.669924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.669950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.670039] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.670066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.670185] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.670211] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.670333] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.670360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.670490] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.670515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.670619] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.670660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.670811] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.670840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.670919] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.670945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.671024] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.671050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.671155] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.671192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.671318] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.671360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.671540] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.671577] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.671730] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.671758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.671874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.671900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.672012] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.672039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.672123] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.187 [2024-12-06 17:54:26.672149] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.187 qpair failed and we were unable to recover it. 00:38:45.187 [2024-12-06 17:54:26.672280] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.672328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.672421] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.672448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.672539] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.672567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.672677] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.672709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.672798] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.672825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.672945] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.672972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.673087] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.673123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.673251] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.673277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.673411] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.673448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.673611] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.673648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.673796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.673822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.673903] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.673929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.674055] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.674092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.674301] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.674338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.674484] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.674531] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.674611] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.674637] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.674730] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.674756] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.674872] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.674898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.675008] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.675045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.675194] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.675247] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.675404] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.675440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.675566] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.675607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.675745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.675771] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.675885] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.675914] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.676054] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.676103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.676216] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.676268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.676438] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.676488] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.676579] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.676606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.676695] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.676722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.676809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.676836] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.676992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.677037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.677162] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.677200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.677320] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.677347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.677459] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.677485] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.677567] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.677594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.677689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.677715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.677801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.677828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.677971] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.677997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.188 qpair failed and we were unable to recover it. 00:38:45.188 [2024-12-06 17:54:26.678089] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.188 [2024-12-06 17:54:26.678115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.678202] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.678229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.678313] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.678339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.678477] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.678504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.678589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.678615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.678706] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.678738] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.678823] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.678849] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.678958] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.678984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.679130] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.679167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.679285] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.679334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.679480] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.679505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.679618] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.679643] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.679746] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.679785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.679878] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.679907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.679999] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.680025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.680115] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.680143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.680290] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.680336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.680428] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.680454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.680597] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.680624] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.680725] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.680752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.680855] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.680881] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.680968] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.680994] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.681103] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.681129] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.681250] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.681277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.681363] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.681390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.681520] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.681559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.681655] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.681693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.681827] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.681853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.681962] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.681988] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.682120] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.682167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.682258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.682285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.682365] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.682392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.682470] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.682500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.682575] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.682602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.682693] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.682720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.682801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.682827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.682938] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.682965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.683101] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.683127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.189 qpair failed and we were unable to recover it. 00:38:45.189 [2024-12-06 17:54:26.683215] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.189 [2024-12-06 17:54:26.683241] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.683335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.683362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.683444] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.683471] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.683567] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.683593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.683703] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.683742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.683882] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.683909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.683998] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.684025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.684111] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.684136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.684276] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.684305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.684420] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.684445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.684558] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.684585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.684722] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.684761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.684864] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.684909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.685042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.685070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.685152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.685178] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.685289] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.685341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.685427] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.685454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.685541] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.685568] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.685695] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.685733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.685880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.685907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.686017] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.686043] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.686129] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.686155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.686248] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.686274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.686356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.686383] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.686466] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.686493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.686610] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.686637] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.686762] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.686789] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.686898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.686925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.687019] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.687046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.687158] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.687183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.687301] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.687330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.687448] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.687473] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.687550] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.687577] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.687674] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.687701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.190 [2024-12-06 17:54:26.687776] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.190 [2024-12-06 17:54:26.687807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.190 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.687889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.687914] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.687995] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.688022] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.688106] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.688132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.688208] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.688234] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.688318] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.688344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.688459] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.688485] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.688602] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.688629] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.688748] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.688774] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.688853] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.688879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.688954] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.688981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.689096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.689121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.689209] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.689235] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.689346] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.689383] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.689542] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.689579] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.689757] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.689783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.689865] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.689890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.689992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.690018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.690093] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.690118] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.690261] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.690298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.690419] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.690444] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.690613] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.690638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.690757] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.690785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.690871] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.690898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.690985] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.691011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.691155] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.691206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.691343] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.691392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.691524] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.691551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.691644] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.691676] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.691770] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.691796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.691881] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.691907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.692064] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.692101] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.692311] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.692347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.692459] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.692506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.692623] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.692651] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.692759] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.692785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.692861] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.692887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.692972] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.692998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.191 [2024-12-06 17:54:26.693124] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.191 [2024-12-06 17:54:26.693150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.191 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.693327] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.693364] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.693506] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.693536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.693650] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.693684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.693767] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.693792] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.693885] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.693913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.694006] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.694033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.694144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.694171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.694260] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.694286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.694371] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.694398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.694482] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.694508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.694633] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.694659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.694758] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.694785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.694869] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.694895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.694972] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.694998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.695101] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.695127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.695219] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.695245] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.695353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.695379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.695528] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.695554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.695674] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.695702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.695789] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.695817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.695915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.695954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.696051] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.696080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.696271] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.696297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.696379] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.696405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.696484] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.696509] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.696627] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.696654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.696789] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.696815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.696930] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.696956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.697068] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.697099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.697213] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.697240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.697333] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.697359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.697453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.697493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.697581] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.697607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.697722] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.697750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.697838] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.697864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.697943] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.697969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.698042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.698068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.192 [2024-12-06 17:54:26.698151] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.192 [2024-12-06 17:54:26.698179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.192 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.698295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.698324] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.698439] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.698465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.698581] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.698608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.698691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.698718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.698839] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.698866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.698999] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.699052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.699199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.699250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.699369] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.699395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.699503] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.699529] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.699623] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.699648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.699744] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.699770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.699852] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.699878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.700067] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.700092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.700176] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.700201] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.700286] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.700314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.700404] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.700430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.700541] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.700567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.700657] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.700694] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.700789] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.700814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.700898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.700924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.701033] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.701060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.701250] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.701275] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.701389] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.701415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.701496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.701522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.701605] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.701630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.701714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.701740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.701811] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.701837] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.701931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.701969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.702053] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.702080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.702198] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.702224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.702310] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.702338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.702468] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.702495] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.702591] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.702619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.702710] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.702737] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.702847] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.702873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.702961] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.702987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.703104] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.703130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.703240] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.193 [2024-12-06 17:54:26.703266] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.193 qpair failed and we were unable to recover it. 00:38:45.193 [2024-12-06 17:54:26.703375] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.703400] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.703517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.703543] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.703632] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.703657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.703746] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.703772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.703855] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.703881] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.704023] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.704049] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.704131] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.704162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.704242] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.704267] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.704353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.704379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.704493] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.704519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.704603] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.704629] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.704785] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.704812] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.704924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.704949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.705084] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.705110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.705227] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.705253] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.705328] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.705354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.705436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.705462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.705573] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.705599] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.705743] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.705769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.705857] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.705883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.705964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.705990] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.706066] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.706091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.706203] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.706229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.706321] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.706349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.706457] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.706496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.706600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.706639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.706774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.706819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.706954] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.706980] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.707125] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.707151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.707295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.707343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.707434] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.707459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.707575] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.707601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.707706] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.707732] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.707819] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.707848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.194 [2024-12-06 17:54:26.707941] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.194 [2024-12-06 17:54:26.707967] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.194 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.708076] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.708102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.708191] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.708216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.708350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.708379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.708501] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.708528] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.708643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.708675] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.708793] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.708818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.708907] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.708933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.709025] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.709051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.709139] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.709165] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.709281] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.709307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.709391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.709416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.709501] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.709527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.709615] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.709641] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.709798] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.709832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.709921] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.709970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.710098] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.710151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.710298] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.710325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.710401] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.710427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.710556] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.710596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.710681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.710710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.710827] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.710855] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.710989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.711037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.711149] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.711199] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.711346] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.711394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.711515] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.711542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.711625] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.711660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.711753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.711780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.711893] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.711920] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.712063] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.712110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.712246] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.712293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.712435] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.712485] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.712563] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.712589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.712674] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.712701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.712843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.712888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.713058] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.713107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.713249] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.713297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.713381] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.713409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.713503] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.195 [2024-12-06 17:54:26.713529] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.195 qpair failed and we were unable to recover it. 00:38:45.195 [2024-12-06 17:54:26.713618] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.713645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.713804] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.713854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.713991] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.714047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.714200] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.714240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.714399] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.714436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.714578] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.714604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.714719] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.714746] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.714861] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.714886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.714973] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.714998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.715109] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.715144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.715348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.715388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.715528] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.715566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.715696] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.715742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.715820] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.715845] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.715963] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.715992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.716134] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.716184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.716277] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.716304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.716414] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.716440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.716528] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.716555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.716675] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.716703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.716819] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.716845] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.716924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.716951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.717035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.717061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.717153] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.717179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.717261] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.717288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.717413] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.717452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.717539] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.717567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.717686] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.717717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.717837] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.717865] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.717952] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.717978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.718057] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.718083] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.718226] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.718263] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.718385] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.718428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.718570] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.718606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.718768] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.718796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.718918] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.718944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.719058] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.719084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.719231] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.196 [2024-12-06 17:54:26.719280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.196 qpair failed and we were unable to recover it. 00:38:45.196 [2024-12-06 17:54:26.719421] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.719447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.719532] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.719558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.719638] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.719671] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.719772] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.719799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.719940] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.719967] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.720058] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.720085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.720202] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.720228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.720345] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.720372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.720460] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.720487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.720591] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.720630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.720732] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.720760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.720844] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.720870] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.720981] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.721032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.721138] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.721189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.721278] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.721305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.721414] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.721439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.721516] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.721547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.721668] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.721695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.721786] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.721812] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.721922] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.721948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.722116] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.722150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.722318] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.722367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.722480] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.722507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.722651] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.722687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.722769] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.722795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.722886] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.722911] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.723108] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.723162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.723312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.723352] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.723508] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.723534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.723646] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.723684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.723809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.723835] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.723925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.723951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.724092] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.724117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.724222] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.724259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.724377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.724412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.724560] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.724597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.724765] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.724792] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.724884] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.197 [2024-12-06 17:54:26.724910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.197 qpair failed and we were unable to recover it. 00:38:45.197 [2024-12-06 17:54:26.725053] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.725078] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.725184] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.725209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.725349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.725392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.725473] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.725498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.725612] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.725639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.725738] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.725776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.725870] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.725898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.726021] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.726047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.726236] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.726274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.726401] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.726445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.726567] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.726593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.726709] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.726736] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.726822] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.726848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.726960] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.726986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.727093] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.727119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.727231] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.727257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.727344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.727369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.727473] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.727512] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.727638] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.727690] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.727790] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.727818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.727910] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.727937] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.728024] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.728052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.728193] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.728219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.728360] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.728410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.728529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.728557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.728651] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.728688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.728830] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.728857] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.728943] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.728969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.729082] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.729109] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.729250] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.729301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.729476] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.729525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.729631] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.729657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.729761] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.729787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.729869] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.729896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.729990] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.730029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.730184] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.730211] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.730345] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.730382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.730498] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.730534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.198 [2024-12-06 17:54:26.730689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.198 [2024-12-06 17:54:26.730734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.198 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.730827] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.730853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.730944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.730971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.731086] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.731135] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.731238] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.731290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.731430] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.731480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.731587] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.731613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.731725] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.731757] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.731849] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.731875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.731989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.732015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.732127] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.732153] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.732244] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.732270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.732396] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.732436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.732565] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.732605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.732705] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.732732] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.732846] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.732872] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.732987] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.733013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.733122] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.733147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.733235] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.733261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.733383] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.733411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.733528] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.733558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.733684] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.733711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.733839] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.733865] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.733992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.734043] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.734120] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.734145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.734263] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.734300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.734445] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.734481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.734618] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.734657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.734792] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.734820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.734955] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.735002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.735176] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.735222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.735357] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.735394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.735525] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.735550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.735661] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.735695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.199 [2024-12-06 17:54:26.735863] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.199 [2024-12-06 17:54:26.735922] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.199 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.736054] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.736091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.736255] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.736304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.736414] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.736439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.736521] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.736546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.736672] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.736698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.736815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.736840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.736919] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.736944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.737050] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.737076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.737162] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.737187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.737290] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.737315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.737394] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.737420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.737563] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.737588] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.737714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.737740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.737833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.737860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.737938] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.737963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.738047] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.738073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.738192] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.738217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.738291] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.738316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.738417] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.738456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.738589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.738628] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.738765] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.738805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.738922] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.738950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.739061] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.739088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.739177] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.739208] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.739353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.739402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.739501] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.739540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.739635] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.739674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.739800] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.739827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.739951] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.739989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.740151] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.740192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.740370] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.740424] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.740517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.740544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.740639] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.740670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.740755] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.740781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.740865] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.740891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.741015] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.741066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.741144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.741170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.200 [2024-12-06 17:54:26.741259] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.200 [2024-12-06 17:54:26.741287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.200 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.741411] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.741449] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.741568] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.741595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.741715] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.741742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.741832] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.741858] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.741995] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.742021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.742128] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.742153] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.742273] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.742301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.742387] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.742414] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.742528] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.742555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.742642] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.742673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.742788] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.742814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.742955] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.743004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.743111] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.743159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.743233] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.743258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.743401] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.743427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.743538] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.743570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.743702] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.743741] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.743831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.743858] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.744002] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.744039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.744184] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.744220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.744392] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.744441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.744557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.744584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.744675] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.744701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.744790] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.744816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.744957] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.745004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.745141] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.745196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.745330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.745378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.745494] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.745522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.745661] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.745694] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.745793] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.745820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.745955] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.745992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.746110] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.746145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.746256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.746292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.746446] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.746482] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.746636] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.746685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.746825] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.746851] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.746978] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.747004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.747144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.201 [2024-12-06 17:54:26.747191] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.201 qpair failed and we were unable to recover it. 00:38:45.201 [2024-12-06 17:54:26.747319] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.747357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.747488] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.747513] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.747599] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.747625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.747720] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.747746] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.747858] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.747889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.747975] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.748001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.748113] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.748139] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.748257] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.748283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.748393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.748419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.748552] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.748591] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.748721] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.748760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.748852] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.748879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.748993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.749019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.749098] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.749124] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.749233] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.749259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.749399] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.749454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.749569] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.749598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.749685] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.749713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.749834] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.749861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.749945] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.749971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.750070] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.750096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.750236] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.750263] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.750378] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.750405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.750539] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.750577] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.750700] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.750728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.750847] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.750874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.750964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.750991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.751115] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.751142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.751251] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.751278] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.751364] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.751391] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.751510] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.751539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.751661] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.751701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.751785] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.751810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.751950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.751976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.752111] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.752147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.752358] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.752396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.752551] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.752581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.752691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.202 [2024-12-06 17:54:26.752717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.202 qpair failed and we were unable to recover it. 00:38:45.202 [2024-12-06 17:54:26.752836] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.752861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.752971] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.752997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.753138] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.753176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.753357] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.753396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.753524] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.753549] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.753660] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.753691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.753799] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.753829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.753920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.753949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.754033] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.754059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.754148] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.754175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.754285] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.754341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.754478] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.754505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.754585] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.754611] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.754711] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.754737] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.754846] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.754873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.754984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.755010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.755093] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.755120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.755238] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.755265] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.755382] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.755409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.755529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.755555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.755645] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.755682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.755772] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.755798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.755877] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.755904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.755997] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.756023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.756172] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.756198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.756317] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.756345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.756429] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.756455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.756535] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.756561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.756698] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.756724] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.756806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.756832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.756911] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.756937] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.757047] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.757073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.757237] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.757282] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.757451] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.203 [2024-12-06 17:54:26.757488] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.203 qpair failed and we were unable to recover it. 00:38:45.203 [2024-12-06 17:54:26.757632] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.757657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.757780] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.757806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.757882] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.757908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.758023] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.758048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.758166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.758204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.758397] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.758434] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.758576] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.758601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.758686] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.758712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.758824] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.758850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.758932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.758959] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.759079] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.759104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.759219] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.759256] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.759464] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.759508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.759631] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.759681] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.759799] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.759825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.759940] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.759966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.760080] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.760105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.760262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.760298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.760460] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.760496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.760650] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.760683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.760801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.760827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.760955] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.760991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.761103] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.761140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.761267] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.761303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.761520] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.761582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.761710] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.761739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.761924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.761971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.762119] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.762169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.762306] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.762354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.762436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.762463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.762569] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.762595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.762704] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.762732] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.762844] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.762870] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.762994] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.763020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.763132] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.763159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.763271] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.763297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.763448] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.763474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.763555] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.763582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.763678] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.763705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.204 qpair failed and we were unable to recover it. 00:38:45.204 [2024-12-06 17:54:26.763804] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.204 [2024-12-06 17:54:26.763831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.763914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.763939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.764052] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.764078] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.764176] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.764212] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.764354] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.764390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.764545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.764571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.764684] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.764733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.764892] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.764931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.765064] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.765102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.765279] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.765315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.765459] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.765496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.765651] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.765713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.765856] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.765893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.766013] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.766075] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.766235] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.766274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.766398] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.766436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.766589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.766627] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.766788] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.766813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.766923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.766948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.767053] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.767078] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.767216] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.767255] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.767425] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.767463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.767648] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.767696] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.767808] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.767833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.767912] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.767937] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.768051] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.768076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.768224] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.768263] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.768444] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.768489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.768642] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.768690] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.768821] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.768846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.768985] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.769011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.769142] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.769179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.769305] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.769343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.769522] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.769560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.769688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.769735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.769853] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.769878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.770035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.770072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.770230] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.770279] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.770437] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.770475] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.770648] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.770681] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.770768] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.770798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.205 [2024-12-06 17:54:26.770909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.205 [2024-12-06 17:54:26.770953] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.205 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.771112] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.771150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.771364] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.771403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.771581] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.771644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.771828] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.771853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.771963] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.772010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.772188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.772224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.772367] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.772410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.772558] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.772619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.772788] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.772812] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.772926] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.772951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.773042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.773090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.773221] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.773259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.773456] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.773520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.773715] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.773742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.773882] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.773908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.774124] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.774160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.774298] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.774334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.774499] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.774537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.774648] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.774687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.774835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.774861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.774962] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.774999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.775126] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.775151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.775300] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.775325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.775497] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.775535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.775662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.775692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.775809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.775833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.775948] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.775982] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.776120] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.776157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.776311] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.776349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.776466] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.776504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.776679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.776719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.776837] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.776865] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.776966] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.777006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.777216] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.777264] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.777396] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.777456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.777576] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.777603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.777725] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.777752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.777865] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.777891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.778038] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.778067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.778181] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.778226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.206 [2024-12-06 17:54:26.778350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.206 [2024-12-06 17:54:26.778388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.206 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.778573] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.778611] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.778765] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.778790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.778906] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.778931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.779051] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.779089] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.779217] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.779262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.779422] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.779470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.779597] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.779634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.779814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.779840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.779966] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.780006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.780190] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.780229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.780476] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.780541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.780738] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.780763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.780879] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.780905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.781065] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.781090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.781193] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.781219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.781381] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.781417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.781538] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.781563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.781681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.781707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.781844] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.781869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.781954] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.781979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.782090] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.782114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.782257] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.782295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.782418] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.782460] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.782591] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.782617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.782769] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.782796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.782908] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.782932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.783057] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.783099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.783230] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.783269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.783455] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.783492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.783631] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.783680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.783819] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.783843] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.783954] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.783979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.784093] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.784119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.784262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.784301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.784487] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.784546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.784740] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.784765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.784858] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.784883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.785029] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.785085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.207 qpair failed and we were unable to recover it. 00:38:45.207 [2024-12-06 17:54:26.785219] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.207 [2024-12-06 17:54:26.785256] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.785378] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.785414] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.785524] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.785562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.785689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.785742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.785853] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.785878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.786031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.786058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.786143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.786185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.786368] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.786422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.786599] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.786625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.786766] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.786791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.786931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.786986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.787121] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.787173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.787343] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.787384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.787599] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.787638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.787817] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.787842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.787919] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.787943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.788105] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.788130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.788240] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.788265] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.788376] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.788401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.788514] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.788538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.788691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.788730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.788894] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.788933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.789129] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.789166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.789315] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.789353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.789481] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.789519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.789690] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.789728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.789865] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.789903] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.790038] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.790075] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.790236] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.790275] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.790425] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.790463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.790585] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.790623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.790784] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.790823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.790983] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.791022] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.791177] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.791215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.791341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.791378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.791526] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.791564] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.791738] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.791778] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.791914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.791951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.792156] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.792193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.792379] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.792424] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.792546] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.792584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.792713] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.792751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.792905] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.792943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.793204] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.793242] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.208 [2024-12-06 17:54:26.793374] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.208 [2024-12-06 17:54:26.793413] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.208 qpair failed and we were unable to recover it. 00:38:45.209 [2024-12-06 17:54:26.793549] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.209 [2024-12-06 17:54:26.793603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.209 qpair failed and we were unable to recover it. 00:38:45.209 [2024-12-06 17:54:26.793784] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.209 [2024-12-06 17:54:26.793821] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.209 qpair failed and we were unable to recover it. 00:38:45.209 [2024-12-06 17:54:26.793990] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.209 [2024-12-06 17:54:26.794027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.209 qpair failed and we were unable to recover it. 00:38:45.209 [2024-12-06 17:54:26.794146] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.209 [2024-12-06 17:54:26.794183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.209 qpair failed and we were unable to recover it. 00:38:45.209 [2024-12-06 17:54:26.794371] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.209 [2024-12-06 17:54:26.794427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.209 qpair failed and we were unable to recover it. 00:38:45.209 [2024-12-06 17:54:26.794632] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.209 [2024-12-06 17:54:26.794701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.209 qpair failed and we were unable to recover it. 00:38:45.209 [2024-12-06 17:54:26.794839] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.209 [2024-12-06 17:54:26.794877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.209 qpair failed and we were unable to recover it. 00:38:45.209 [2024-12-06 17:54:26.795021] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.209 [2024-12-06 17:54:26.795061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.209 qpair failed and we were unable to recover it. 00:38:45.209 [2024-12-06 17:54:26.795198] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.209 [2024-12-06 17:54:26.795237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.209 qpair failed and we were unable to recover it. 00:38:45.209 [2024-12-06 17:54:26.795404] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.209 [2024-12-06 17:54:26.795445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.209 qpair failed and we were unable to recover it. 00:38:45.209 [2024-12-06 17:54:26.795614] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.209 [2024-12-06 17:54:26.795655] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.209 qpair failed and we were unable to recover it. 00:38:45.209 [2024-12-06 17:54:26.795837] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.209 [2024-12-06 17:54:26.795877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.209 qpair failed and we were unable to recover it. 00:38:45.209 [2024-12-06 17:54:26.796028] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.209 [2024-12-06 17:54:26.796067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.209 qpair failed and we were unable to recover it. 00:38:45.209 [2024-12-06 17:54:26.796227] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.209 [2024-12-06 17:54:26.796268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.209 qpair failed and we were unable to recover it. 00:38:45.209 [2024-12-06 17:54:26.796462] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.209 [2024-12-06 17:54:26.796501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.209 qpair failed and we were unable to recover it. 00:38:45.209 [2024-12-06 17:54:26.796661] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.209 [2024-12-06 17:54:26.796707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.209 qpair failed and we were unable to recover it. 00:38:45.209 [2024-12-06 17:54:26.796822] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.209 [2024-12-06 17:54:26.796858] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.209 qpair failed and we were unable to recover it. 00:38:45.209 [2024-12-06 17:54:26.797009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.209 [2024-12-06 17:54:26.797049] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.209 qpair failed and we were unable to recover it. 00:38:45.209 [2024-12-06 17:54:26.797210] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.209 [2024-12-06 17:54:26.797250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.209 qpair failed and we were unable to recover it. 00:38:45.209 [2024-12-06 17:54:26.797413] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.209 [2024-12-06 17:54:26.797452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.209 qpair failed and we were unable to recover it. 00:38:45.209 [2024-12-06 17:54:26.797585] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.209 [2024-12-06 17:54:26.797624] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.209 qpair failed and we were unable to recover it. 00:38:45.209 [2024-12-06 17:54:26.797767] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.209 [2024-12-06 17:54:26.797809] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.209 qpair failed and we were unable to recover it. 00:38:45.209 [2024-12-06 17:54:26.797980] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.209 [2024-12-06 17:54:26.798023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.209 qpair failed and we were unable to recover it. 00:38:45.209 [2024-12-06 17:54:26.798188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.209 [2024-12-06 17:54:26.798228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.209 qpair failed and we were unable to recover it. 00:38:45.209 [2024-12-06 17:54:26.798378] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.209 [2024-12-06 17:54:26.798413] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.209 qpair failed and we were unable to recover it. 00:38:45.209 [2024-12-06 17:54:26.798558] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.209 [2024-12-06 17:54:26.798594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.209 qpair failed and we were unable to recover it. 00:38:45.209 [2024-12-06 17:54:26.798825] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.209 [2024-12-06 17:54:26.798863] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.209 qpair failed and we were unable to recover it. 00:38:45.209 [2024-12-06 17:54:26.799018] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.799055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.799221] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.799275] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.799421] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.799460] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.799720] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.799762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.800009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.800080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.800264] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.800304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.800505] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.800541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.800721] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.800765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.800945] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.800981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.801160] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.801195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.801354] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.801391] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.801524] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.801564] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.801728] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.801765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.801916] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.801971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.802110] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.802152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.802314] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.802357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.802512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.802552] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.802714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.802754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.802970] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.803008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.803216] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.803255] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.803415] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.803455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.803623] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.803685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.803854] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.803896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.804085] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.804126] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.804260] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.804300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.804456] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.804501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.804684] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.804730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.804850] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.804889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.805033] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.805073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.805196] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.805235] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.210 [2024-12-06 17:54:26.805362] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.210 [2024-12-06 17:54:26.805407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.210 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.805576] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.805616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.805797] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.805838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.805955] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.805999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.806176] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.806217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.806436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.806477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.806614] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.806654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.806832] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.806872] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.807035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.807074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.807222] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.807263] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.807457] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.807498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.807692] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.807734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.807929] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.807972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.808141] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.808183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.808350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.808394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.808578] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.808616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.808787] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.808821] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.808950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.808997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.809229] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.809268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.809406] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.809457] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.809598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.809630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.809754] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.809786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.809902] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.809944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.810126] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.810170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.810357] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.810397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.810558] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.810590] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.810754] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.810794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.810966] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.811025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.811197] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.811240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.811387] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.811421] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.811562] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.811594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.811751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.811805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.211 qpair failed and we were unable to recover it. 00:38:45.211 [2024-12-06 17:54:26.812001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.211 [2024-12-06 17:54:26.812077] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.812309] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.812377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.812524] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.812557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.812684] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.812718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.812941] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.812999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.813173] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.813230] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.813413] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.813445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.813584] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.813617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.813762] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.813817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.813971] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.814028] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.814186] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.814240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.814431] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.814472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.814645] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.814698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.814859] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.814898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.815047] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.815100] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.815327] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.815371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.815576] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.815610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.815795] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.815838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.816015] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.816059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.816246] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.816292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.816444] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.816478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.816617] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.816651] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.816839] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.816878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.817053] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.817091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.817261] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.817306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.817522] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.817561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.817735] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.817770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.817924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.817983] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.818135] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.818181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.818319] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.818364] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.818519] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.818553] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.818714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.818749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.818882] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.212 [2024-12-06 17:54:26.818921] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.212 qpair failed and we were unable to recover it. 00:38:45.212 [2024-12-06 17:54:26.819089] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.213 [2024-12-06 17:54:26.819127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.213 qpair failed and we were unable to recover it. 00:38:45.213 [2024-12-06 17:54:26.819364] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.213 [2024-12-06 17:54:26.819409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.213 qpair failed and we were unable to recover it. 00:38:45.213 [2024-12-06 17:54:26.819575] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.213 [2024-12-06 17:54:26.819609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.213 qpair failed and we were unable to recover it. 00:38:45.213 [2024-12-06 17:54:26.819770] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.213 [2024-12-06 17:54:26.819805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.213 qpair failed and we were unable to recover it. 00:38:45.213 [2024-12-06 17:54:26.820020] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.213 [2024-12-06 17:54:26.820064] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.213 qpair failed and we were unable to recover it. 00:38:45.213 [2024-12-06 17:54:26.820275] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.213 [2024-12-06 17:54:26.820319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.213 qpair failed and we were unable to recover it. 00:38:45.213 [2024-12-06 17:54:26.820519] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.213 [2024-12-06 17:54:26.820554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.213 qpair failed and we were unable to recover it. 00:38:45.213 [2024-12-06 17:54:26.820684] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.213 [2024-12-06 17:54:26.820719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.213 qpair failed and we were unable to recover it. 00:38:45.213 [2024-12-06 17:54:26.820877] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.213 [2024-12-06 17:54:26.820915] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.213 qpair failed and we were unable to recover it. 00:38:45.213 [2024-12-06 17:54:26.821056] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.213 [2024-12-06 17:54:26.821112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.213 qpair failed and we were unable to recover it. 00:38:45.213 [2024-12-06 17:54:26.821295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.213 [2024-12-06 17:54:26.821345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.213 qpair failed and we were unable to recover it. 00:38:45.213 [2024-12-06 17:54:26.821500] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.213 [2024-12-06 17:54:26.821534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.213 qpair failed and we were unable to recover it. 00:38:45.213 [2024-12-06 17:54:26.821652] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.213 [2024-12-06 17:54:26.821711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.213 qpair failed and we were unable to recover it. 00:38:45.213 [2024-12-06 17:54:26.821874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.213 [2024-12-06 17:54:26.821912] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.213 qpair failed and we were unable to recover it. 00:38:45.213 [2024-12-06 17:54:26.822106] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.213 [2024-12-06 17:54:26.822144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.213 qpair failed and we were unable to recover it. 00:38:45.213 [2024-12-06 17:54:26.822266] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.213 [2024-12-06 17:54:26.822305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.213 qpair failed and we were unable to recover it. 00:38:45.213 [2024-12-06 17:54:26.822466] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.213 [2024-12-06 17:54:26.822502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.213 qpair failed and we were unable to recover it. 00:38:45.213 [2024-12-06 17:54:26.822682] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.213 [2024-12-06 17:54:26.822740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.213 qpair failed and we were unable to recover it. 00:38:45.213 [2024-12-06 17:54:26.822869] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.213 [2024-12-06 17:54:26.822918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.213 qpair failed and we were unable to recover it. 00:38:45.213 [2024-12-06 17:54:26.823086] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.213 [2024-12-06 17:54:26.823129] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.213 qpair failed and we were unable to recover it. 00:38:45.213 [2024-12-06 17:54:26.823315] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.213 [2024-12-06 17:54:26.823354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.213 qpair failed and we were unable to recover it. 00:38:45.213 [2024-12-06 17:54:26.823510] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.213 [2024-12-06 17:54:26.823544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.213 qpair failed and we were unable to recover it. 00:38:45.213 [2024-12-06 17:54:26.823715] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.213 [2024-12-06 17:54:26.823754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.213 qpair failed and we were unable to recover it. 00:38:45.213 [2024-12-06 17:54:26.823923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.213 [2024-12-06 17:54:26.823977] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.213 qpair failed and we were unable to recover it. 00:38:45.213 [2024-12-06 17:54:26.824143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.213 [2024-12-06 17:54:26.824189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.213 qpair failed and we were unable to recover it. 00:38:45.213 [2024-12-06 17:54:26.824370] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.213 [2024-12-06 17:54:26.824403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.213 qpair failed and we were unable to recover it. 00:38:45.213 [2024-12-06 17:54:26.824515] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.824551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.824688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.824742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.824907] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.824945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.825137] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.825182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.825347] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.825405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.825574] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.825607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.825781] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.825840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.825986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.826032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.826228] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.826273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.826460] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.826500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.826641] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.826719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.826856] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.826889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.827079] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.827124] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.827276] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.827325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.827475] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.827509] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.827730] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.827781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.827947] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.827985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.828173] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.828218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.828406] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.828440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.828609] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.828643] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.828789] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.828828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.829005] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.829046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.829275] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.829309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.829453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.829486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.829654] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.829718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.829871] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.829908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.830104] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.830153] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.830332] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.830377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.830555] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.830589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.830764] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.830803] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.830923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.830961] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.831204] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.831241] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.214 qpair failed and we were unable to recover it. 00:38:45.214 [2024-12-06 17:54:26.831360] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.214 [2024-12-06 17:54:26.831394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.831564] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.831598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.831747] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.831786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.831995] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.832042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.832227] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.832274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.832429] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.832485] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.832627] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.832659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.832802] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.832844] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.833009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.833058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.833244] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.833290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.833502] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.833536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.833715] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.833749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.833988] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.834035] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.834168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.834214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.834372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.834426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.834623] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.834656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.834832] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.834874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.835017] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.835062] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.835198] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.835257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.835414] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.835459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.835636] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.835696] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.835871] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.835910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.836108] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.836154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.836354] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.836398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.836627] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.836674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.836843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.836883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.837010] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.837069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.837233] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.837265] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.837470] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.837503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.837612] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.837650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.837836] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.837874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.838020] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.838058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.838256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.215 [2024-12-06 17:54:26.838301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.215 qpair failed and we were unable to recover it. 00:38:45.215 [2024-12-06 17:54:26.838481] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.838515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.838645] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.838692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.838838] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.838874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.839076] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.839120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.839258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.839301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.839491] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.839525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.839678] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.839713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.839813] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.839864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.840013] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.840065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.840245] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.840289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.840475] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.840510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.840651] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.840693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.840823] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.840876] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.841022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.841062] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.841250] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.841307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.841514] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.841547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.841662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.841706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.841865] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.841902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.842091] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.842136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.842371] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.842406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.842575] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.842607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.842742] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.842786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.842942] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.842986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.843152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.843202] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.843462] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.843496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.843706] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.843746] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.843906] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.843961] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.844122] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.844170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.844368] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.844421] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.844557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.844591] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.844724] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.844758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.844903] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.844935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.216 [2024-12-06 17:54:26.845113] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.216 [2024-12-06 17:54:26.845169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.216 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.845323] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.845368] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.217 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.845546] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.845581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.217 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.845784] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.845830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.217 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.846009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.846058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.217 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.846251] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.846297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.217 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.846472] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.846533] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.217 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.846682] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.846715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.217 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.846869] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.846913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.217 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.847073] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.847111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.217 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.847382] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.847419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.217 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.847530] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.847562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.217 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.847725] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.847773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.217 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.847994] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.848046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.217 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.848200] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.848269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.217 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.848399] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.848449] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.217 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.848622] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.848655] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.217 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.848828] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.848860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.217 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.849046] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.849105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.217 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.849292] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.849337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.217 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.849494] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.849527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.217 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.849681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.849716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.217 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.849828] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.849859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.217 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.850031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.850068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.217 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.850260] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.850305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.217 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.850496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.850532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.217 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.850687] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.850724] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.217 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.850932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.850988] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.217 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.851150] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.851195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.217 qpair failed and we were unable to recover it. 00:38:45.217 [2024-12-06 17:54:26.851447] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.217 [2024-12-06 17:54:26.851489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.218 qpair failed and we were unable to recover it. 00:38:45.218 [2024-12-06 17:54:26.851604] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.218 [2024-12-06 17:54:26.851642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.218 qpair failed and we were unable to recover it. 00:38:45.218 [2024-12-06 17:54:26.851851] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.218 [2024-12-06 17:54:26.851889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.218 qpair failed and we were unable to recover it. 00:38:45.218 [2024-12-06 17:54:26.852038] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.218 [2024-12-06 17:54:26.852076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.218 qpair failed and we were unable to recover it. 00:38:45.218 [2024-12-06 17:54:26.852308] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.218 [2024-12-06 17:54:26.852366] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.218 qpair failed and we were unable to recover it. 00:38:45.218 [2024-12-06 17:54:26.852481] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.218 [2024-12-06 17:54:26.852514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.218 qpair failed and we were unable to recover it. 00:38:45.218 [2024-12-06 17:54:26.852677] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.218 [2024-12-06 17:54:26.852711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.218 qpair failed and we were unable to recover it. 00:38:45.218 [2024-12-06 17:54:26.852909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.218 [2024-12-06 17:54:26.852954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.218 qpair failed and we were unable to recover it. 00:38:45.218 [2024-12-06 17:54:26.853130] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.218 [2024-12-06 17:54:26.853190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.218 qpair failed and we were unable to recover it. 00:38:45.218 [2024-12-06 17:54:26.853408] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.218 [2024-12-06 17:54:26.853442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.218 qpair failed and we were unable to recover it. 00:38:45.218 [2024-12-06 17:54:26.853573] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.218 [2024-12-06 17:54:26.853629] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.218 qpair failed and we were unable to recover it. 00:38:45.218 [2024-12-06 17:54:26.853864] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.218 [2024-12-06 17:54:26.853934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.218 qpair failed and we were unable to recover it. 00:38:45.218 [2024-12-06 17:54:26.854118] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.218 [2024-12-06 17:54:26.854177] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.218 qpair failed and we were unable to recover it. 00:38:45.218 [2024-12-06 17:54:26.854379] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.218 [2024-12-06 17:54:26.854422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.218 qpair failed and we were unable to recover it. 00:38:45.218 [2024-12-06 17:54:26.854597] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.218 [2024-12-06 17:54:26.854653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.218 qpair failed and we were unable to recover it. 00:38:45.218 [2024-12-06 17:54:26.854806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.218 [2024-12-06 17:54:26.854840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.218 qpair failed and we were unable to recover it. 00:38:45.218 [2024-12-06 17:54:26.855007] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.218 [2024-12-06 17:54:26.855073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.218 qpair failed and we were unable to recover it. 00:38:45.218 [2024-12-06 17:54:26.855342] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.218 [2024-12-06 17:54:26.855397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.218 qpair failed and we were unable to recover it. 00:38:45.218 [2024-12-06 17:54:26.855588] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.218 [2024-12-06 17:54:26.855640] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.218 qpair failed and we were unable to recover it. 00:38:45.218 [2024-12-06 17:54:26.855838] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.218 [2024-12-06 17:54:26.855907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.218 qpair failed and we were unable to recover it. 00:38:45.218 [2024-12-06 17:54:26.856105] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.218 [2024-12-06 17:54:26.856160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.218 qpair failed and we were unable to recover it. 00:38:45.218 [2024-12-06 17:54:26.856360] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.218 [2024-12-06 17:54:26.856413] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.218 qpair failed and we were unable to recover it. 00:38:45.218 [2024-12-06 17:54:26.856621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.218 [2024-12-06 17:54:26.856692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.218 qpair failed and we were unable to recover it. 00:38:45.218 [2024-12-06 17:54:26.856845] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.218 [2024-12-06 17:54:26.856880] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.218 qpair failed and we were unable to recover it. 00:38:45.218 [2024-12-06 17:54:26.857042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.857081] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.219 qpair failed and we were unable to recover it. 00:38:45.219 [2024-12-06 17:54:26.857218] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.857257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.219 qpair failed and we were unable to recover it. 00:38:45.219 [2024-12-06 17:54:26.857457] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.857514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.219 qpair failed and we were unable to recover it. 00:38:45.219 [2024-12-06 17:54:26.857742] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.857793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.219 qpair failed and we were unable to recover it. 00:38:45.219 [2024-12-06 17:54:26.857966] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.858008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.219 qpair failed and we were unable to recover it. 00:38:45.219 [2024-12-06 17:54:26.858131] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.858171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.219 qpair failed and we were unable to recover it. 00:38:45.219 [2024-12-06 17:54:26.858300] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.858341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.219 qpair failed and we were unable to recover it. 00:38:45.219 [2024-12-06 17:54:26.858552] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.858591] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.219 qpair failed and we were unable to recover it. 00:38:45.219 [2024-12-06 17:54:26.858750] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.858786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.219 qpair failed and we were unable to recover it. 00:38:45.219 [2024-12-06 17:54:26.858891] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.858924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.219 qpair failed and we were unable to recover it. 00:38:45.219 [2024-12-06 17:54:26.859075] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.859109] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.219 qpair failed and we were unable to recover it. 00:38:45.219 [2024-12-06 17:54:26.859265] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.859318] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.219 qpair failed and we were unable to recover it. 00:38:45.219 [2024-12-06 17:54:26.859486] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.859524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.219 qpair failed and we were unable to recover it. 00:38:45.219 [2024-12-06 17:54:26.859655] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.859722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.219 qpair failed and we were unable to recover it. 00:38:45.219 [2024-12-06 17:54:26.859826] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.859860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.219 qpair failed and we were unable to recover it. 00:38:45.219 [2024-12-06 17:54:26.859996] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.860030] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.219 qpair failed and we were unable to recover it. 00:38:45.219 [2024-12-06 17:54:26.860189] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.860242] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.219 qpair failed and we were unable to recover it. 00:38:45.219 [2024-12-06 17:54:26.860386] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.860425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.219 qpair failed and we were unable to recover it. 00:38:45.219 [2024-12-06 17:54:26.860560] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.860598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.219 qpair failed and we were unable to recover it. 00:38:45.219 [2024-12-06 17:54:26.860727] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.860761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.219 qpair failed and we were unable to recover it. 00:38:45.219 [2024-12-06 17:54:26.860903] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.860937] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.219 qpair failed and we were unable to recover it. 00:38:45.219 [2024-12-06 17:54:26.861152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.861191] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.219 qpair failed and we were unable to recover it. 00:38:45.219 [2024-12-06 17:54:26.861296] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.861334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.219 qpair failed and we were unable to recover it. 00:38:45.219 [2024-12-06 17:54:26.861453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.861502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.219 qpair failed and we were unable to recover it. 00:38:45.219 [2024-12-06 17:54:26.861678] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.861712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.219 qpair failed and we were unable to recover it. 00:38:45.219 [2024-12-06 17:54:26.861814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.861848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.219 qpair failed and we were unable to recover it. 00:38:45.219 [2024-12-06 17:54:26.861975] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.862013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.219 qpair failed and we were unable to recover it. 00:38:45.219 [2024-12-06 17:54:26.862161] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.862199] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.219 qpair failed and we were unable to recover it. 00:38:45.219 [2024-12-06 17:54:26.862331] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.219 [2024-12-06 17:54:26.862364] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.862516] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.862556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.862751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.862813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.863062] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.863117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.863312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.863367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.863589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.863629] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.863835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.863869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.863995] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.864033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.864197] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.864235] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.864377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.864428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.864564] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.864602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.864778] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.864813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.864951] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.864992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.865144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.865182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.865339] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.865380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.865545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.865584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.865728] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.865763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.865902] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.865936] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.866082] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.866121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.866252] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.866307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.866424] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.866462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.866583] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.866635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.866786] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.866821] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.866931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.866965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.867145] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.867183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.867342] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.867382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.867545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.867578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.867691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.867725] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.867863] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.867897] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.868043] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.868082] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.220 [2024-12-06 17:54:26.868248] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.220 [2024-12-06 17:54:26.868287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.220 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.868435] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.868473] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.221 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.868636] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.868684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.221 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.868850] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.868884] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.221 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.869022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.869056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.221 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.869240] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.869301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.221 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.869432] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.869470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.221 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.869680] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.869714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.221 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.869829] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.869863] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.221 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.869980] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.870018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.221 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.870214] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.870252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.221 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.870358] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.870397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.221 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.870556] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.870594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.221 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.870741] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.870775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.221 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.870885] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.870919] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.221 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.871032] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.871065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.221 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.871283] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.871328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.221 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.871564] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.871603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.221 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.871816] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.871849] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.221 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.871966] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.872000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.221 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.872145] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.872198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.221 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.872342] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.872381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.221 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.872533] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.872572] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.221 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.872735] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.872769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.221 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.872909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.872942] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.221 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.873062] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.873101] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.221 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.873223] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.873261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.221 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.873420] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.873453] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.221 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.873629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.873676] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.221 qpair failed and we were unable to recover it. 00:38:45.221 [2024-12-06 17:54:26.873835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.221 [2024-12-06 17:54:26.873868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.874030] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.874077] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.874199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.874238] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.874453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.874490] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.874607] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.874646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.874821] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.874855] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.875020] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.875058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.875198] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.875237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.875380] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.875419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.875540] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.875580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.875758] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.875792] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.875905] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.875948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.876115] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.876154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.876341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.876379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.876535] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.876572] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.876723] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.876758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.876897] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.876930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.877118] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.877152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.877296] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.877335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.877543] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.877576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.877722] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.877755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.877897] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.877931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.878073] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.878107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.878283] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.878316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.878485] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.878523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.878651] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.878713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.878859] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.878893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.879031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.879069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.879237] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.879276] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.879398] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.879436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.879638] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.222 [2024-12-06 17:54:26.879688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.222 qpair failed and we were unable to recover it. 00:38:45.222 [2024-12-06 17:54:26.879805] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.879838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.879958] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.879992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.880157] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.880196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.880351] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.880390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.880547] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.880585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.880760] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.880795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.880931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.880964] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.881106] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.881144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.881358] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.881396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.881586] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.881624] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.881810] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.881844] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.881999] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.882037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.882164] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.882204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.882387] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.882425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.882588] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.882626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.882826] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.882878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.883022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.883065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.883274] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.883314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.883435] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.883476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.883683] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.883736] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.883906] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.883958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.884175] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.884224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.884415] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.884463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.884622] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.884705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.884826] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.884861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.885036] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.885084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.885353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.885402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.885627] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.885678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.885844] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.223 [2024-12-06 17:54:26.885878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.223 qpair failed and we were unable to recover it. 00:38:45.223 [2024-12-06 17:54:26.885993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.224 [2024-12-06 17:54:26.886042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.224 qpair failed and we were unable to recover it. 00:38:45.224 [2024-12-06 17:54:26.886230] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.224 [2024-12-06 17:54:26.886270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.224 qpair failed and we were unable to recover it. 00:38:45.224 [2024-12-06 17:54:26.886424] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.224 [2024-12-06 17:54:26.886464] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.224 qpair failed and we were unable to recover it. 00:38:45.224 [2024-12-06 17:54:26.886592] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.224 [2024-12-06 17:54:26.886631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.224 qpair failed and we were unable to recover it. 00:38:45.224 [2024-12-06 17:54:26.886818] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.224 [2024-12-06 17:54:26.886853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.224 qpair failed and we were unable to recover it. 00:38:45.224 [2024-12-06 17:54:26.887022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.224 [2024-12-06 17:54:26.887080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.224 qpair failed and we were unable to recover it. 00:38:45.224 [2024-12-06 17:54:26.887310] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.224 [2024-12-06 17:54:26.887360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.224 qpair failed and we were unable to recover it. 00:38:45.224 [2024-12-06 17:54:26.887612] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.224 [2024-12-06 17:54:26.887711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.224 qpair failed and we were unable to recover it. 00:38:45.224 [2024-12-06 17:54:26.887857] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.224 [2024-12-06 17:54:26.887892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.224 qpair failed and we were unable to recover it. 00:38:45.224 [2024-12-06 17:54:26.888113] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.224 [2024-12-06 17:54:26.888162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.224 qpair failed and we were unable to recover it. 00:38:45.224 [2024-12-06 17:54:26.888357] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.224 [2024-12-06 17:54:26.888405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.224 qpair failed and we were unable to recover it. 00:38:45.224 [2024-12-06 17:54:26.888681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.224 [2024-12-06 17:54:26.888742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.224 qpair failed and we were unable to recover it. 00:38:45.224 [2024-12-06 17:54:26.888883] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.224 [2024-12-06 17:54:26.888917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.224 qpair failed and we were unable to recover it. 00:38:45.224 [2024-12-06 17:54:26.889104] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.224 [2024-12-06 17:54:26.889138] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.224 qpair failed and we were unable to recover it. 00:38:45.224 [2024-12-06 17:54:26.889284] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.224 [2024-12-06 17:54:26.889320] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.224 qpair failed and we were unable to recover it. 00:38:45.224 [2024-12-06 17:54:26.889529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.224 [2024-12-06 17:54:26.889578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.224 qpair failed and we were unable to recover it. 00:38:45.224 [2024-12-06 17:54:26.889798] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.224 [2024-12-06 17:54:26.889833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.224 qpair failed and we were unable to recover it. 00:38:45.224 [2024-12-06 17:54:26.889950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.224 [2024-12-06 17:54:26.889984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.224 qpair failed and we were unable to recover it. 00:38:45.224 [2024-12-06 17:54:26.890124] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.224 [2024-12-06 17:54:26.890158] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.224 qpair failed and we were unable to recover it. 00:38:45.224 [2024-12-06 17:54:26.890301] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.224 [2024-12-06 17:54:26.890350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.224 qpair failed and we were unable to recover it. 00:38:45.224 [2024-12-06 17:54:26.890497] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.224 [2024-12-06 17:54:26.890545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.224 qpair failed and we were unable to recover it. 00:38:45.224 [2024-12-06 17:54:26.890779] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.224 [2024-12-06 17:54:26.890830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.224 qpair failed and we were unable to recover it. 00:38:45.224 [2024-12-06 17:54:26.891066] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.224 [2024-12-06 17:54:26.891116] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.224 qpair failed and we were unable to recover it. 00:38:45.224 [2024-12-06 17:54:26.891308] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.224 [2024-12-06 17:54:26.891357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.224 qpair failed and we were unable to recover it. 00:38:45.224 [2024-12-06 17:54:26.891608] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.224 [2024-12-06 17:54:26.891676] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.891858] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.225 [2024-12-06 17:54:26.891909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.892109] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.225 [2024-12-06 17:54:26.892159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.892356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.225 [2024-12-06 17:54:26.892405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.892623] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.225 [2024-12-06 17:54:26.892719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.892887] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.225 [2024-12-06 17:54:26.892936] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.893164] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.225 [2024-12-06 17:54:26.893213] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.893361] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.225 [2024-12-06 17:54:26.893412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.893581] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.225 [2024-12-06 17:54:26.893631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.893849] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.225 [2024-12-06 17:54:26.893899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.894094] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.225 [2024-12-06 17:54:26.894143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.894332] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.225 [2024-12-06 17:54:26.894381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.894582] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.225 [2024-12-06 17:54:26.894632] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.894856] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.225 [2024-12-06 17:54:26.894909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.895123] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.225 [2024-12-06 17:54:26.895175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.895362] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.225 [2024-12-06 17:54:26.895415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.895615] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.225 [2024-12-06 17:54:26.895680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.895924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.225 [2024-12-06 17:54:26.895977] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.896148] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.225 [2024-12-06 17:54:26.896199] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.896385] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.225 [2024-12-06 17:54:26.896434] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.896707] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.225 [2024-12-06 17:54:26.896775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.896989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.225 [2024-12-06 17:54:26.897090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.897336] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.225 [2024-12-06 17:54:26.897388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.897606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.225 [2024-12-06 17:54:26.897659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.897905] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.225 [2024-12-06 17:54:26.897958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.898125] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.225 [2024-12-06 17:54:26.898173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.898409] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.225 [2024-12-06 17:54:26.898443] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.898583] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.225 [2024-12-06 17:54:26.898617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.898779] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.225 [2024-12-06 17:54:26.898830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.225 qpair failed and we were unable to recover it. 00:38:45.225 [2024-12-06 17:54:26.899057] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.899106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.899339] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.899391] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.899605] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.899714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.899971] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.900029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.900232] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.900284] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.900512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.900552] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.900691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.900732] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.900889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.900928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.901095] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.901148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.901390] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.901443] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.901708] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.901748] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.901883] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.901923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.902108] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.902161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.902325] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.902379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.902584] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.902635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.902847] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.902900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.903101] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.903154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.903399] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.903451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.903650] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.903712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.903964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.904017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.904249] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.904301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.904543] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.904595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.904815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.904856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.904989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.905028] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.905241] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.905293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.905508] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.905547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.905710] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.905752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.905909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.905948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.906210] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.906262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.906431] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.906483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.226 [2024-12-06 17:54:26.906686] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.226 [2024-12-06 17:54:26.906740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.226 qpair failed and we were unable to recover it. 00:38:45.227 [2024-12-06 17:54:26.906941] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.227 [2024-12-06 17:54:26.906992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.227 qpair failed and we were unable to recover it. 00:38:45.227 [2024-12-06 17:54:26.907199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.227 [2024-12-06 17:54:26.907260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.227 qpair failed and we were unable to recover it. 00:38:45.227 [2024-12-06 17:54:26.907461] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.227 [2024-12-06 17:54:26.907514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.227 qpair failed and we were unable to recover it. 00:38:45.227 [2024-12-06 17:54:26.907693] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.227 [2024-12-06 17:54:26.907746] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.227 qpair failed and we were unable to recover it. 00:38:45.227 [2024-12-06 17:54:26.907923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.227 [2024-12-06 17:54:26.907977] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.227 qpair failed and we were unable to recover it. 00:38:45.227 [2024-12-06 17:54:26.908224] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.227 [2024-12-06 17:54:26.908277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.227 qpair failed and we were unable to recover it. 00:38:45.227 [2024-12-06 17:54:26.908479] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.227 [2024-12-06 17:54:26.908531] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.227 qpair failed and we were unable to recover it. 00:38:45.227 [2024-12-06 17:54:26.908715] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.227 [2024-12-06 17:54:26.908768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.227 qpair failed and we were unable to recover it. 00:38:45.227 [2024-12-06 17:54:26.908962] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.227 [2024-12-06 17:54:26.909014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.227 qpair failed and we were unable to recover it. 00:38:45.227 [2024-12-06 17:54:26.909239] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.227 [2024-12-06 17:54:26.909293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.227 qpair failed and we were unable to recover it. 00:38:45.227 [2024-12-06 17:54:26.909548] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.227 [2024-12-06 17:54:26.909601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.227 qpair failed and we were unable to recover it. 00:38:45.227 [2024-12-06 17:54:26.909806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.227 [2024-12-06 17:54:26.909859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.227 qpair failed and we were unable to recover it. 00:38:45.227 [2024-12-06 17:54:26.910076] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.227 [2024-12-06 17:54:26.910130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.227 qpair failed and we were unable to recover it. 00:38:45.227 [2024-12-06 17:54:26.910384] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.227 [2024-12-06 17:54:26.910438] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.227 qpair failed and we were unable to recover it. 00:38:45.227 [2024-12-06 17:54:26.910721] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.227 [2024-12-06 17:54:26.910776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.227 qpair failed and we were unable to recover it. 00:38:45.227 [2024-12-06 17:54:26.910980] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.227 [2024-12-06 17:54:26.911033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.227 qpair failed and we were unable to recover it. 00:38:45.512 [2024-12-06 17:54:26.911229] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.512 [2024-12-06 17:54:26.911282] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.512 qpair failed and we were unable to recover it. 00:38:45.512 [2024-12-06 17:54:26.911511] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.512 [2024-12-06 17:54:26.911550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.512 qpair failed and we were unable to recover it. 00:38:45.512 [2024-12-06 17:54:26.911685] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.512 [2024-12-06 17:54:26.911724] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.512 qpair failed and we were unable to recover it. 00:38:45.512 [2024-12-06 17:54:26.911891] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.512 [2024-12-06 17:54:26.911931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.512 qpair failed and we were unable to recover it. 00:38:45.512 [2024-12-06 17:54:26.912074] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.512 [2024-12-06 17:54:26.912126] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.512 qpair failed and we were unable to recover it. 00:38:45.512 [2024-12-06 17:54:26.912326] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.512 [2024-12-06 17:54:26.912378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.512 qpair failed and we were unable to recover it. 00:38:45.512 [2024-12-06 17:54:26.912525] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.512 [2024-12-06 17:54:26.912577] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.512 qpair failed and we were unable to recover it. 00:38:45.512 [2024-12-06 17:54:26.912812] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.512 [2024-12-06 17:54:26.912852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.512 qpair failed and we were unable to recover it. 00:38:45.512 [2024-12-06 17:54:26.912989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.512 [2024-12-06 17:54:26.913029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.512 qpair failed and we were unable to recover it. 00:38:45.512 [2024-12-06 17:54:26.913248] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.512 [2024-12-06 17:54:26.913300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.512 qpair failed and we were unable to recover it. 00:38:45.512 [2024-12-06 17:54:26.913460] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.512 [2024-12-06 17:54:26.913499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.512 qpair failed and we were unable to recover it. 00:38:45.512 [2024-12-06 17:54:26.913661] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.512 [2024-12-06 17:54:26.913707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.512 qpair failed and we were unable to recover it. 00:38:45.512 [2024-12-06 17:54:26.913888] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.512 [2024-12-06 17:54:26.913945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.512 qpair failed and we were unable to recover it. 00:38:45.512 [2024-12-06 17:54:26.914082] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.512 [2024-12-06 17:54:26.914122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.512 qpair failed and we were unable to recover it. 00:38:45.512 [2024-12-06 17:54:26.914283] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.512 [2024-12-06 17:54:26.914322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.512 qpair failed and we were unable to recover it. 00:38:45.512 [2024-12-06 17:54:26.914481] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.512 [2024-12-06 17:54:26.914519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.512 qpair failed and we were unable to recover it. 00:38:45.512 [2024-12-06 17:54:26.914633] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.512 [2024-12-06 17:54:26.914684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.512 qpair failed and we were unable to recover it. 00:38:45.512 [2024-12-06 17:54:26.914821] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.512 [2024-12-06 17:54:26.914860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.512 qpair failed and we were unable to recover it. 00:38:45.512 [2024-12-06 17:54:26.914989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.512 [2024-12-06 17:54:26.915027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.512 qpair failed and we were unable to recover it. 00:38:45.512 [2024-12-06 17:54:26.915155] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.512 [2024-12-06 17:54:26.915193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.512 qpair failed and we were unable to recover it. 00:38:45.512 [2024-12-06 17:54:26.915321] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.512 [2024-12-06 17:54:26.915359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.512 qpair failed and we were unable to recover it. 00:38:45.512 [2024-12-06 17:54:26.915515] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.512 [2024-12-06 17:54:26.915557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.512 qpair failed and we were unable to recover it. 00:38:45.512 [2024-12-06 17:54:26.915718] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.512 [2024-12-06 17:54:26.915759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.512 qpair failed and we were unable to recover it. 00:38:45.512 [2024-12-06 17:54:26.915956] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.512 [2024-12-06 17:54:26.915995] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.512 qpair failed and we were unable to recover it. 00:38:45.512 [2024-12-06 17:54:26.916118] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.512 [2024-12-06 17:54:26.916158] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.512 qpair failed and we were unable to recover it. 00:38:45.512 [2024-12-06 17:54:26.916316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.512 [2024-12-06 17:54:26.916355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.512 qpair failed and we were unable to recover it. 00:38:45.512 [2024-12-06 17:54:26.916552] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.512 [2024-12-06 17:54:26.916592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.512 qpair failed and we were unable to recover it. 00:38:45.512 [2024-12-06 17:54:26.916736] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.916778] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.916935] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.917015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.917318] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.917384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.917628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.917725] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.917982] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.918047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.918238] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.918293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.918512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.918552] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.918715] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.918755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.918870] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.918909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.919084] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.919136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.919360] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.919416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.919601] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.919657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.919866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.919923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.920141] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.920198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.920418] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.920474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.920686] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.920727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.920930] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.920986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.921153] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.921209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.921431] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.921488] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.921700] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.921758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.921970] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.922026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.922275] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.922332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.922627] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.922700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.922886] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.922944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.923149] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.923206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.923420] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.923485] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.923727] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.923786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.923989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.924044] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.924301] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.924357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.924562] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.924618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.924857] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.924913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.925086] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.925142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.925361] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.925419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.925650] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.925718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.925932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.925988] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.926243] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.926301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.926564] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.926620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.926853] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.513 [2024-12-06 17:54:26.926909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.513 qpair failed and we were unable to recover it. 00:38:45.513 [2024-12-06 17:54:26.927087] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.927143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.927372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.927428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.927645] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.927720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.927942] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.927999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.928251] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.928307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.928566] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.928622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.928902] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.928958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.929125] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.929181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.929432] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.929489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.929654] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.929728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.929945] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.930003] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.930268] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.930325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.930517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.930583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.930827] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.930885] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.931107] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.931164] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.931380] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.931437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.931695] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.931754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.932018] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.932074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.932343] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.932398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.932626] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.932699] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.932948] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.933004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.933259] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.933315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.933541] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.933596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.933876] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.933932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.934142] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.934204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.934572] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.934638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.934911] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.934971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.935188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.935258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.935522] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.935582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.935843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.935901] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.936085] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.936142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.936393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.936450] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.936683] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.936760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.936943] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.937006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.937229] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.937291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.937521] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.937582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.937861] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.937923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.514 [2024-12-06 17:54:26.938127] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.514 [2024-12-06 17:54:26.938187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.514 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.938404] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.938464] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.938704] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.938765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.939038] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.939099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.939391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.939451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.939689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.939750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.939994] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.940055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.940275] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.940336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.940558] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.940617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.940858] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.940918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.941181] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.941241] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.941525] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.941585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.941835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.941896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.942129] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.942190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.942432] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.942492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.942754] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.942815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.943064] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.943135] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.943431] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.943491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.943725] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.943787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.944041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.944099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.944331] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.944387] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.944676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.944738] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.945009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.945069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.945300] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.945360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.945646] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.945738] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.945990] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.946046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.946304] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.946359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.946600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.946659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.946964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.947025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.947211] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.947271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.947505] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.947574] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.947789] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.947853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.948086] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.948149] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.948353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.948414] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.948662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.948739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.948944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.949005] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.949233] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.949294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.949562] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.949623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.949877] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.515 [2024-12-06 17:54:26.949938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.515 qpair failed and we were unable to recover it. 00:38:45.515 [2024-12-06 17:54:26.950164] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.950224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.950468] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.950528] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.950732] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.950793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.951016] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.951078] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.951338] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.951399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.951691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.951752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.951985] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.952046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.952308] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.952369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.952652] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.952726] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.952959] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.953035] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.953312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.953378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.953627] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.953729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.953986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.954053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.954250] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.954316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.954565] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.954631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.954959] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.955025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.955325] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.955390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.955654] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.955738] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.956000] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.956068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.956323] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.956388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.956699] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.956766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.957063] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.957129] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.957420] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.957485] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.957785] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.957853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.958114] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.958179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.958433] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.958503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.958751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.958819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.959086] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.959151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.959375] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.959441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.959731] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.959798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.960101] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.516 [2024-12-06 17:54:26.960166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.516 qpair failed and we were unable to recover it. 00:38:45.516 [2024-12-06 17:54:26.960430] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.960508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.960788] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.960856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.961224] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.961288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.961583] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.961648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.961938] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.962003] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.962217] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.962282] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.962485] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.962553] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.962820] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.962888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.963186] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.963252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.963535] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.963602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.963856] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.963923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.964174] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.964242] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.964491] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.964555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.964819] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.964885] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.965188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.965253] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.965556] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.965621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.965870] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.965935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.966195] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.966261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.966554] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.966618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.966931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.966997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.967218] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.967283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.967526] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.967590] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.967854] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.967922] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.968215] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.968283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.968485] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.968550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.968862] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.968930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.969185] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.969251] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.969514] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.969581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.969869] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.969936] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.970197] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.970262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.970551] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.970617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.970931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.970996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.971295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.971360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.971684] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.971741] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.971969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.972027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.972287] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.972351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.972610] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.972695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.972905] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.517 [2024-12-06 17:54:26.972971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.517 qpair failed and we were unable to recover it. 00:38:45.517 [2024-12-06 17:54:26.973278] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.973333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.973614] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.973697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.973990] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.974065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.974363] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.974427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.974715] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.974782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.974997] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.975063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.975323] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.975388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.975631] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.975710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.975910] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.975976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.976270] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.976335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.976593] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.976660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.976930] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.976995] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.977246] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.977311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.977610] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.977700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.978000] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.978066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.978368] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.978433] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.978733] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.978799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.979119] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.979184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.979448] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.979513] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.979780] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.979848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.980144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.980209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.980513] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.980578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.980824] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.980890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.981143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.981209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.981497] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.981561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.981815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.981882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.982083] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.982148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.982388] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.982452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.982739] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.982831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.983095] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.983162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.983423] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.983488] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.983775] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.983842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.984132] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.984198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.984448] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.984514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.984803] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.984860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.985089] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.985146] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.985404] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.985472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.985714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.518 [2024-12-06 17:54:26.985781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.518 qpair failed and we were unable to recover it. 00:38:45.518 [2024-12-06 17:54:26.986070] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.986136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.986437] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.986502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.986746] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.986828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.987094] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.987160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.987406] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.987481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.987724] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.987790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.988009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.988077] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.988359] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.988424] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.988720] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.988787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.989079] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.989145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.989392] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.989457] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.989662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.989762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.989976] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.990041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.990234] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.990299] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.990587] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.990653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.990951] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.991016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.991303] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.991368] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.991565] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.991629] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.991917] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.991984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.992237] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.992304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.992598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.992681] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.992943] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.993010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.993304] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.993369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.993595] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.993662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.993949] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.994015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.994261] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.994327] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.994616] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.994700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.994965] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.995031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.995312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.995378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.995662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.995745] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.996045] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.996111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.996362] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.996427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.996697] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.996763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.997056] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.997120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.997368] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.997433] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.997628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.997728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.998029] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.998093] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.998350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.519 [2024-12-06 17:54:26.998415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.519 qpair failed and we were unable to recover it. 00:38:45.519 [2024-12-06 17:54:26.998656] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:26.998740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:26.999004] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:26.999068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:26.999322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:26.999388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:26.999632] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:26.999715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:26.999972] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.000037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.000279] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.000346] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.000550] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.000627] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.000914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.000979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.001172] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.001238] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.001530] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.001596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.001854] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.001922] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.002138] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.002204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.002413] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.002478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.002703] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.002770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.003071] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.003136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.003354] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.003421] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.003694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.003760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.004058] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.004123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.004410] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.004476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.004694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.004761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.005030] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.005096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.005353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.005418] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.005612] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.005706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.005955] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.006021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.006263] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.006330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.006631] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.006717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.007020] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.007086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.007291] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.007357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.007642] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.007727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.008035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.008101] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.008322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.008388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.008682] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.008749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.009041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.009107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.009383] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.009451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.009758] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.009826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.010136] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.010202] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.010459] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.010526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.010817] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.520 [2024-12-06 17:54:27.010884] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.520 qpair failed and we were unable to recover it. 00:38:45.520 [2024-12-06 17:54:27.011179] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.011245] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.011542] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.011609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.011900] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.011966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.012163] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.012228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.012473] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.012539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.012833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.012900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.013200] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.013266] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.013505] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.013571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.013847] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.013925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.014220] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.014285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.014579] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.014645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.014958] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.015023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.015280] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.015348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.015628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.015713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.015961] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.016026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.016231] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.016300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.016555] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.016620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.016930] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.016996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.017248] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.017315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.017560] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.017627] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.017925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.017991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.018280] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.018346] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.018660] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.018744] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.018952] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.019021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.019311] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.019377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.019693] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.019762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.020006] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.020072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.020334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.020399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.020696] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.020763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.020973] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.021039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.021326] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.021391] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.021648] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.021734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.022042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.022108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.521 qpair failed and we were unable to recover it. 00:38:45.521 [2024-12-06 17:54:27.022350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.521 [2024-12-06 17:54:27.022415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.022681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.022749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.023061] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.023127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.023373] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.023437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.023718] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.023786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.024098] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.024163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.024419] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.024484] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.024739] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.024806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.024997] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.025065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.025317] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.025385] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.025633] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.025712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.025967] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.026036] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.026235] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.026302] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.026604] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.026683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.026909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.026974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.027185] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.027253] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.027507] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.027573] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.027889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.027957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.028158] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.028228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.028481] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.028547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.028791] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.028859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.029150] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.029216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.029407] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.029474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.029767] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.029835] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.030093] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.030159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.030425] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.030489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.030796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.030863] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.031117] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.031183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.031472] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.031537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.031860] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.031928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.032182] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.032247] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.032536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.032601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.032892] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.032958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.033223] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.033288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.033581] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.033646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.522 [2024-12-06 17:54:27.033942] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.522 [2024-12-06 17:54:27.034009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.522 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.034263] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.034329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.034578] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.034643] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.034918] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.034984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.035237] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.035302] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.035560] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.035625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.035936] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.036002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.036264] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.036340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.036643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.036730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.037031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.037097] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.037295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.037362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.037543] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.037609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.037914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.037981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.038231] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.038296] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.038511] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.038579] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.038892] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.038960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.039261] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.039326] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.039630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.039729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.039991] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.040057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.040303] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.040369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.040629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.040714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.040947] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.041013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.041258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.041323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.041570] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.041635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.041969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.042035] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.042238] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.042308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.042557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.042622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.042890] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.042955] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.043196] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.043261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.043513] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.043580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.043843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.043910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.044158] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.044224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.044467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.044533] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.044729] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.044797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.045014] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.045080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.045350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.045416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.045677] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.045744] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.523 qpair failed and we were unable to recover it. 00:38:45.523 [2024-12-06 17:54:27.046054] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.523 [2024-12-06 17:54:27.046119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.046366] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.046431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.046654] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.046739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.046993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.047058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.047309] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.047377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.047645] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.047744] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.048039] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.048104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.048364] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.048430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.048728] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.048796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.049094] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.049159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.049452] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.049527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.049793] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.049860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.050076] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.050141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.050390] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.050455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.050778] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.050859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.051137] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.051193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.051427] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.051483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.051770] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.051824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.052059] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.052113] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.052289] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.052340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.052546] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.052600] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.052864] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.052928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.053181] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.053247] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.053495] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.053559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.053809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.053876] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.054159] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.054225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.054436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.054501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.054801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.054868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.055110] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.055176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.055481] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.055547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.055811] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.055879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.056172] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.056238] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.056455] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.056520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.056812] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.056878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.057125] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.057192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.057414] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.057479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.057727] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.057794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.058055] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.524 [2024-12-06 17:54:27.058121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.524 qpair failed and we were unable to recover it. 00:38:45.524 [2024-12-06 17:54:27.058385] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.058452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.058723] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.058790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.059042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.059112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.059401] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.059467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.059770] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.059837] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.060123] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.060188] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.060393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.060461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.060720] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.060787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.061018] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.061085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.061328] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.061396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.061654] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.061737] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.061981] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.062046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.062219] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.062295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.062598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.062684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.062945] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.063010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.063306] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.063371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.063621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.063720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.063928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.063994] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.064226] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.064292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.064585] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.064650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.064928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.064994] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.065226] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.065292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.065513] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.065580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.065900] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.065966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.066273] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.066339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.066583] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.066647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.066978] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.067045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.067320] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.067386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.067718] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.067784] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.068016] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.068082] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.068296] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.068362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.068578] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.068643] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.068893] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.068958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.069173] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.069239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.069536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.069604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.069885] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.069952] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.070216] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.070282] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.070539] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.525 [2024-12-06 17:54:27.070605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.525 qpair failed and we were unable to recover it. 00:38:45.525 [2024-12-06 17:54:27.070878] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.070944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.071207] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.071274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.071519] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.071585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.071837] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.071905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.072150] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.072216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.072511] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.072576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.072901] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.072968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.073174] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.073241] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.073493] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.073558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.073836] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.073903] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.074207] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.074272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.074570] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.074636] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.074887] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.074952] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.075243] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.075309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.075560] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.075636] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.075905] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.075970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.076187] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.076253] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.076511] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.076578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.076813] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.076881] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.077145] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.077210] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.077505] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.077571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.077880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.077947] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.078236] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.078301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.078546] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.078613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.078851] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.078917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.079210] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.079276] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.079535] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.079601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.079896] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.079967] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.080246] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.080313] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.080573] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.080639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.080921] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.080986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.081228] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.081293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.081543] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.526 [2024-12-06 17:54:27.081609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.526 qpair failed and we were unable to recover it. 00:38:45.526 [2024-12-06 17:54:27.081937] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.082003] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.082294] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.082361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.082645] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.082731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.082985] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.083050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.083353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.083418] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.083712] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.083780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.084069] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.084134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.084350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.084415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.084646] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.084730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.085020] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.085087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.085334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.085399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.085694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.085762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.086054] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.086119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.086370] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.086438] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.086692] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.086778] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.087036] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.087101] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.087295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.087361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.087564] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.087630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.087921] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.087987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.088277] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.088343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.088598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.088662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.088943] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.089020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.089269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.089335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.089535] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.089602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.089870] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.089937] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.090175] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.090240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.090500] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.090566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.090833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.090902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.091193] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.091258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.091543] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.091609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.091932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.091998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.092288] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.092353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.092600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.092684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.092993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.093059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.093293] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.093358] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.093598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.093685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.093949] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.094018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.094309] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.527 [2024-12-06 17:54:27.094375] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.527 qpair failed and we were unable to recover it. 00:38:45.527 [2024-12-06 17:54:27.094622] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.094708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.095005] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.095072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.095323] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.095388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.095635] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.095742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.096038] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.096104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.096398] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.096463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.096748] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.096816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.097063] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.097129] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.097344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.097410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.097616] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.097701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.097957] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.098027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.098228] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.098293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.098532] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.098597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.098811] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.098878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.099138] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.099206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.099499] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.099565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.099858] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.099929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.100178] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.100244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.100494] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.100561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.100779] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.100847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.101102] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.101168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.101429] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.101495] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.101797] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.101863] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.102131] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.102207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.102459] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.102528] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.102774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.102841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.103100] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.103165] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.103403] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.103469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.103714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.103780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.104034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.104100] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.104389] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.104455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.104722] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.104789] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.105096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.105162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.105376] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.105442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.105744] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.105811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.106098] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.106164] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.106411] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.106478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.106723] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.528 [2024-12-06 17:54:27.106791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.528 qpair failed and we were unable to recover it. 00:38:45.528 [2024-12-06 17:54:27.107089] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.107156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.107415] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.107480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.107720] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.107788] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.108001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.108067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.108331] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.108398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.108646] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.108726] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.109038] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.109103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.109375] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.109445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.109742] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.109810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.110020] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.110085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.110349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.110416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.110727] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.110795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.111061] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.111127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.111427] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.111493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.111781] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.111848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.112149] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.112214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.112476] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.112541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.112807] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.112874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.113170] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.113237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.113483] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.113549] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.113785] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.113853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.114074] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.114143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.114453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.114518] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.114820] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.114888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.115156] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.115223] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.115472] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.115547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.115757] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.115825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.116082] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.116147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.116355] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.116424] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.116636] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.116719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.116996] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.117063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.117256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.117324] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.117571] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.117639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.117920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.117987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.118281] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.118349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.118602] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.118686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.118904] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.118970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.119272] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.529 [2024-12-06 17:54:27.119337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.529 qpair failed and we were unable to recover it. 00:38:45.529 [2024-12-06 17:54:27.119630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.119723] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.120018] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.120084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.120343] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.120408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.120708] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.120776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.121065] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.121130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.121376] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.121441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.121702] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.121770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.122056] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.122123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.122420] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.122486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.122750] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.122818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.123024] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.123093] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.123342] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.123407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.123613] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.123709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.123967] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.124032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.124324] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.124389] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.124638] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.124728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.124934] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.124998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.125198] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.125262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.125442] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.125509] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.125754] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.125820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.126066] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.126132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.126393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.126458] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.126713] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.126781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.127036] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.127101] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.127344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.127409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.127628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.127730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.128032] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.128097] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.128339] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.128415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.128691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.128760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.129008] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.129076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.129276] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.129342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.129606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.129693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.129957] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.130023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.530 [2024-12-06 17:54:27.130314] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.530 [2024-12-06 17:54:27.130379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.530 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.130624] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.130711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.130930] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.130996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.131257] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.131323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.131611] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.131697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.131951] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.132016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.132308] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.132373] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.132688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.132754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.133007] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.133073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.133332] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.133398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.133695] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.133762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.134012] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.134077] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.134366] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.134432] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.134695] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.134761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.135017] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.135082] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.135334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.135401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.135606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.135711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.135974] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.136040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.136252] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.136318] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.136523] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.136587] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.136904] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.136971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.137233] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.137331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.137598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.137690] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.137908] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.137974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.138234] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.138304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.138565] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.138630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.138890] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.138954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.139256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.139321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.139626] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.139712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.139979] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.140042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.140266] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.140329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.140583] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.140647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.140960] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.141030] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.141344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.141409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.141700] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.141780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.142033] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.142097] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.142336] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.142402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.142681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.142748] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.531 [2024-12-06 17:54:27.142950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.531 [2024-12-06 17:54:27.143016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.531 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.143296] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.143359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.143603] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.143685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.143964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.144029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.144287] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.144352] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.144565] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.144628] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.144885] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.144950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.145199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.145272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.145524] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.145594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.145906] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.145972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.146225] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.146290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.146537] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.146603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.146898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.146965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.147206] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.147272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.147518] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.147581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.147824] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.147889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.148188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.148260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.148514] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.148578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.148882] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.148947] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.149168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.149235] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.149483] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.149546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.149799] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.149866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.150122] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.150187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.150526] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.150622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.150920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.150990] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.151225] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.151314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.151637] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.151748] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.152107] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.152193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.152555] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.152645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.152933] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.153023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.153374] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.153467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.153807] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.153878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.154183] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.154249] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.154540] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.154604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.154880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.154948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.155256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.155346] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.155680] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.155785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.156105] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.156184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.532 [2024-12-06 17:54:27.156545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.532 [2024-12-06 17:54:27.156636] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.532 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.156971] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.157059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.157326] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.157418] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.157661] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.157750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.157965] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.158031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.158247] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.158312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.158599] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.158688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.158917] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.159004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.159323] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.159412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.159716] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.159805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.160133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.160214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.160575] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.160662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.161052] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.161138] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.161429] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.161499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.161809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.161877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.162187] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.162252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.162436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.162501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.162794] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.162871] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.163202] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.163291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.163643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.163755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.164077] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.164165] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.164513] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.164602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.164957] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.165049] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.165395] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.165464] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.165720] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.165787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.166056] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.166124] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.166394] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.166458] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.166659] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.166761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.167073] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.167161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.167521] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.167606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.167971] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.168061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.168424] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.168513] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.168890] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.168964] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.169236] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.169301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.169587] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.169653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.169934] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.169999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.170286] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.170350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.170621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.170748] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.533 [2024-12-06 17:54:27.171103] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.533 [2024-12-06 17:54:27.171205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.533 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.171565] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.171648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.172026] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.172113] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.172473] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.172560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.172900] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.172972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.173268] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.173333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.173564] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.173649] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.173965] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.174033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.174289] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.174354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.174608] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.174695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.174953] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.175032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.175382] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.175470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.175806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.175899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.176243] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.176331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.176655] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.176768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.177126] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.177215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.177517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.177586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.177961] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.178031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.178257] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.178325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.178609] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.178691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.178994] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.179065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.179383] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.179474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.179735] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.179823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.180147] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.180236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.180541] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.180629] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.180999] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.181088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.181448] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.181543] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.181889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.181959] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.182249] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.182314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.182600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.182682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.182912] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.182977] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.183270] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.183359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.183688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.183780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.184127] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.184216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.184565] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.184654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.184988] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.185079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.185430] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.185518] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.185898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.185971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.534 [2024-12-06 17:54:27.186226] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.534 [2024-12-06 17:54:27.186293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.534 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.186586] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.186651] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.186943] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.187021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.187316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.187379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.187629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.187739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.188107] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.188194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.188506] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.188582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.188955] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.189044] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.189366] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.189455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.189745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.189839] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.190191] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.190260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.190551] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.190615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.190884] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.190949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.191184] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.191248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.191479] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.191568] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.191860] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.191951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.192268] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.192358] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.192716] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.192808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.193155] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.193243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.193566] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.193653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.194031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.194102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.194402] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.194467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.194700] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.194768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.194982] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.195050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.195310] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.195374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.195698] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.195791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.196094] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.196182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.196472] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.196557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.196896] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.196988] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.197350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.197447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.197763] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.197833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.198119] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.198185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.198459] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.198524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.198771] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.198836] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.535 [2024-12-06 17:54:27.199091] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.535 [2024-12-06 17:54:27.199159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.535 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.199447] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.199520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.199774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.199841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.200029] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.200095] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.200348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.200413] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.200621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.200707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.200959] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.201024] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.201327] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.201392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.201617] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.201700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.201987] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.202057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.202335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.202399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.202652] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.202734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.202989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.203052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.203301] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.203364] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.203631] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.203715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.203994] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.204059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.204318] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.204381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.204627] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.204710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.204913] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.204986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.205271] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.205335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.205586] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.205651] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.205915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.205980] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.206240] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.206310] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.206600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.206683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.206934] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.206998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.207190] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.207253] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.207545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.207609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.207829] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.207904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.208224] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.208288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.208553] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.208618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.208858] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.208923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.209222] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.209296] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.209564] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.209628] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.209893] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.209958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.210249] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.210312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.210601] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.210695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.210974] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.211043] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.211291] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.211355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.536 [2024-12-06 17:54:27.211557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.536 [2024-12-06 17:54:27.211620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.536 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.211878] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.211941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.212245] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.212317] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.212517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.212586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.212899] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.212963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.213192] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.213255] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.213499] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.213562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.213821] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.213888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.214155] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.214228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.214475] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.214540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.214839] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.214905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.215189] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.215254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.215545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.215614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.215866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.215931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.216220] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.216283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.216481] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.216548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.216807] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.216872] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.217116] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.217182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.217441] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.217506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.217700] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.217765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.218021] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.218086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.218333] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.218403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.218710] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.218776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.219079] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.219143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.219414] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.219478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.219723] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.219789] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.220040] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.220108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.220375] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.220441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.220713] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.220779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.221070] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.221135] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.221357] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.221430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.221699] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.221765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.222019] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.222084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.222339] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.222404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.222694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.222759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.223015] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.223084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.223389] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.223453] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.223699] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.223776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.537 [2024-12-06 17:54:27.223975] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.537 [2024-12-06 17:54:27.224039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.537 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.224236] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.224308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.224581] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.224645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.224985] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.225051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.225284] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.225347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.225586] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.225649] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.225895] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.225965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.226165] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.226230] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.226494] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.226559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.226837] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.226902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.227189] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.227268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.227574] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.227639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.227919] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.227982] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.228200] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.228264] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.228462] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.228526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.228774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.228846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.229076] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.229146] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.229401] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.229467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.229656] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.229742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.230016] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.230089] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.230348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.230412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.230717] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.230782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.231042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.231107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.231403] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.231467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.231766] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.231832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.232097] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.232161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.232387] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.232452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.232656] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.232750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.232979] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.233044] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.233279] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.233343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.233584] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.233647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.233960] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.234023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.234325] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.234388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.234644] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.234730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.234991] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.235055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.235367] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.235430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.235685] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.235761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.236022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.236086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.538 [2024-12-06 17:54:27.236342] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.538 [2024-12-06 17:54:27.236408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.538 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.236677] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.236760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.237048] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.237112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.237359] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.237423] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.237687] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.237760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.237963] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.238029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.238318] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.238382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.238625] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.238705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.238999] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.239063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.239317] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.239381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.239602] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.239682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.239945] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.240009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.240266] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.240331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.240633] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.240726] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.240977] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.241042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.241295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.241358] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.241643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.241728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.242002] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.242067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.242312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.242375] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.242639] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.242718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.242981] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.243045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.243299] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.243363] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.243608] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.243689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.243951] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.244014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.244226] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.244290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.244591] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.244655] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.244920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.244983] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.245280] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.245344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.245559] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.245626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.245886] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.245950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.246199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.246263] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.246483] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.246549] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.246797] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.246861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.247071] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.539 [2024-12-06 17:54:27.247137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.539 qpair failed and we were unable to recover it. 00:38:45.539 [2024-12-06 17:54:27.247407] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.247472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.247776] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.247841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.248153] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.248216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.248448] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.248511] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.248779] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.248847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.249086] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.249150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.249441] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.249504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.249749] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.249835] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.250133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.250197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.250383] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.250447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.250741] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.250806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.251068] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.251132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.251424] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.251488] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.251738] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.251803] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.252095] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.252158] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.252403] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.252467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.252740] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.252805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.253060] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.253123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.253417] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.253480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.253729] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.253795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.254041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.254104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.254334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.254398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.254693] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.254759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.254999] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.255063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.255362] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.255425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.255741] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.255806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.256028] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.256091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.256333] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.256397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.256726] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.256791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.256984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.257049] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.257312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.257375] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.257624] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.257702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.258010] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.258074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.258269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.258332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.258601] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.258680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.258985] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.259049] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.259310] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.259372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.259612] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.540 [2024-12-06 17:54:27.259691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.540 qpair failed and we were unable to recover it. 00:38:45.540 [2024-12-06 17:54:27.259914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.259977] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.260217] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.260282] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.260546] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.260609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.260955] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.261019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.261263] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.261328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.261619] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.261700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.261921] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.261985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.262242] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.262306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.262519] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.262583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.262868] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.262944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.263243] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.263307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.263599] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.263679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.263954] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.264019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.264266] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.264329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.264616] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.264712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.264977] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.265041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.265245] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.265309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.265567] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.265630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.265921] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.265984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.266188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.266252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.266539] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.266602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.266878] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.266945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.267235] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.267299] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.267618] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.267699] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.267966] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.268030] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.268332] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.268396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.268715] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.268781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.269072] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.269136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.269391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.269454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.269702] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.269770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.270062] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.270126] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.270330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.270394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.270576] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.270642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.270915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.270979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.271266] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.271329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.271622] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.271705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.272020] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.272085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.272329] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.541 [2024-12-06 17:54:27.272392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.541 qpair failed and we were unable to recover it. 00:38:45.541 [2024-12-06 17:54:27.272632] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.272710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.272986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.273050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.273239] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.273302] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.273588] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.273652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.273889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.273954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.274193] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.274256] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.274510] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.274573] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.274882] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.274948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.275234] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.275298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.275596] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.275659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.275972] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.276036] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.276253] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.276328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.276623] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.276705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.276954] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.277020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.277238] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.277305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.277607] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.277689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.277945] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.278008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.278199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.278262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.278555] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.278619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.278896] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.278960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.279205] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.279272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.279578] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.279642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.279950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.280014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.280314] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.280378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.280629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.280711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.281016] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.281080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.281294] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.281360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.281608] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.281691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.281898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.281962] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.282226] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.282289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.282547] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.282610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.282856] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.282921] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.283120] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.283183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.283486] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.283549] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.283771] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.283835] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.284049] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.284113] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.284360] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.542 [2024-12-06 17:54:27.284425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.542 qpair failed and we were unable to recover it. 00:38:45.542 [2024-12-06 17:54:27.284716] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.284782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.285012] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.285076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.285314] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.285377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.285639] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.285720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.285943] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.286007] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.286240] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.286305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.286595] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.286658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.286915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.286980] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.287275] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.287340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.287602] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.287679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.287932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.287995] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.288285] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.288350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.288656] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.288738] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.288993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.289056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.289344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.289419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.289688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.289755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.290047] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.290111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.290411] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.290474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.290758] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.290823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.291128] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.291190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.291405] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.291468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.291693] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.291759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.292051] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.292114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.292355] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.292418] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.292661] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.292739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.293041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.293105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.293344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.293407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.293711] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.293776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.293995] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.294060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.294287] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.294350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.294592] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.294654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.294898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.294964] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.295208] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.295272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.295511] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.295573] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.295803] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.295870] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.296124] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.296189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.296439] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.543 [2024-12-06 17:54:27.296503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.543 qpair failed and we were unable to recover it. 00:38:45.543 [2024-12-06 17:54:27.296753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.296818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.297063] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.297127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.297418] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.297481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.297773] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.297838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.298148] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.298213] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.298473] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.298535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.298823] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.298887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.299186] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.299250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.299551] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.299615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.299923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.299986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.300239] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.300304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.300595] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.300657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.300929] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.300989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.301271] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.301332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.301568] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.301633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.301848] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.301912] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.302204] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.302268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.302523] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.302607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.302886] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.302961] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.303216] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.303288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.303536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.303601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.303874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.303941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.304154] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.304219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.304480] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.304544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.304804] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.304870] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.305165] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.305229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.305455] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.305526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.305772] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.305838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.306123] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.306187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.306398] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.306463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.306719] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.306785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.544 qpair failed and we were unable to recover it. 00:38:45.544 [2024-12-06 17:54:27.307045] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.544 [2024-12-06 17:54:27.307112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.307370] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.307437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.307697] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.307762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.307982] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.308052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.308298] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.308362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.308615] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.308698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.308896] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.308959] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.309214] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.309278] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.309554] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.309619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.309911] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.309977] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.310192] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.310258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.310515] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.310579] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.310853] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.310925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.311229] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.311295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.311544] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.311608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.311823] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.311888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.312086] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.312152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.312410] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.312478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.312766] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.312831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.313035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.313099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.313307] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.313372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.313585] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.313662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.314046] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.314114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.314355] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.314423] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.314630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.314759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.315014] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.315082] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.315341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.315423] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.315629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.315719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.316017] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.316081] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.316368] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.316430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.316701] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.316771] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.317021] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.317086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.317334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.317399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.317624] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.317708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.318007] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.318079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.318324] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.318388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.318597] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.318661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.318922] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.545 [2024-12-06 17:54:27.318986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.545 qpair failed and we were unable to recover it. 00:38:45.545 [2024-12-06 17:54:27.319251] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.546 [2024-12-06 17:54:27.319315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.546 qpair failed and we were unable to recover it. 00:38:45.546 [2024-12-06 17:54:27.319600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.546 [2024-12-06 17:54:27.319694] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.546 qpair failed and we were unable to recover it. 00:38:45.546 [2024-12-06 17:54:27.319927] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.546 [2024-12-06 17:54:27.319994] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.546 qpair failed and we were unable to recover it. 00:38:45.546 [2024-12-06 17:54:27.320248] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.546 [2024-12-06 17:54:27.320312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.546 qpair failed and we were unable to recover it. 00:38:45.546 [2024-12-06 17:54:27.320505] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.546 [2024-12-06 17:54:27.320571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.546 qpair failed and we were unable to recover it. 00:38:45.546 [2024-12-06 17:54:27.320831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.546 [2024-12-06 17:54:27.320896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.546 qpair failed and we were unable to recover it. 00:38:45.546 [2024-12-06 17:54:27.321169] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.546 [2024-12-06 17:54:27.321239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.546 qpair failed and we were unable to recover it. 00:38:45.546 [2024-12-06 17:54:27.321437] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.546 [2024-12-06 17:54:27.321506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.546 qpair failed and we were unable to recover it. 00:38:45.546 [2024-12-06 17:54:27.321725] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.546 [2024-12-06 17:54:27.321791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.546 qpair failed and we were unable to recover it. 00:38:45.546 [2024-12-06 17:54:27.322090] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.546 [2024-12-06 17:54:27.322155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.546 qpair failed and we were unable to recover it. 00:38:45.546 [2024-12-06 17:54:27.322412] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.546 [2024-12-06 17:54:27.322487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.546 qpair failed and we were unable to recover it. 00:38:45.546 [2024-12-06 17:54:27.322745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.546 [2024-12-06 17:54:27.322814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.546 qpair failed and we were unable to recover it. 00:38:45.546 [2024-12-06 17:54:27.323065] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.546 [2024-12-06 17:54:27.323130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.546 qpair failed and we were unable to recover it. 00:38:45.546 [2024-12-06 17:54:27.323348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.546 [2024-12-06 17:54:27.323411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.546 qpair failed and we were unable to recover it. 00:38:45.546 [2024-12-06 17:54:27.323654] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.546 [2024-12-06 17:54:27.323740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.546 qpair failed and we were unable to recover it. 00:38:45.546 [2024-12-06 17:54:27.324009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.546 [2024-12-06 17:54:27.324076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.546 qpair failed and we were unable to recover it. 00:38:45.546 [2024-12-06 17:54:27.324278] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.546 [2024-12-06 17:54:27.324343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.546 qpair failed and we were unable to recover it. 00:38:45.546 [2024-12-06 17:54:27.324538] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.546 [2024-12-06 17:54:27.324602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.546 qpair failed and we were unable to recover it. 00:38:45.546 [2024-12-06 17:54:27.324924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.546 [2024-12-06 17:54:27.324990] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.546 qpair failed and we were unable to recover it. 00:38:45.546 [2024-12-06 17:54:27.325243] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.823 [2024-12-06 17:54:27.325313] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.823 qpair failed and we were unable to recover it. 00:38:45.823 [2024-12-06 17:54:27.325612] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.823 [2024-12-06 17:54:27.325701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.823 qpair failed and we were unable to recover it. 00:38:45.823 [2024-12-06 17:54:27.325924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.823 [2024-12-06 17:54:27.325989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.823 qpair failed and we were unable to recover it. 00:38:45.823 [2024-12-06 17:54:27.326230] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.823 [2024-12-06 17:54:27.326293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.823 qpair failed and we were unable to recover it. 00:38:45.823 [2024-12-06 17:54:27.326517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.823 [2024-12-06 17:54:27.326582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.823 qpair failed and we were unable to recover it. 00:38:45.823 [2024-12-06 17:54:27.326887] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.823 [2024-12-06 17:54:27.326954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.823 qpair failed and we were unable to recover it. 00:38:45.823 [2024-12-06 17:54:27.327179] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.823 [2024-12-06 17:54:27.327243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.823 qpair failed and we were unable to recover it. 00:38:45.823 [2024-12-06 17:54:27.327456] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.823 [2024-12-06 17:54:27.327519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.823 qpair failed and we were unable to recover it. 00:38:45.823 [2024-12-06 17:54:27.327820] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.327886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.328073] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.328158] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.328431] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.328498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.328763] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.328829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.329025] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.329090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.329374] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.329437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.329703] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.329774] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.329999] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.330064] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.330306] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.330370] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.330634] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.330718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.330982] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.331047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.331354] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.331424] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.331710] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.331776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.332015] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.332078] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.332275] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.332339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.332657] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.332744] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.333014] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.333079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.333264] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.333328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.333518] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.333583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.333886] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.333952] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.334248] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.334321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.334524] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.334588] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.334856] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.334922] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.335128] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.335191] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.335478] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.335541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.335794] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.335861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.336157] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.336227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.336544] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.336607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.336928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.336999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.337226] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.337291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.337512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.337576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.337841] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.337907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.338089] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.338152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.338403] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.338467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.338718] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.338790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.339052] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.339116] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.339351] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.339415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.824 qpair failed and we were unable to recover it. 00:38:45.824 [2024-12-06 17:54:27.339695] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.824 [2024-12-06 17:54:27.339761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.340011] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.340081] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.340345] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.340411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.340719] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.340784] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.341026] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.341102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.341351] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.341418] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.341721] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.341792] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.342041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.342106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.342375] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.342440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.342696] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.342762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.343004] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.343070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.343282] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.343349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.343589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.343654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.343880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.343945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.344133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.344196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.344439] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.344511] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.344780] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.344847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.345161] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.345224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.345417] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.345480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.345775] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.345844] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.346107] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.346171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.346469] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.346533] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.346784] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.346850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.347107] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.347170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.347457] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.347522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.347782] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.347850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.348044] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.348107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.348397] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.348460] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.348715] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.348780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.349046] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.349118] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.349381] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.349448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.349717] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.349783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.350025] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.350089] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.350337] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.350414] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.350686] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.350752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.351042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.351106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.351353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.351418] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.351683] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.351748] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.825 qpair failed and we were unable to recover it. 00:38:45.825 [2024-12-06 17:54:27.351997] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.825 [2024-12-06 17:54:27.352067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.352295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.352360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.352647] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.352742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.353039] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.353102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.353359] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.353424] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.353634] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.353719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.353941] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.354015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.354265] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.354330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.354565] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.354629] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.354887] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.354956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.355242] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.355307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.355536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.355600] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.355864] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.355928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.356199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.356270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.356497] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.356562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.356800] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.356865] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.357110] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.357174] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.357398] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.357462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.357757] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.357830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.358054] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.358121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.358344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.358407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.358650] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.358733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.358992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.359056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.359290] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.359359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.359681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.359748] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.359982] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.360046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.360253] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.360317] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.360559] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.360624] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.360888] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.360952] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.361206] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.361269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.361517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.361581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.361856] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.361921] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.362221] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.362285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.362556] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.362621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.362937] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.363001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.363241] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.363305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.363562] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.363625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.363898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.363962] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.826 qpair failed and we were unable to recover it. 00:38:45.826 [2024-12-06 17:54:27.364226] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.826 [2024-12-06 17:54:27.364289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.364546] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.364610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.364820] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.364885] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.365174] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.365237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.365529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.365594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.365882] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.365986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.366280] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.366359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.366625] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.366721] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.366999] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.367066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.367340] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.367405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.367699] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.367766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.367980] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.368045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.368324] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.368395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.368639] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.368740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.368971] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.369037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.369343] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.369412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.369717] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.369785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.370010] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.370076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.370379] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.370445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.370660] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.370747] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.370957] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.371021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.371279] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.371355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.371612] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.371715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.371953] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.372017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.372274] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.372344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.372657] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.372746] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.373017] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.373083] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.373390] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.373456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.373739] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.373807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.374082] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.374158] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.374453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.374518] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.374779] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.374846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.375102] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.375180] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.375475] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.375541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.375836] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.375903] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.376140] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.376206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.376507] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.376576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.376897] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.376962] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.377249] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.827 [2024-12-06 17:54:27.377313] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.827 qpair failed and we were unable to recover it. 00:38:45.827 [2024-12-06 17:54:27.377529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.377595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.377832] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.377897] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.378161] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.378225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.378511] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.378576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.378872] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.378937] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.379185] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.379249] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.379495] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.379559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.379783] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.379848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.380094] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.380157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.380422] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.380486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.380704] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.380781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.381032] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.381098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.381350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.381414] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.381678] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.381743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.381952] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.382016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.382271] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.382336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.382593] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.382656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.382868] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.382932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.383224] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.383288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.383536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.383599] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.383867] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.383932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.384220] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.384285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.384574] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.384638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.384874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.384938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.385192] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.385266] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.385629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.385717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.385981] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.386055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.386370] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.386439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.386728] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.386796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.387063] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.387127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.387347] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.387434] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.387702] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.387770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.388008] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.388072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.828 [2024-12-06 17:54:27.388321] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.828 [2024-12-06 17:54:27.388388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.828 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.388686] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.388752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.389039] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.389102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.389393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.389457] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.389711] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.389777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.390040] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.390105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.390309] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.390376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.390696] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.390762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.391004] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.391068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.391331] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.391396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.391705] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.391770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.392060] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.392124] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.392372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.392436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.392693] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.392758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.392969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.393034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.393335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.393398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.393656] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.393735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.393980] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.394045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.394334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.394409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.394738] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.394804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.395004] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.395068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.395364] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.395428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.395695] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.395761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.396050] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.396114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.396356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.396420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.396684] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.396751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.397003] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.397067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.397352] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.397416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.397620] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.397701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.397965] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.398029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.398322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.398386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.398630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.398743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.399024] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.399089] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.399373] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.399437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.399655] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.399738] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.399950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.400014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.400254] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.400318] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.400606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.400689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.401001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.829 [2024-12-06 17:54:27.401066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.829 qpair failed and we were unable to recover it. 00:38:45.829 [2024-12-06 17:54:27.401315] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.401381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.401643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.401727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.402029] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.402094] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.402355] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.402419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.402691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.402756] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.403000] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.403064] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.403306] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.403380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.403584] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.403648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.403882] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.403946] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.404212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.404275] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.404523] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.404586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.404903] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.404968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.405220] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.405283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.405549] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.405614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.405881] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.405944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.406202] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.406267] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.406465] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.406530] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.406786] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.406851] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.407150] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.407213] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.407462] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.407527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.407840] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.407905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.408205] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.408269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.408494] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.408559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.408862] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.408927] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.409196] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.409259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.409558] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.409623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.409831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.409894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.410107] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.410172] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.410461] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.410526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.410792] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.410857] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.411117] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.411181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.411413] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.411477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.411699] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.411767] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.412055] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.412119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.412396] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.412460] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.412656] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.412734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.413022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.413086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.413319] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.413384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.830 [2024-12-06 17:54:27.413595] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.830 [2024-12-06 17:54:27.413658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.830 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.413924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.413991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.414290] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.414356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.414646] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.414742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.415034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.415098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.415316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.415382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.415596] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.415661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.415973] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.416037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.416285] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.416349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.416641] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.416737] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.416984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.417048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.417287] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.417351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.417600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.417682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.417937] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.418001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.418249] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.418314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.418609] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.418701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.418946] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.419010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.419301] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.419366] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.419662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.419747] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.419957] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.420022] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.420267] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.420330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.420622] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.420705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.420925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.420989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.421232] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.421296] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.421585] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.421649] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.421928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.421992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.422261] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.422324] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.422571] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.422635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.422878] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.422943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.423181] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.423244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.423497] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.423561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.423824] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.423892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.424129] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.424193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.424445] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.424510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.424753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.424821] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.425113] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.425177] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.425417] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.425493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.425745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.425812] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.426062] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.426127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.831 qpair failed and we were unable to recover it. 00:38:45.831 [2024-12-06 17:54:27.426380] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.831 [2024-12-06 17:54:27.426456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.426757] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.426823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.427139] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.427205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.427502] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.427576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.427890] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.427965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.428169] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.428235] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.428440] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.428506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.428786] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.428854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.429151] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.429216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.429465] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.429530] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.429775] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.429842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.430120] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.430190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.430486] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.430556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.430846] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.430920] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.431218] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.431283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.431542] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.431616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.431926] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.431991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.432274] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.432342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.432652] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.432738] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.433013] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.433085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.433355] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.433420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.433637] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.433731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.434005] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.434071] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.434372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.434437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.434716] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.434785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.435080] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.435145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.435439] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.435515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.435744] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.435812] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.436075] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.436145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.436424] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.436493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.436820] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.436888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.437151] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.437216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.437499] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.437568] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.437858] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.437926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.832 qpair failed and we were unable to recover it. 00:38:45.832 [2024-12-06 17:54:27.438242] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.832 [2024-12-06 17:54:27.438308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.438570] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.438635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.438876] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.438952] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.439206] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.439270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.439569] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.439645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.439989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.440056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.440286] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.440355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.440651] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.440754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.441010] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.441075] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.441326] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.441402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.441703] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.441771] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.441993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.442069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.442357] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.442422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.442732] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.442802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.443109] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.443181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.443479] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.443547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.443809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.443877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.444145] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.444210] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.444524] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.444598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.444850] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.444927] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.445229] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.445294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.445515] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.445588] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.445866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.445934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.446192] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.446257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.446520] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.446584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.446815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.446881] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.447180] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.447253] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.447549] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.447614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.447910] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.447980] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.448224] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.448290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.448545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.448618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.448916] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.448982] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.449245] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.449311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.449560] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.449631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.449920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.449985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.450207] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.450284] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.450539] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.450605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.450939] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.833 [2024-12-06 17:54:27.451016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.833 qpair failed and we were unable to recover it. 00:38:45.833 [2024-12-06 17:54:27.451325] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.451389] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.451644] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.451733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.452025] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.452089] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.452341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.452404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.452698] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.452764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.453034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.453098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.453344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.453409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.453705] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.453815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.454102] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.454170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.454395] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.454460] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.454762] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.454828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.455095] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.455162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.455410] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.455473] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.455692] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.455757] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.456001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.456065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.456349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.456412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.456690] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.456755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.457053] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.457117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.457331] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.457395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.457646] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.457735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.457982] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.458046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.458313] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.458377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.458635] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.458737] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.459020] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.459085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.459367] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.459430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.459739] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.459805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.460088] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.460152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.460397] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.460460] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.460694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.460760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.460996] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.461060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.461303] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.461367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.461631] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.461709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.461971] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.462037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.462264] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.462328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.462577] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.462651] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.462974] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.463038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.463310] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.463374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.463687] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.463754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.834 [2024-12-06 17:54:27.464012] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.834 [2024-12-06 17:54:27.464076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.834 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.464369] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.464434] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.464728] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.464794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.465094] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.465159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.465398] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.465463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.465758] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.465823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.466066] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.466130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.466411] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.466476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.466786] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.466852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.467142] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.467206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.467512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.467576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.467862] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.467927] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.468221] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.468284] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.468535] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.468599] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.468884] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.468948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.469190] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.469254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.469493] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.469557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.469844] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.469944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.470183] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.470253] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.470502] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.470569] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.470840] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.470906] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.471206] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.471268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.471484] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.471547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.471843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.471922] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.472201] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.472265] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.472570] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.472633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.472902] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.472968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.473256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.473322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.473620] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.473700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.473958] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.474021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.474273] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.474338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.474604] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.474685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.474946] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.475012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.475306] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.475369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.475614] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.475697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.475956] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.476020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.476268] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.476331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.476633] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.476716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.477002] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.835 [2024-12-06 17:54:27.477066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.835 qpair failed and we were unable to recover it. 00:38:45.835 [2024-12-06 17:54:27.477269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.477333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.477558] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.477621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.477931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.477996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.478295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.478359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.478607] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.478692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.478947] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.479012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.479272] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.479335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.479594] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.479657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.479908] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.479974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.480224] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.480289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.480548] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.480611] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.480948] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.481015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.481212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.481275] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.481566] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.481630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.481895] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.481964] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.482212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.482277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.482520] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.482584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.482860] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.482925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.483168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.483232] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.483444] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.483508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.483753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.483819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.484062] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.484126] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.484417] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.484480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.484772] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.484835] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.485137] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.485211] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.485469] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.485532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.485774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.485838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.486047] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.486111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.486401] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.486465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.486769] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.486833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.487135] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.487199] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.487495] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.487559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.487891] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.487957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.488252] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.488317] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.488582] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.488644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.488919] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.488986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.489286] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.489351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.489609] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.836 [2024-12-06 17:54:27.489689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.836 qpair failed and we were unable to recover it. 00:38:45.836 [2024-12-06 17:54:27.489970] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.490033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.490322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.490387] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.490697] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.490762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.491028] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.491092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.491399] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.491463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.491710] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.491777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.491993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.492059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.492274] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.492337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.492625] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.492704] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.492925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.492989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.493275] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.493337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.493529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.493593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.493810] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.493874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.494096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.494160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.494381] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.494445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.494758] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.494823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.495110] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.495174] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.495421] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.495484] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.495728] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.495794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.496080] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.496144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.496341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.496406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.496694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.496759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.497008] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.497073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.497319] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.497382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.497628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.497709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.497980] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.498046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.498272] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.498346] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.498636] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.498715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.498964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.499027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.499292] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.499355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.499605] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.499686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.499936] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.500001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.500259] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.500323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.500568] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.837 [2024-12-06 17:54:27.500632] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.837 qpair failed and we were unable to recover it. 00:38:45.837 [2024-12-06 17:54:27.500925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.500989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.501238] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.501303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.501568] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.501631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.501944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.502007] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.502313] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.502377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.502624] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.502710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.502978] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.503042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.503330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.503394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.503600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.503682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.503891] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.503954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.504196] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.504259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.504564] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.504629] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.504944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.505008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.505220] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.505284] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.505573] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.505637] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.505911] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.505974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.506233] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.506297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.506492] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.506557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.506815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.506879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.507152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.507217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.507467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.507532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.507777] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.507842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.508144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.508207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.508448] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.508514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.508804] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.508870] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.509077] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.509140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.509352] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.509416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.509659] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.509740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.510002] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.510067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.510257] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.510321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.510575] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.510640] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.510949] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.511013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.511255] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.511330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.511584] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.511648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.511908] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.511971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.512220] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.512283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.512579] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.512642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.512986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.838 [2024-12-06 17:54:27.513050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.838 qpair failed and we were unable to recover it. 00:38:45.838 [2024-12-06 17:54:27.513311] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.513375] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.513629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.513713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.513973] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.514047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.514290] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.514354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.514576] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.514639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.514909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.514971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.515223] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.515287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.515579] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.515642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.515922] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.515985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.516234] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.516300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.516516] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.516579] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.516873] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.516939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.517200] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.517264] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.517539] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.517602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.517831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.517895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.518149] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.518212] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.518504] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.518569] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.518803] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.518867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.519116] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.519180] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.519392] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.519457] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.519707] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.519772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.520066] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.520131] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.520395] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.520460] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.520689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.520756] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.521047] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.521110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.521396] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.521460] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.521723] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.521788] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.522043] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.522106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.522352] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.522415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.522679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.522745] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.522953] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.523018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.523321] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.523385] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.523639] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.523724] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.524023] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.524086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.524345] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.524419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.524727] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.524794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.524992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.525059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.525349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.525414] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.839 qpair failed and we were unable to recover it. 00:38:45.839 [2024-12-06 17:54:27.525719] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.839 [2024-12-06 17:54:27.525785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.526031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.526094] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.526347] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.526411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.526676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.526744] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.527011] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.527075] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.527315] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.527378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.527701] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.527766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.528006] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.528068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.528322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.528385] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.528633] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.528716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.528983] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.529046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.529302] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.529365] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.529600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.529680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.529984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.530047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.530253] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.530316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.530587] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.530650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.530919] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.530982] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.531231] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.531295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.531588] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.531652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.531931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.532004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.532270] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.532334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.532625] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.532708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.533007] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.533070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.533324] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.533390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.533648] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.533731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.533989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.534051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.534347] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.534410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.534706] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.534773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.535036] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.535099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.535402] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.535465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.535753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.535817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.536113] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.536177] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.536433] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.536497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.536750] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.536816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.537076] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.537140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.537393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.537455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.537700] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.537776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.538055] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.538119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.840 [2024-12-06 17:54:27.538364] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.840 [2024-12-06 17:54:27.538428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.840 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.538686] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.538752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.538963] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.539030] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.539283] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.539347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.539606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.539704] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.539959] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.540024] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.540252] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.540316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.540624] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.540707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.540980] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.541043] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.541329] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.541392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.541699] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.541766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.542067] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.542130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.542397] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.542461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.542719] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.542785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.543048] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.543111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.543370] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.543433] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.543649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.543749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.544037] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.544100] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.544396] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.544460] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.544756] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.544822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.545007] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.545073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.545330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.545394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.545637] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.545721] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.545970] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.546034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.546326] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.546390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.546617] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.546697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.546917] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.546981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.547230] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.547294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.547582] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.547645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.547883] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.547948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.548203] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.548267] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.548570] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.548633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.548912] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.548976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.549166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.549229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.549472] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.549536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.549832] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.841 [2024-12-06 17:54:27.549897] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.841 qpair failed and we were unable to recover it. 00:38:45.841 [2024-12-06 17:54:27.550187] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.550250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.550503] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.550567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.550818] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.550882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.551184] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.551247] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.551515] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.551578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.551846] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.551910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.552166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.552234] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.552488] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.552559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.552815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.552879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.553075] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.553139] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.553343] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.553415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.553658] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.553745] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.553941] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.554005] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.554256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.554320] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.554557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.554621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.554913] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.554978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.555199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.555266] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.555526] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.555590] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.555899] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.555964] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.556261] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.556325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.556580] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.556650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.556977] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.557041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.557242] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.557307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.557580] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.557645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.557885] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.557951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.558210] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.558279] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.558486] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.558550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.558810] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.558876] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.559168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.559239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.559461] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.559537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.559784] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.559851] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.560145] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.560209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.560416] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.560479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.560714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.560782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.560994] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.561063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.561325] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.561389] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.842 qpair failed and we were unable to recover it. 00:38:45.842 [2024-12-06 17:54:27.561635] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.842 [2024-12-06 17:54:27.561714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.561916] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.561982] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.562226] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.562297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.562544] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.562610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.562817] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.562882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.563093] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.563156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.563406] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.563469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.563774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.563841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.564096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.564160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.564429] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.564494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.564787] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.564852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.565121] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.565191] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.565447] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.565513] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.565748] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.565814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.566023] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.566087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.566345] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.566409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.566635] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.566712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.566960] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.567029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.567294] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.567360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.567598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.567660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.567970] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.568034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.568269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.568339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.568543] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.568608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.568891] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.568957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.569174] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.569239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.569490] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.569555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.569812] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.569880] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.570131] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.570202] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.570494] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.570558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.570786] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.570852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.571103] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.571174] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.571398] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.571463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.571766] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.571833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.572085] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.572159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.572442] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.572507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.572789] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.572854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.573066] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.573136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.573391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.843 [2024-12-06 17:54:27.573455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.843 qpair failed and we were unable to recover it. 00:38:45.843 [2024-12-06 17:54:27.573754] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.573819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.574078] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.574142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.574402] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.574467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.574691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.574756] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.574990] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.575054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.575297] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.575361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.575590] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.575655] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.575970] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.576041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.576298] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.576362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.576620] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.576720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.577016] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.577086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.577331] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.577396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.577600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.577684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.577914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.577977] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.578211] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.578273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.578458] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.578527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.578816] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.578888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.579181] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.579244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.579548] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.579611] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.579839] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.579904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.580150] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.580222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.580494] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.580559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.580783] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.580848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.581142] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.581205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.581456] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.581527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.581742] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.581808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.582064] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.582127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.582351] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.582414] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.582701] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.582768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.583066] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.583129] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.583373] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.583438] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.583693] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.583759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.584015] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.584079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.584337] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.584402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.584619] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.584714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.584931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.585008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.585242] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.585306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.585553] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.844 [2024-12-06 17:54:27.585617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.844 qpair failed and we were unable to recover it. 00:38:45.844 [2024-12-06 17:54:27.585916] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.585982] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.586274] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.586338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.586563] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.586627] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.586933] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.586996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.587293] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.587362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.587621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.587706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.587956] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.588020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.588258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.588321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.588538] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.588608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.588889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.588954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.589164] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.589230] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.589509] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.589574] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.589854] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.589920] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.590154] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.590226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.590425] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.590490] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.590785] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.590850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.591064] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.591127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.591418] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.591483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.591711] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.591785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.592076] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.592142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.592440] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.592505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.592770] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.592835] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.593129] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.593194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.593456] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.593520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.593815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.593881] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.594125] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.594188] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.594432] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.594494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.594761] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.594827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.595080] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.595147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.595404] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.595469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.595689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.595763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.596056] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.596120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.596372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.596435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.596728] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.596798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.597097] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.597162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.597458] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.597521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.597812] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.597877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.598122] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.845 [2024-12-06 17:54:27.598203] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.845 qpair failed and we were unable to recover it. 00:38:45.845 [2024-12-06 17:54:27.598502] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.598566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.598830] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.598896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.599140] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.599204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.599411] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.599473] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.599748] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.599815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.600030] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.600094] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.600345] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.600409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.600589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.600652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.600963] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.601032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.601280] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.601344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.601605] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.601684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.601931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.601994] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.602290] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.602354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.602625] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.602711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.602953] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.603017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.603262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.603325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.603614] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.603698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.603950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.604013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.604268] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.604337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.604639] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.604724] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.604977] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.605041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.605336] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.605400] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.605654] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.605743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.605968] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.606036] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.606335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.606399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.606648] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.606729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.607019] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.607092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.607358] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.607424] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.607661] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.607746] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.608000] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.608064] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.608303] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.608366] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.608624] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.608728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.608976] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.609041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.609328] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.609391] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.846 qpair failed and we were unable to recover it. 00:38:45.846 [2024-12-06 17:54:27.609637] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.846 [2024-12-06 17:54:27.609721] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.609982] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.610053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.610343] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.610408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.610700] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.610765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.611049] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.611112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.611341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.611415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.611639] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.611723] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.611933] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.611998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.612241] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.612305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.612530] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.612592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.612927] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.613000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.613261] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.613326] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.613575] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.613638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.613947] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.614011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.614219] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.614283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.614496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.614571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.614882] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.614948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.615207] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.615274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.615532] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.615596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.615900] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.615975] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.616235] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.616299] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.616586] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.616650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.616955] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.617019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.617267] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.617330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.617599] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.617723] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.618001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.618070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.618380] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.618445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.618685] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.618751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.619005] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.619072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.619376] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.619439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.619647] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.619735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.619983] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.620050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.620292] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.620390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.620662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.620774] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.621035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.621099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.621387] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.621451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.621642] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.621731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.621995] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.622066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.847 [2024-12-06 17:54:27.622330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.847 [2024-12-06 17:54:27.622397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.847 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.622609] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.622693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.622946] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.623012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.623308] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.623373] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.623624] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.623711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.623967] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.624032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.624284] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.624348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.624578] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.624656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.624940] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.625012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.625238] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.625307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.625559] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.625623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.625868] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.625932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.626125] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.626188] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.626437] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.626510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.626715] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.626790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.626997] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.627063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.627367] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.627431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.627662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.627752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.628020] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.628094] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.628372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.628437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.628647] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.628725] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.629004] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.629069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.629259] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.629322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.629559] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.629623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.629931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.629997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.630272] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.630335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.630552] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.630615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.630883] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.630947] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.631208] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.631278] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.631528] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.631591] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.631874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.631939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.632188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.632252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.632553] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.632624] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.632966] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.633031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.633328] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.633394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.633701] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.633766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.634052] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.634122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.634390] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.634455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.848 [2024-12-06 17:54:27.634684] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.848 [2024-12-06 17:54:27.634749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.848 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.635001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.635065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.635325] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.635400] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.635622] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.635712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.635984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.636050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.636265] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.636328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.636578] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.636642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.636914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.636980] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.637231] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.637295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.637546] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.637610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.637914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.637979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.638224] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.638288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.638538] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.638603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.638865] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.638931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.639216] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.639280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.639525] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.639591] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.639883] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.639956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.640199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.640265] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.640507] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.640569] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.640887] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.640953] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.641204] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.641268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.641536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.641600] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.641838] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.641902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.642173] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.642238] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.642501] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.642563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.642813] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.642887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.643112] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.643185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.643435] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.643500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.643749] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.643815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.644100] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.644163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.644397] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.644466] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:45.849 [2024-12-06 17:54:27.644790] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:45.849 [2024-12-06 17:54:27.644857] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:45.849 qpair failed and we were unable to recover it. 00:38:46.123 [2024-12-06 17:54:27.645062] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.123 [2024-12-06 17:54:27.645129] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.123 qpair failed and we were unable to recover it. 00:38:46.123 [2024-12-06 17:54:27.645333] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.123 [2024-12-06 17:54:27.645397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.123 qpair failed and we were unable to recover it. 00:38:46.123 [2024-12-06 17:54:27.645606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.123 [2024-12-06 17:54:27.645710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.123 qpair failed and we were unable to recover it. 00:38:46.123 [2024-12-06 17:54:27.645983] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.123 [2024-12-06 17:54:27.646048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.123 qpair failed and we were unable to recover it. 00:38:46.123 [2024-12-06 17:54:27.646266] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.123 [2024-12-06 17:54:27.646340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.123 qpair failed and we were unable to recover it. 00:38:46.123 [2024-12-06 17:54:27.646534] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.123 [2024-12-06 17:54:27.646600] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.123 qpair failed and we were unable to recover it. 00:38:46.123 [2024-12-06 17:54:27.646877] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.123 [2024-12-06 17:54:27.646941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.123 qpair failed and we were unable to recover it. 00:38:46.123 [2024-12-06 17:54:27.647180] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.123 [2024-12-06 17:54:27.647250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.123 qpair failed and we were unable to recover it. 00:38:46.123 [2024-12-06 17:54:27.647456] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.123 [2024-12-06 17:54:27.647532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.123 qpair failed and we were unable to recover it. 00:38:46.123 [2024-12-06 17:54:27.647783] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.123 [2024-12-06 17:54:27.647848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.123 qpair failed and we were unable to recover it. 00:38:46.123 [2024-12-06 17:54:27.648061] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.123 [2024-12-06 17:54:27.648127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.123 qpair failed and we were unable to recover it. 00:38:46.123 [2024-12-06 17:54:27.648424] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.648488] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.648701] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.648773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.648980] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.649045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.649307] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.649371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.649603] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.649693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.649986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.650050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.650339] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.650408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.650618] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.650701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.650954] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.651018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.651267] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.651331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.651576] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.651647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.651892] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.651956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.652213] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.652277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.652534] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.652598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.652822] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.652886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.653179] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.653250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.653524] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.653589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.653829] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.653895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.654182] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.654246] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.654540] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.654611] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.654918] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.654983] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.655278] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.655341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.655585] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.655650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.655883] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.655946] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.656209] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.656274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.656520] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.656585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.656887] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.656951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.657239] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.657303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.657556] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.657622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.657953] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.658053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.658362] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.658430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.658724] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.658793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.659093] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.659159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.659470] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.659547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.659813] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.659879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.660162] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.660226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.660435] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.660499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.124 [2024-12-06 17:54:27.660725] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.124 [2024-12-06 17:54:27.660791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.124 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.661041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.661104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.661357] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.661422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.661689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.661755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.661975] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.662039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.662257] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.662321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.662554] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.662618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.662927] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.662992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.663240] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.663303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.663550] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.663614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.663909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.663977] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.664227] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.664291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.664545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.664610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.664890] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.664954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.665196] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.665260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.665512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.665576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.665803] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.665869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.666155] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.666219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.666438] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.666503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.666762] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.666828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.667079] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.667144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.667339] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.667404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.667625] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.667702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.667959] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.668023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.668289] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.668353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.668595] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.668659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.668891] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.668955] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.669165] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.669228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.669493] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.669557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.669860] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.669925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.670136] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.670199] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.670467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.670531] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.670902] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.671002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.671256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.671331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.671628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.671716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.671941] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.672009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.672247] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.672314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.125 [2024-12-06 17:54:27.672551] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.125 [2024-12-06 17:54:27.672616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.125 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.672860] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.672925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.673180] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.673246] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.673506] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.673575] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.673891] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.673957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.674153] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.674223] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.674450] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.674517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.674781] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.674847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.675115] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.675180] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.675469] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.675532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.675790] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.675866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.676160] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.676225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.676457] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.676520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.676814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.676892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.677173] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.677242] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.677545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.677610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.677875] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.677939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.678193] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.678257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.678544] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.678607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.678843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.678910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.679156] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.679221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.679518] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.679581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.679881] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.679946] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.680184] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.680250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.680519] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.680583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.680888] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.680954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.681205] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.681269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.681574] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.681643] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.681894] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.681965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.682181] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.682246] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.682494] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.682556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.682868] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.682933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.683190] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.683256] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.683475] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.683540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.683809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.683875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.684170] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.126 [2024-12-06 17:54:27.684233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.126 qpair failed and we were unable to recover it. 00:38:46.126 [2024-12-06 17:54:27.684480] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.684543] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.684840] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.684906] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.685159] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.685224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.685490] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.685554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.685833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.685899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.686113] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.686177] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.686448] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.686514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.686795] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.686861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.687112] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.687176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.687463] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.687526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.687768] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.687833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.688064] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.688136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.688384] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.688449] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.688714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.688779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.689072] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.689135] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.689373] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.689439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.689695] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.689762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.689977] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.690052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.690253] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.690317] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.690598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.690660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.690880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.690947] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.691205] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.691273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.691522] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.691585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.691918] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.691984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.692225] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.692297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.692548] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.692615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.692898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.692962] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.693200] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.693263] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.693510] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.693573] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.693887] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.693957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.694251] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.694317] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.694612] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.694693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.694950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.695014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.695221] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.695292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.695581] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.695645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.695932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.695996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.696217] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.127 [2024-12-06 17:54:27.696279] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.127 qpair failed and we were unable to recover it. 00:38:46.127 [2024-12-06 17:54:27.696571] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.696640] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.696912] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.696979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.697189] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.697253] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.697541] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.697604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.697816] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.697881] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.698120] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.698187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.698496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.698561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.698794] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.698862] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.699153] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.699218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.699506] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.699575] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.699814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.699885] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.700135] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.700199] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.700483] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.700546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.700790] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.700856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.701130] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.701198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.701425] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.701488] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.701704] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.701769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.702041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.702105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.702351] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.702413] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.702674] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.702746] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.703046] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.703121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.703330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.703393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.703624] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.703713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.703931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.704003] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.704261] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.704333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.704530] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.704595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.704845] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.704909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.705160] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.705224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.705484] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.705555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.128 qpair failed and we were unable to recover it. 00:38:46.128 [2024-12-06 17:54:27.705828] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.128 [2024-12-06 17:54:27.705893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.706063] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.706127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.706371] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.706434] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.706637] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.706718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.707016] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.707083] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.707350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.707414] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.707700] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.707764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.707981] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.708046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.708312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.708378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.708628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.708711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.708969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.709033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.709292] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.709357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.709615] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.709694] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.709991] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.710054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.710300] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.710364] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.710622] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.710702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.710964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.711027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.711282] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.711345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.711542] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.711616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.711928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.711992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.712277] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.712340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.712589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.712652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.712961] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.713025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.713289] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.713352] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.713606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.713684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.713981] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.714045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.714288] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.714351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.714616] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.714695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.714958] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.715023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.715316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.715379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.715582] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.715645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.715942] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.716005] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.716275] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.716339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.716631] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.716711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.716974] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.717037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.717326] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.129 [2024-12-06 17:54:27.717391] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.129 qpair failed and we were unable to recover it. 00:38:46.129 [2024-12-06 17:54:27.717694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.717759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.717969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.718034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.718279] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.718342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.718621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.718700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.718950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.719014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.719319] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.719381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.719571] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.719634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.719869] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.719934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.720166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.720230] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.720538] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.720602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.720835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.720899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.721190] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.721254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.721555] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.721619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.721928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.721992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.722199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.722262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.722475] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.722538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.722759] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.722825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.723116] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.723180] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.723477] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.723541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.723751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.723816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.724066] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.724131] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.724383] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.724448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.724700] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.724774] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.725062] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.725126] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.725430] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.725493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.725741] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.725806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.726104] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.726167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.726378] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.726440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.726701] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.726766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.727008] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.727071] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.727290] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.727353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.727643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.727742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.727993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.728057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.728344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.728407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.728652] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.728733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.728985] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.729048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.729355] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.729419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.729688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.130 [2024-12-06 17:54:27.729752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.130 qpair failed and we were unable to recover it. 00:38:46.130 [2024-12-06 17:54:27.730052] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.730115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.730412] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.730475] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.730691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.730758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.731039] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.731102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.731356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.731418] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.731719] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.731785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.732030] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.732093] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.732293] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.732357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.732567] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.732631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.732902] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.732965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.733218] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.733281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.733556] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.733620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.733923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.733986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.734284] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.734347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.734539] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.734603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.734823] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.734887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.735129] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.735192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.735449] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.735512] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.735796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.735862] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.736156] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.736219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.736531] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.736594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.736898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.736963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.737154] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.737217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.737508] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.737571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.737830] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.737907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.738135] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.738198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.738409] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.738472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.738769] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.738834] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.739120] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.739182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.739469] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.739533] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.739836] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.739902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.740151] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.740214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.740437] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.740500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.740721] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.740787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.741077] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.741140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.741440] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.741503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.741803] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.741868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.742169] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.131 [2024-12-06 17:54:27.742232] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.131 qpair failed and we were unable to recover it. 00:38:46.131 [2024-12-06 17:54:27.742499] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.742562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.742819] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.742882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.743168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.743232] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.743485] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.743548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.743821] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.743887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.744132] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.744195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.744477] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.744541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.744837] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.744903] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.745201] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.745265] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.745512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.745576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.745881] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.745945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.746236] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.746300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.746506] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.746570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.746886] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.746950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.747239] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.747301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.747552] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.747615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.747861] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.747927] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.748176] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.748239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.748503] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.748565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.748854] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.748918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.749201] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.749264] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.749466] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.749528] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.749809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.749874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.750186] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.750249] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.750553] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.750615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.750889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.750953] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.751212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.751287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.751540] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.751602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.751924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.751989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.752249] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.752312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.752606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.752682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.752969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.753032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.753250] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.753314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.753555] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.753620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.753854] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.753919] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.754177] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.754240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.754446] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.754509] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.754757] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.132 [2024-12-06 17:54:27.754823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.132 qpair failed and we were unable to recover it. 00:38:46.132 [2024-12-06 17:54:27.755072] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.755135] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.755378] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.755441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.755749] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.755815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.756103] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.756165] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.756423] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.756486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.756737] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.756802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.757044] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.757106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.757403] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.757466] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.757730] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.757795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.758091] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.758154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.758416] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.758478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.758777] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.758843] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.759094] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.759156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.759414] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.759478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.759752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.759817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.760126] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.760189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.760383] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.760448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.760736] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.760801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.761004] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.761066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.761316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.761379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.761679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.761745] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.761951] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.762014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.762253] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.762315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.762562] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.762625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.762903] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.762969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.763212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.763275] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.763517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.763581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.763853] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.763919] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.764207] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.764281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.764533] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.764597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.764831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.764896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.765181] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.765243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.133 [2024-12-06 17:54:27.765536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.133 [2024-12-06 17:54:27.765599] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.133 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.765925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.765990] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.766296] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.766359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.766601] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.766682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.766945] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.767009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.767256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.767320] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.767568] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.767632] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.767901] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.767964] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.768208] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.768271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.768554] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.768617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.768896] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.768958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.769197] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.769260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.769506] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.769572] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.769880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.769944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.770225] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.770289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.770488] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.770552] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.770801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.770867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.771076] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.771140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.771424] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.771488] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.771746] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.771810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.772093] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.772156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.772453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.772517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.772810] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.772875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.773146] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.773210] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.773453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.773517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.773806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.773870] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.774110] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.774174] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.774416] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.774480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.774754] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.774818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.775081] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.775144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.775392] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.775458] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.775715] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.775780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.776038] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.776101] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.776296] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.776360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.776584] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.776647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.776913] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.776977] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.777261] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.777336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.134 qpair failed and we were unable to recover it. 00:38:46.134 [2024-12-06 17:54:27.777589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.134 [2024-12-06 17:54:27.777655] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.777921] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.777986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.778268] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.778331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.778581] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.778645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.778953] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.779017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.779263] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.779326] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.779616] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.779713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.779964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.780028] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.780317] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.780380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.780636] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.780720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.781023] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.781088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.781377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.781440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.781698] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.781764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.781959] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.782025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.782266] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.782329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.782586] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.782649] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.782875] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.782938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.783200] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.783264] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.783558] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.783621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.783888] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.783951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.784244] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.784307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.784545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.784608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.784876] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.784940] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.785174] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.785237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.785525] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.785588] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.785874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.785939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.786195] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.786259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.786499] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.786563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.786868] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.786933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.787159] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.787223] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.787467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.787530] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.787816] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.787881] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.788077] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.788141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.788396] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.135 [2024-12-06 17:54:27.788459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.135 qpair failed and we were unable to recover it. 00:38:46.135 [2024-12-06 17:54:27.788769] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.788833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.789090] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.789155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.789443] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.789507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.789752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.789816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.790075] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.790139] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.790350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.790424] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.790636] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.790713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.790958] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.791023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.791267] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.791331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.791587] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.791650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.791969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.792033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.792285] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.792349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.792635] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.792717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.792904] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.792967] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.793261] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.793325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.793523] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.793589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.793917] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.793982] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.794249] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.794312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.794553] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.794616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.794874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.794938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.795191] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.795254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.795498] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.795561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.795789] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.795854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.796102] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.796164] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.796416] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.796481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.796697] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.796764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.797008] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.797071] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.797375] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.797439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.797731] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.797796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.798111] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.798175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.798433] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.798497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.798742] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.798809] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.799071] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.799135] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.136 qpair failed and we were unable to recover it. 00:38:46.136 [2024-12-06 17:54:27.799429] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.136 [2024-12-06 17:54:27.799492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.137 qpair failed and we were unable to recover it. 00:38:46.137 [2024-12-06 17:54:27.799698] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.137 [2024-12-06 17:54:27.799763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.137 qpair failed and we were unable to recover it. 00:38:46.137 [2024-12-06 17:54:27.800001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.137 [2024-12-06 17:54:27.800066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.137 qpair failed and we were unable to recover it. 00:38:46.137 [2024-12-06 17:54:27.800273] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.137 [2024-12-06 17:54:27.800337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.137 qpair failed and we were unable to recover it. 00:38:46.137 [2024-12-06 17:54:27.800616] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.137 [2024-12-06 17:54:27.800699] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.137 qpair failed and we were unable to recover it. 00:38:46.137 [2024-12-06 17:54:27.800951] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.137 [2024-12-06 17:54:27.801014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.137 qpair failed and we were unable to recover it. 00:38:46.137 [2024-12-06 17:54:27.801256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.137 [2024-12-06 17:54:27.801319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.137 qpair failed and we were unable to recover it. 00:38:46.137 [2024-12-06 17:54:27.801539] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.137 [2024-12-06 17:54:27.801602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.137 qpair failed and we were unable to recover it. 00:38:46.137 [2024-12-06 17:54:27.801866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.137 [2024-12-06 17:54:27.801930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.137 qpair failed and we were unable to recover it. 00:38:46.137 [2024-12-06 17:54:27.802223] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.137 [2024-12-06 17:54:27.802286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.137 qpair failed and we were unable to recover it. 00:38:46.137 [2024-12-06 17:54:27.802487] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.137 [2024-12-06 17:54:27.802551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.137 qpair failed and we were unable to recover it. 00:38:46.137 [2024-12-06 17:54:27.802801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.137 [2024-12-06 17:54:27.802870] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.137 qpair failed and we were unable to recover it. 00:38:46.137 [2024-12-06 17:54:27.803173] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.137 [2024-12-06 17:54:27.803248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.137 qpair failed and we were unable to recover it. 00:38:46.137 [2024-12-06 17:54:27.803518] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.137 [2024-12-06 17:54:27.803583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.137 qpair failed and we were unable to recover it. 00:38:46.137 [2024-12-06 17:54:27.803890] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.137 [2024-12-06 17:54:27.803954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.137 qpair failed and we were unable to recover it. 00:38:46.137 [2024-12-06 17:54:27.804251] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.137 [2024-12-06 17:54:27.804324] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.137 qpair failed and we were unable to recover it. 00:38:46.137 [2024-12-06 17:54:27.804548] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.137 [2024-12-06 17:54:27.804618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.137 qpair failed and we were unable to recover it. 00:38:46.137 [2024-12-06 17:54:27.804895] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.137 [2024-12-06 17:54:27.804959] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.137 qpair failed and we were unable to recover it. 00:38:46.137 [2024-12-06 17:54:27.805208] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.137 [2024-12-06 17:54:27.805271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.137 qpair failed and we were unable to recover it. 00:38:46.137 [2024-12-06 17:54:27.805473] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.137 [2024-12-06 17:54:27.805536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.137 qpair failed and we were unable to recover it. 00:38:46.137 [2024-12-06 17:54:27.805783] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.137 [2024-12-06 17:54:27.805854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.137 qpair failed and we were unable to recover it. 00:38:46.137 [2024-12-06 17:54:27.806092] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.137 [2024-12-06 17:54:27.806163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.137 qpair failed and we were unable to recover it. 00:38:46.137 [2024-12-06 17:54:27.806423] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.137 [2024-12-06 17:54:27.806486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.137 qpair failed and we were unable to recover it. 00:38:46.137 [2024-12-06 17:54:27.806765] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.137 [2024-12-06 17:54:27.806829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.137 qpair failed and we were unable to recover it. 00:38:46.137 [2024-12-06 17:54:27.807121] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.137 [2024-12-06 17:54:27.807184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.137 qpair failed and we were unable to recover it. 00:38:46.137 [2024-12-06 17:54:27.807440] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.137 [2024-12-06 17:54:27.807504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.137 qpair failed and we were unable to recover it. 00:38:46.137 [2024-12-06 17:54:27.807772] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.137 [2024-12-06 17:54:27.807838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.137 qpair failed and we were unable to recover it. 00:38:46.137 [2024-12-06 17:54:27.808079] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.138 [2024-12-06 17:54:27.808143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.138 qpair failed and we were unable to recover it. 00:38:46.138 [2024-12-06 17:54:27.808404] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.138 [2024-12-06 17:54:27.808466] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.138 qpair failed and we were unable to recover it. 00:38:46.138 [2024-12-06 17:54:27.808747] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.138 [2024-12-06 17:54:27.808818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.138 qpair failed and we were unable to recover it. 00:38:46.138 [2024-12-06 17:54:27.809108] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.138 [2024-12-06 17:54:27.809178] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.138 qpair failed and we were unable to recover it. 00:38:46.138 [2024-12-06 17:54:27.809463] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.138 [2024-12-06 17:54:27.809525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.138 qpair failed and we were unable to recover it. 00:38:46.138 [2024-12-06 17:54:27.809753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.138 [2024-12-06 17:54:27.809819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.138 qpair failed and we were unable to recover it. 00:38:46.138 [2024-12-06 17:54:27.810066] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.138 [2024-12-06 17:54:27.810130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.138 qpair failed and we were unable to recover it. 00:38:46.138 [2024-12-06 17:54:27.810405] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.138 [2024-12-06 17:54:27.810470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.138 qpair failed and we were unable to recover it. 00:38:46.138 [2024-12-06 17:54:27.810734] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.138 [2024-12-06 17:54:27.810800] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.138 qpair failed and we were unable to recover it. 00:38:46.138 [2024-12-06 17:54:27.811083] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.138 [2024-12-06 17:54:27.811147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.138 qpair failed and we were unable to recover it. 00:38:46.138 [2024-12-06 17:54:27.811356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.138 [2024-12-06 17:54:27.811421] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.138 qpair failed and we were unable to recover it. 00:38:46.138 [2024-12-06 17:54:27.811718] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.138 [2024-12-06 17:54:27.811783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.138 qpair failed and we were unable to recover it. 00:38:46.138 [2024-12-06 17:54:27.812119] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.138 [2024-12-06 17:54:27.812218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.138 qpair failed and we were unable to recover it. 00:38:46.138 [2024-12-06 17:54:27.812498] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.138 [2024-12-06 17:54:27.812568] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.138 qpair failed and we were unable to recover it. 00:38:46.138 [2024-12-06 17:54:27.812831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.138 [2024-12-06 17:54:27.812901] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.138 qpair failed and we were unable to recover it. 00:38:46.138 [2024-12-06 17:54:27.813202] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.138 [2024-12-06 17:54:27.813266] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.138 qpair failed and we were unable to recover it. 00:38:46.138 [2024-12-06 17:54:27.813517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.138 [2024-12-06 17:54:27.813582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.138 qpair failed and we were unable to recover it. 00:38:46.138 [2024-12-06 17:54:27.813821] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.138 [2024-12-06 17:54:27.813888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.138 qpair failed and we were unable to recover it. 00:38:46.138 [2024-12-06 17:54:27.814133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.138 [2024-12-06 17:54:27.814198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.138 qpair failed and we were unable to recover it. 00:38:46.138 [2024-12-06 17:54:27.814443] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.138 [2024-12-06 17:54:27.814508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.138 qpair failed and we were unable to recover it. 00:38:46.138 [2024-12-06 17:54:27.814708] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.138 [2024-12-06 17:54:27.814776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.138 qpair failed and we were unable to recover it. 00:38:46.138 [2024-12-06 17:54:27.815016] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.138 [2024-12-06 17:54:27.815079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.138 qpair failed and we were unable to recover it. 00:38:46.138 [2024-12-06 17:54:27.815329] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.138 [2024-12-06 17:54:27.815393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.138 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.815628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.815708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.815969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.816034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.816318] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.816382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.816692] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.816758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.817018] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.817083] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.817316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.817384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.817686] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.817751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.817975] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.818039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.818290] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.818355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.818612] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.818695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.818924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.818989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.819185] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.819250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.819450] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.819513] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.819740] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.819806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.820064] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.820128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.820365] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.820428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.820728] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.820805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.821063] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.821128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.821372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.821437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.821663] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.821740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.822028] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.822093] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.822388] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.822452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.822747] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.822812] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.823064] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.823132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.823390] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.823454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.823745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.823811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.824032] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.824097] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.824353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.824418] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.139 [2024-12-06 17:54:27.824681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.139 [2024-12-06 17:54:27.824746] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.139 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.824996] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.825062] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.825297] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.825362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.825610] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.825696] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.825951] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.826016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.826231] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.826295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.826544] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.826609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.826866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.826929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.827145] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.827207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.827430] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.827491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.827762] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.827829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.828069] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.828130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.828413] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.828478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.828713] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.828781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.829032] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.829096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.829388] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.829464] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.829762] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.829829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.830081] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.830145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.830403] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.830468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.830725] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.830791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.831042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.831107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.831307] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.831371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.831589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.831654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.831912] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.831978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.832263] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.832328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.832583] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.832650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.832960] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.833027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.833251] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.833316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.833519] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.833585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.140 [2024-12-06 17:54:27.833891] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.140 [2024-12-06 17:54:27.833993] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.140 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.834271] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.834339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.834635] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.834717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.835014] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.835081] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.835336] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.835405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.835724] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.835792] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.836045] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.836111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.836327] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.836396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.836630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.836711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.836921] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.836989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.837246] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.837310] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.837573] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.837638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.837915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.837981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.838283] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.838348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.838575] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.838642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.838933] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.838999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.839258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.839323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.839584] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.839649] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.839982] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.840049] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.840350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.840415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.840689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.840757] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.841016] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.841082] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.841340] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.841405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.841656] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.841741] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.842009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.842075] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.842371] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.842436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.842731] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.842797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.843054] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.843134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.843426] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.843492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.843710] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.843776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.843980] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.844046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.844358] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.844423] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.844686] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.844752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.845046] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.845112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.845360] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.845426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.845684] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.845751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.846008] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.846074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.846333] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.846399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.141 [2024-12-06 17:54:27.846696] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.141 [2024-12-06 17:54:27.846763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.141 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.847020] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.847086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.847338] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.847405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.847711] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.847778] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.848035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.848102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.848343] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.848409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.848681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.848749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.848979] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.849045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.849333] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.849399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.849649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.849733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.849988] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.850053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.850340] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.850406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.850647] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.850731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.850977] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.851042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.851253] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.851318] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.851554] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.851619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.851909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.851985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.852288] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.852354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.852592] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.852658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.852940] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.853005] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.853277] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.853342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.853604] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.853691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.853948] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.854013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.854225] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.854291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.854492] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.854559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.854807] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.854874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.855075] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.855141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.855391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.855455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.855713] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.855781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.856009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.856075] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.856418] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.856524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.856789] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.142 [2024-12-06 17:54:27.856861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.142 qpair failed and we were unable to recover it. 00:38:46.142 [2024-12-06 17:54:27.857161] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.143 [2024-12-06 17:54:27.857229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.143 qpair failed and we were unable to recover it. 00:38:46.143 [2024-12-06 17:54:27.857516] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.143 [2024-12-06 17:54:27.857581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.143 qpair failed and we were unable to recover it. 00:38:46.143 [2024-12-06 17:54:27.857860] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.143 [2024-12-06 17:54:27.857928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.143 qpair failed and we were unable to recover it. 00:38:46.143 [2024-12-06 17:54:27.858172] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.143 [2024-12-06 17:54:27.858238] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.143 qpair failed and we were unable to recover it. 00:38:46.143 [2024-12-06 17:54:27.858529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.143 [2024-12-06 17:54:27.858594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.143 qpair failed and we were unable to recover it. 00:38:46.143 [2024-12-06 17:54:27.858920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.143 [2024-12-06 17:54:27.858986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.143 qpair failed and we were unable to recover it. 00:38:46.143 [2024-12-06 17:54:27.859271] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.143 [2024-12-06 17:54:27.859336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.143 qpair failed and we were unable to recover it. 00:38:46.143 [2024-12-06 17:54:27.859593] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.143 [2024-12-06 17:54:27.859658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.143 qpair failed and we were unable to recover it. 00:38:46.143 [2024-12-06 17:54:27.859905] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.143 [2024-12-06 17:54:27.859970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.143 qpair failed and we were unable to recover it. 00:38:46.143 [2024-12-06 17:54:27.860227] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.143 [2024-12-06 17:54:27.860293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.143 qpair failed and we were unable to recover it. 00:38:46.143 [2024-12-06 17:54:27.860591] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.143 [2024-12-06 17:54:27.860655] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.143 qpair failed and we were unable to recover it. 00:38:46.143 [2024-12-06 17:54:27.860972] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.143 [2024-12-06 17:54:27.861037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.143 qpair failed and we were unable to recover it. 00:38:46.143 [2024-12-06 17:54:27.861334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.143 [2024-12-06 17:54:27.861400] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.143 qpair failed and we were unable to recover it. 00:38:46.143 [2024-12-06 17:54:27.861592] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.143 [2024-12-06 17:54:27.861656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.143 qpair failed and we were unable to recover it. 00:38:46.143 [2024-12-06 17:54:27.861898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.143 [2024-12-06 17:54:27.861963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.143 qpair failed and we were unable to recover it. 00:38:46.143 [2024-12-06 17:54:27.862211] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.143 [2024-12-06 17:54:27.862276] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.143 qpair failed and we were unable to recover it. 00:38:46.143 [2024-12-06 17:54:27.862511] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.143 [2024-12-06 17:54:27.862575] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.143 qpair failed and we were unable to recover it. 00:38:46.143 [2024-12-06 17:54:27.862810] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.143 [2024-12-06 17:54:27.862877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.143 qpair failed and we were unable to recover it. 00:38:46.143 [2024-12-06 17:54:27.863097] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.143 [2024-12-06 17:54:27.863162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.143 qpair failed and we were unable to recover it. 00:38:46.143 [2024-12-06 17:54:27.863426] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.143 [2024-12-06 17:54:27.863490] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.143 qpair failed and we were unable to recover it. 00:38:46.143 [2024-12-06 17:54:27.863744] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.143 [2024-12-06 17:54:27.863814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.143 qpair failed and we were unable to recover it. 00:38:46.143 [2024-12-06 17:54:27.864068] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.143 [2024-12-06 17:54:27.864133] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.143 qpair failed and we were unable to recover it. 00:38:46.143 [2024-12-06 17:54:27.864352] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.143 [2024-12-06 17:54:27.864418] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.143 qpair failed and we were unable to recover it. 00:38:46.143 [2024-12-06 17:54:27.864642] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.143 [2024-12-06 17:54:27.864740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.143 qpair failed and we were unable to recover it. 00:38:46.143 [2024-12-06 17:54:27.864952] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.143 [2024-12-06 17:54:27.865017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.143 qpair failed and we were unable to recover it. 00:38:46.143 [2024-12-06 17:54:27.865300] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.143 [2024-12-06 17:54:27.865376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.143 qpair failed and we were unable to recover it. 00:38:46.143 [2024-12-06 17:54:27.865629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.143 [2024-12-06 17:54:27.865710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.143 qpair failed and we were unable to recover it. 00:38:46.143 [2024-12-06 17:54:27.865951] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.143 [2024-12-06 17:54:27.866017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.143 qpair failed and we were unable to recover it. 00:38:46.143 [2024-12-06 17:54:27.866272] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.144 [2024-12-06 17:54:27.866338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.144 qpair failed and we were unable to recover it. 00:38:46.144 [2024-12-06 17:54:27.866546] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.144 [2024-12-06 17:54:27.866611] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.144 qpair failed and we were unable to recover it. 00:38:46.144 [2024-12-06 17:54:27.866912] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.144 [2024-12-06 17:54:27.867011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.144 qpair failed and we were unable to recover it. 00:38:46.144 [2024-12-06 17:54:27.867294] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.144 [2024-12-06 17:54:27.867369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.144 qpair failed and we were unable to recover it. 00:38:46.144 [2024-12-06 17:54:27.867625] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.144 [2024-12-06 17:54:27.867712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.144 qpair failed and we were unable to recover it. 00:38:46.144 [2024-12-06 17:54:27.867951] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.144 [2024-12-06 17:54:27.868017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.144 qpair failed and we were unable to recover it. 00:38:46.144 [2024-12-06 17:54:27.868311] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.144 [2024-12-06 17:54:27.868379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.144 qpair failed and we were unable to recover it. 00:38:46.144 [2024-12-06 17:54:27.868641] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.144 [2024-12-06 17:54:27.868722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.144 qpair failed and we were unable to recover it. 00:38:46.144 [2024-12-06 17:54:27.868984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.144 [2024-12-06 17:54:27.869051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.144 qpair failed and we were unable to recover it. 00:38:46.144 [2024-12-06 17:54:27.869284] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.144 [2024-12-06 17:54:27.869351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.144 qpair failed and we were unable to recover it. 00:38:46.144 [2024-12-06 17:54:27.869643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.144 [2024-12-06 17:54:27.869724] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.144 qpair failed and we were unable to recover it. 00:38:46.144 [2024-12-06 17:54:27.869987] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.144 [2024-12-06 17:54:27.870052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.144 qpair failed and we were unable to recover it. 00:38:46.144 [2024-12-06 17:54:27.870272] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.144 [2024-12-06 17:54:27.870337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.144 qpair failed and we were unable to recover it. 00:38:46.144 [2024-12-06 17:54:27.870602] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.144 [2024-12-06 17:54:27.870680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.144 qpair failed and we were unable to recover it. 00:38:46.144 [2024-12-06 17:54:27.870905] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.144 [2024-12-06 17:54:27.870970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.144 qpair failed and we were unable to recover it. 00:38:46.144 [2024-12-06 17:54:27.871229] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.144 [2024-12-06 17:54:27.871294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.144 qpair failed and we were unable to recover it. 00:38:46.144 [2024-12-06 17:54:27.871516] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.144 [2024-12-06 17:54:27.871581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.144 qpair failed and we were unable to recover it. 00:38:46.144 [2024-12-06 17:54:27.871878] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.144 [2024-12-06 17:54:27.871945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.144 qpair failed and we were unable to recover it. 00:38:46.144 [2024-12-06 17:54:27.872241] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.144 [2024-12-06 17:54:27.872306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.144 qpair failed and we were unable to recover it. 00:38:46.144 [2024-12-06 17:54:27.872560] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.144 [2024-12-06 17:54:27.872626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.144 qpair failed and we were unable to recover it. 00:38:46.144 [2024-12-06 17:54:27.872903] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.144 [2024-12-06 17:54:27.872969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.144 qpair failed and we were unable to recover it. 00:38:46.144 [2024-12-06 17:54:27.873208] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.144 [2024-12-06 17:54:27.873274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.144 qpair failed and we were unable to recover it. 00:38:46.144 [2024-12-06 17:54:27.873522] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.144 [2024-12-06 17:54:27.873588] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.144 qpair failed and we were unable to recover it. 00:38:46.144 [2024-12-06 17:54:27.873862] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.144 [2024-12-06 17:54:27.873928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.144 qpair failed and we were unable to recover it. 00:38:46.144 [2024-12-06 17:54:27.874225] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.144 [2024-12-06 17:54:27.874301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.144 qpair failed and we were unable to recover it. 00:38:46.144 [2024-12-06 17:54:27.874599] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.144 [2024-12-06 17:54:27.874683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.144 qpair failed and we were unable to recover it. 00:38:46.144 [2024-12-06 17:54:27.874937] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.145 [2024-12-06 17:54:27.875003] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.145 qpair failed and we were unable to recover it. 00:38:46.145 [2024-12-06 17:54:27.875272] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.145 [2024-12-06 17:54:27.875338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.145 qpair failed and we were unable to recover it. 00:38:46.145 [2024-12-06 17:54:27.875632] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.145 [2024-12-06 17:54:27.875716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.145 qpair failed and we were unable to recover it. 00:38:46.145 [2024-12-06 17:54:27.876021] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.145 [2024-12-06 17:54:27.876086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.145 qpair failed and we were unable to recover it. 00:38:46.145 [2024-12-06 17:54:27.876348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.145 [2024-12-06 17:54:27.876413] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.145 qpair failed and we were unable to recover it. 00:38:46.145 [2024-12-06 17:54:27.876694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.145 [2024-12-06 17:54:27.876761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.145 qpair failed and we were unable to recover it. 00:38:46.145 [2024-12-06 17:54:27.877063] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.145 [2024-12-06 17:54:27.877129] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.145 qpair failed and we were unable to recover it. 00:38:46.145 [2024-12-06 17:54:27.877372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.145 [2024-12-06 17:54:27.877437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.145 qpair failed and we were unable to recover it. 00:38:46.145 [2024-12-06 17:54:27.877641] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.145 [2024-12-06 17:54:27.877728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.145 qpair failed and we were unable to recover it. 00:38:46.145 [2024-12-06 17:54:27.878004] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.145 [2024-12-06 17:54:27.878069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.145 qpair failed and we were unable to recover it. 00:38:46.145 [2024-12-06 17:54:27.878365] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.145 [2024-12-06 17:54:27.878432] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.145 qpair failed and we were unable to recover it. 00:38:46.145 [2024-12-06 17:54:27.878621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.145 [2024-12-06 17:54:27.878706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.145 qpair failed and we were unable to recover it. 00:38:46.145 [2024-12-06 17:54:27.878979] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.145 [2024-12-06 17:54:27.879045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.145 qpair failed and we were unable to recover it. 00:38:46.145 [2024-12-06 17:54:27.879338] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.145 [2024-12-06 17:54:27.879403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.145 qpair failed and we were unable to recover it. 00:38:46.145 [2024-12-06 17:54:27.879659] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.145 [2024-12-06 17:54:27.879743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.145 qpair failed and we were unable to recover it. 00:38:46.145 [2024-12-06 17:54:27.880001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.145 [2024-12-06 17:54:27.880067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.145 qpair failed and we were unable to recover it. 00:38:46.145 [2024-12-06 17:54:27.880315] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.145 [2024-12-06 17:54:27.880380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.145 qpair failed and we were unable to recover it. 00:38:46.145 [2024-12-06 17:54:27.880626] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.145 [2024-12-06 17:54:27.880726] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.145 qpair failed and we were unable to recover it. 00:38:46.145 [2024-12-06 17:54:27.881020] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.145 [2024-12-06 17:54:27.881086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.145 qpair failed and we were unable to recover it. 00:38:46.145 [2024-12-06 17:54:27.881359] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.145 [2024-12-06 17:54:27.881425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.145 qpair failed and we were unable to recover it. 00:38:46.145 [2024-12-06 17:54:27.881718] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.145 [2024-12-06 17:54:27.881785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.145 qpair failed and we were unable to recover it. 00:38:46.145 [2024-12-06 17:54:27.882094] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.145 [2024-12-06 17:54:27.882160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.145 qpair failed and we were unable to recover it. 00:38:46.145 [2024-12-06 17:54:27.882410] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.145 [2024-12-06 17:54:27.882476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.145 qpair failed and we were unable to recover it. 00:38:46.145 [2024-12-06 17:54:27.882735] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.145 [2024-12-06 17:54:27.882801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.145 qpair failed and we were unable to recover it. 00:38:46.145 [2024-12-06 17:54:27.883020] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.145 [2024-12-06 17:54:27.883086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.145 qpair failed and we were unable to recover it. 00:38:46.145 [2024-12-06 17:54:27.883322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.145 [2024-12-06 17:54:27.883388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.145 qpair failed and we were unable to recover it. 00:38:46.145 [2024-12-06 17:54:27.883692] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.145 [2024-12-06 17:54:27.883758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.145 qpair failed and we were unable to recover it. 00:38:46.145 [2024-12-06 17:54:27.884014] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.145 [2024-12-06 17:54:27.884080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.146 qpair failed and we were unable to recover it. 00:38:46.146 [2024-12-06 17:54:27.884369] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.146 [2024-12-06 17:54:27.884435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.146 qpair failed and we were unable to recover it. 00:38:46.146 [2024-12-06 17:54:27.884693] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.146 [2024-12-06 17:54:27.884760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.146 qpair failed and we were unable to recover it. 00:38:46.146 [2024-12-06 17:54:27.885017] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.146 [2024-12-06 17:54:27.885085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.146 qpair failed and we were unable to recover it. 00:38:46.146 [2024-12-06 17:54:27.885337] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.146 [2024-12-06 17:54:27.885404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.146 qpair failed and we were unable to recover it. 00:38:46.146 [2024-12-06 17:54:27.885705] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.146 [2024-12-06 17:54:27.885772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.146 qpair failed and we were unable to recover it. 00:38:46.146 [2024-12-06 17:54:27.886032] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.146 [2024-12-06 17:54:27.886098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.146 qpair failed and we were unable to recover it. 00:38:46.146 [2024-12-06 17:54:27.886348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.146 [2024-12-06 17:54:27.886418] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.146 qpair failed and we were unable to recover it. 00:38:46.146 [2024-12-06 17:54:27.886728] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.146 [2024-12-06 17:54:27.886794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.146 qpair failed and we were unable to recover it. 00:38:46.146 [2024-12-06 17:54:27.887053] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.146 [2024-12-06 17:54:27.887119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.146 qpair failed and we were unable to recover it. 00:38:46.146 [2024-12-06 17:54:27.887374] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.146 [2024-12-06 17:54:27.887442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.146 qpair failed and we were unable to recover it. 00:38:46.146 [2024-12-06 17:54:27.887706] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.146 [2024-12-06 17:54:27.887773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.146 qpair failed and we were unable to recover it. 00:38:46.146 [2024-12-06 17:54:27.888095] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.146 [2024-12-06 17:54:27.888160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.146 qpair failed and we were unable to recover it. 00:38:46.146 [2024-12-06 17:54:27.888417] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.146 [2024-12-06 17:54:27.888483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.146 qpair failed and we were unable to recover it. 00:38:46.146 [2024-12-06 17:54:27.888722] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.146 [2024-12-06 17:54:27.888789] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.146 qpair failed and we were unable to recover it. 00:38:46.146 [2024-12-06 17:54:27.889022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.146 [2024-12-06 17:54:27.889087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.146 qpair failed and we were unable to recover it. 00:38:46.146 [2024-12-06 17:54:27.889384] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.146 [2024-12-06 17:54:27.889449] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.146 qpair failed and we were unable to recover it. 00:38:46.146 [2024-12-06 17:54:27.889716] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.146 [2024-12-06 17:54:27.889782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.146 qpair failed and we were unable to recover it. 00:38:46.146 [2024-12-06 17:54:27.890071] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.146 [2024-12-06 17:54:27.890136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.146 qpair failed and we were unable to recover it. 00:38:46.146 [2024-12-06 17:54:27.890327] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.146 [2024-12-06 17:54:27.890392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.146 qpair failed and we were unable to recover it. 00:38:46.146 [2024-12-06 17:54:27.890612] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.146 [2024-12-06 17:54:27.890699] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.146 qpair failed and we were unable to recover it. 00:38:46.146 [2024-12-06 17:54:27.891001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.146 [2024-12-06 17:54:27.891067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.146 qpair failed and we were unable to recover it. 00:38:46.146 [2024-12-06 17:54:27.891276] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.146 [2024-12-06 17:54:27.891341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.146 qpair failed and we were unable to recover it. 00:38:46.146 [2024-12-06 17:54:27.891650] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.146 [2024-12-06 17:54:27.891738] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.146 qpair failed and we were unable to recover it. 00:38:46.146 [2024-12-06 17:54:27.892046] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.146 [2024-12-06 17:54:27.892112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.146 qpair failed and we were unable to recover it. 00:38:46.146 [2024-12-06 17:54:27.892409] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.146 [2024-12-06 17:54:27.892484] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.146 qpair failed and we were unable to recover it. 00:38:46.146 [2024-12-06 17:54:27.892727] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.146 [2024-12-06 17:54:27.892793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.146 qpair failed and we were unable to recover it. 00:38:46.146 [2024-12-06 17:54:27.893041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.146 [2024-12-06 17:54:27.893107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.146 qpair failed and we were unable to recover it. 00:38:46.146 [2024-12-06 17:54:27.893349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.893417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.893708] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.893775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.894035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.894100] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.894361] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.894425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.894691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.894758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.895056] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.895120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.895378] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.895443] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.895731] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.895798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.896095] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.896160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.896414] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.896479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.896774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.896841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.897105] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.897180] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.897469] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.897534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.897828] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.897895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.898103] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.898167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.898421] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.898486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.898780] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.898847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.899149] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.899212] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.899515] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.899580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.899839] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.899904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.900194] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.900258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.900520] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.900585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.900905] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.900981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.901190] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.901254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.901515] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.901580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.901909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.901976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.902223] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.902294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.902578] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.902643] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.902926] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.902996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.903289] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.903353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.903598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.903679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.903993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.904058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.904350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.904423] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.904724] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.904791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.905087] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.905151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.905444] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.905510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.905774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.905841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.906107] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.147 [2024-12-06 17:54:27.906172] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.147 qpair failed and we were unable to recover it. 00:38:46.147 [2024-12-06 17:54:27.906427] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.906491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.906871] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.906939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.907195] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.907260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.907547] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.907611] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.907884] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.907951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.908206] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.908270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.908573] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.908639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.908932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.908999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.909292] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.909357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.909644] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.909730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.910030] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.910095] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.910390] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.910455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.910746] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.910814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.911119] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.911185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.911444] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.911519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.911822] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.911889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.912134] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.912199] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.912420] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.912485] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.912785] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.912851] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.913134] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.913199] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.913456] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.913520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.913823] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.913890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.914129] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.914194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.914438] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.914502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.914752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.914819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.915030] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.915098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.915320] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.915385] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.915687] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.915753] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.916062] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.148 [2024-12-06 17:54:27.916126] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.148 qpair failed and we were unable to recover it. 00:38:46.148 [2024-12-06 17:54:27.916391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.916456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.916715] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.916782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.917021] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.917087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.917324] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.917389] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.917690] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.917757] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.918045] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.918111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.918358] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.918425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.918731] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.918820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.919135] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.919200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.919482] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.919546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.919747] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.919815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.920115] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.920180] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.920438] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.920513] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.920772] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.920839] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.921145] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.921210] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.921479] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.921544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.921832] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.921898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.922161] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.922225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.922511] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.922575] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.922894] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.922960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.923179] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.923243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.923538] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.923604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.923931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.923999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.924290] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.924354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.924654] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.924755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.925046] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.925111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.925418] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.925483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.925744] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.925809] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.926008] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.926072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.926364] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.926429] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.926746] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.926811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.149 qpair failed and we were unable to recover it. 00:38:46.149 [2024-12-06 17:54:27.927063] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.149 [2024-12-06 17:54:27.927130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.927386] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.927451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.927745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.927811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.928102] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.928167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.928453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.928520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.928817] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.928883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.929174] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.929238] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.929527] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.929591] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.929911] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.929977] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.930293] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.930358] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.930645] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.930729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.931020] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.931085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.931380] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.931445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.931742] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.931807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.932064] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.932132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.932389] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.932454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.932729] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.932796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.933022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.933088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.933360] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.933424] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.933698] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.933777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.934069] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.934145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.934405] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.934470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.934763] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.934840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.935138] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.935203] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.935463] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.935529] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.935770] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.935837] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.936077] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.936149] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.936442] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.936507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.936773] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.936839] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.937067] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.937131] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.937412] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.937477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.937768] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.937835] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.938141] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.938205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.938484] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.938550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.938795] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.938862] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.150 qpair failed and we were unable to recover it. 00:38:46.150 [2024-12-06 17:54:27.939117] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.150 [2024-12-06 17:54:27.939182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.151 qpair failed and we were unable to recover it. 00:38:46.151 [2024-12-06 17:54:27.939448] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.151 [2024-12-06 17:54:27.939512] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.151 qpair failed and we were unable to recover it. 00:38:46.151 [2024-12-06 17:54:27.939812] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.151 [2024-12-06 17:54:27.939879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.151 qpair failed and we were unable to recover it. 00:38:46.151 [2024-12-06 17:54:27.940131] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.151 [2024-12-06 17:54:27.940194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.151 qpair failed and we were unable to recover it. 00:38:46.151 [2024-12-06 17:54:27.940456] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.151 [2024-12-06 17:54:27.940521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.151 qpair failed and we were unable to recover it. 00:38:46.151 [2024-12-06 17:54:27.940824] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.151 [2024-12-06 17:54:27.940889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.151 qpair failed and we were unable to recover it. 00:38:46.151 [2024-12-06 17:54:27.941181] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.151 [2024-12-06 17:54:27.941246] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.151 qpair failed and we were unable to recover it. 00:38:46.151 [2024-12-06 17:54:27.941541] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.151 [2024-12-06 17:54:27.941605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.151 qpair failed and we were unable to recover it. 00:38:46.151 [2024-12-06 17:54:27.941824] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.151 [2024-12-06 17:54:27.941891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.151 qpair failed and we were unable to recover it. 00:38:46.151 [2024-12-06 17:54:27.942209] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.151 [2024-12-06 17:54:27.942281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.151 qpair failed and we were unable to recover it. 00:38:46.151 [2024-12-06 17:54:27.942530] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.151 [2024-12-06 17:54:27.942595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.151 qpair failed and we were unable to recover it. 00:38:46.151 [2024-12-06 17:54:27.942879] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.151 [2024-12-06 17:54:27.942948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.151 qpair failed and we were unable to recover it. 00:38:46.151 [2024-12-06 17:54:27.943234] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.151 [2024-12-06 17:54:27.943303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.151 qpair failed and we were unable to recover it. 00:38:46.151 [2024-12-06 17:54:27.943557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.151 [2024-12-06 17:54:27.943622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.151 qpair failed and we were unable to recover it. 00:38:46.151 [2024-12-06 17:54:27.943903] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.151 [2024-12-06 17:54:27.943980] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.151 qpair failed and we were unable to recover it. 00:38:46.151 [2024-12-06 17:54:27.944282] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.151 [2024-12-06 17:54:27.944347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.151 qpair failed and we were unable to recover it. 00:38:46.151 [2024-12-06 17:54:27.944708] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.151 [2024-12-06 17:54:27.944796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.151 qpair failed and we were unable to recover it. 00:38:46.151 [2024-12-06 17:54:27.945071] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.151 [2024-12-06 17:54:27.945136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.151 qpair failed and we were unable to recover it. 00:38:46.151 [2024-12-06 17:54:27.945398] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.151 [2024-12-06 17:54:27.945475] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.151 qpair failed and we were unable to recover it. 00:38:46.151 [2024-12-06 17:54:27.945814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.151 [2024-12-06 17:54:27.945883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.151 qpair failed and we were unable to recover it. 00:38:46.151 [2024-12-06 17:54:27.946186] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.151 [2024-12-06 17:54:27.946252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.151 qpair failed and we were unable to recover it. 00:38:46.151 [2024-12-06 17:54:27.946540] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.151 [2024-12-06 17:54:27.946606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.151 qpair failed and we were unable to recover it. 00:38:46.151 [2024-12-06 17:54:27.946882] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.151 [2024-12-06 17:54:27.946969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.151 qpair failed and we were unable to recover it. 00:38:46.151 [2024-12-06 17:54:27.947254] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.151 [2024-12-06 17:54:27.947323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.151 qpair failed and we were unable to recover it. 00:38:46.151 [2024-12-06 17:54:27.947619] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.151 [2024-12-06 17:54:27.947711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.151 qpair failed and we were unable to recover it. 00:38:46.435 [2024-12-06 17:54:27.947939] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.435 [2024-12-06 17:54:27.948004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.435 qpair failed and we were unable to recover it. 00:38:46.435 [2024-12-06 17:54:27.948295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.435 [2024-12-06 17:54:27.948360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.435 qpair failed and we were unable to recover it. 00:38:46.435 [2024-12-06 17:54:27.948661] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.435 [2024-12-06 17:54:27.948773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.435 qpair failed and we were unable to recover it. 00:38:46.435 [2024-12-06 17:54:27.948900] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x7835f0 (9): Bad file descriptor 00:38:46.435 [2024-12-06 17:54:27.949267] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.435 [2024-12-06 17:54:27.949360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.435 qpair failed and we were unable to recover it. 00:38:46.435 [2024-12-06 17:54:27.949605] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.435 [2024-12-06 17:54:27.949658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.435 qpair failed and we were unable to recover it. 00:38:46.435 [2024-12-06 17:54:27.949856] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.435 [2024-12-06 17:54:27.949891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.435 qpair failed and we were unable to recover it. 00:38:46.435 [2024-12-06 17:54:27.950016] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.435 [2024-12-06 17:54:27.950050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.435 qpair failed and we were unable to recover it. 00:38:46.435 [2024-12-06 17:54:27.950264] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.435 [2024-12-06 17:54:27.950313] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.435 qpair failed and we were unable to recover it. 00:38:46.435 [2024-12-06 17:54:27.950533] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.435 [2024-12-06 17:54:27.950596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.435 qpair failed and we were unable to recover it. 00:38:46.435 [2024-12-06 17:54:27.950817] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.435 [2024-12-06 17:54:27.950852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.435 qpair failed and we were unable to recover it. 00:38:46.435 [2024-12-06 17:54:27.950992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.435 [2024-12-06 17:54:27.951026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.435 qpair failed and we were unable to recover it. 00:38:46.435 [2024-12-06 17:54:27.951150] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.435 [2024-12-06 17:54:27.951184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.435 qpair failed and we were unable to recover it. 00:38:46.435 [2024-12-06 17:54:27.951366] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.435 [2024-12-06 17:54:27.951416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.435 qpair failed and we were unable to recover it. 00:38:46.435 [2024-12-06 17:54:27.951654] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.435 [2024-12-06 17:54:27.951736] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.435 qpair failed and we were unable to recover it. 00:38:46.435 [2024-12-06 17:54:27.951882] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.435 [2024-12-06 17:54:27.951916] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.435 qpair failed and we were unable to recover it. 00:38:46.435 [2024-12-06 17:54:27.952136] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.435 [2024-12-06 17:54:27.952201] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.435 qpair failed and we were unable to recover it. 00:38:46.435 [2024-12-06 17:54:27.952511] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.435 [2024-12-06 17:54:27.952575] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.435 qpair failed and we were unable to recover it. 00:38:46.435 [2024-12-06 17:54:27.952779] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.435 [2024-12-06 17:54:27.952814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.435 qpair failed and we were unable to recover it. 00:38:46.435 [2024-12-06 17:54:27.952932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.952967] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.953107] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.953141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.953280] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.953314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.953436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.953504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.953696] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.953740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.953854] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.953888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.954047] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.954112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.954351] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.954404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.954660] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.954741] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.954861] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.954894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.955053] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.955087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.955252] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.955292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.955397] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.955431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.955551] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.955585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.955744] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.955779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.955917] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.955953] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.956202] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.956251] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.956549] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.956597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.956770] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.956804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.956960] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.956995] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.957103] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.957136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.957285] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.957319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.957428] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.957462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.957636] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.957680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.957829] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.957863] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.957990] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.958024] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.958214] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.958249] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.958389] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.958422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.958560] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.958594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.958709] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.958744] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.958852] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.958886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.959000] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.959035] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.959223] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.959272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.959436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.959469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.959695] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.959730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.959832] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.959866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.959974] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.960010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.436 [2024-12-06 17:54:27.960150] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.436 [2024-12-06 17:54:27.960185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.436 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.960297] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.960332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.960500] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.960535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.960680] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.960719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.960863] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.960897] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.961046] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.961079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.961230] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.961296] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.961544] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.961607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.961824] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.961858] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.961978] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.962012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.962153] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.962186] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.962348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.962405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.962657] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.962741] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.962893] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.962926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.963035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.963075] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.963216] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.963261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.963397] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.963431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.963593] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.963627] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.963747] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.963774] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.963868] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.963894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.963985] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.964010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.964107] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.964132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.964302] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.964336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.964515] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.964548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.964719] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.964745] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.964830] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.964855] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.964942] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.964968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.965060] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.965107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.965246] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.965279] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.965413] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.965445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.965584] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.965609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.965736] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.965762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.965842] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.965867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.965994] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.966026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.966155] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.966186] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.966349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.966381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.966508] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.966533] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.966688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.966714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.966805] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.437 [2024-12-06 17:54:27.966830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.437 qpair failed and we were unable to recover it. 00:38:46.437 [2024-12-06 17:54:27.966930] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.966974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.967091] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.967116] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.967248] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.967281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.967418] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.967450] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.967593] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.967618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.967715] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.967741] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.967842] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.967868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.967998] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.968024] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.968114] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.968162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.968317] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.968382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.968644] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.968728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.968816] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.968842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.968990] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.969015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.969116] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.969169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.969353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.969408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.969649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.969707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.969824] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.969849] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.969973] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.969998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.970173] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.970236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.970414] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.970468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.970673] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.970720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.970836] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.970862] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.970993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.971018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.971142] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.971188] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.971364] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.971398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.971529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.971574] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.971730] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.971756] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.971876] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.971902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.971986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.972012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.972128] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.972154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.972297] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.972331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.972545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.972576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.972707] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.972749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.972840] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.972866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.973005] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.973030] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.973148] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.973194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.973329] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.973380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.973526] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.438 [2024-12-06 17:54:27.973571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.438 qpair failed and we were unable to recover it. 00:38:46.438 [2024-12-06 17:54:27.973732] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.973758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.973844] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.973870] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.974011] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.974037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.974122] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.974148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.974313] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.974346] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.974526] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.974558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.974701] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.974749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.974873] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.974898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.975018] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.975044] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.975158] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.975183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.975348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.975386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.975514] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.975556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.975686] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.975728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.975810] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.975836] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.975950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.975976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.976137] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.976170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.976352] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.976385] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.976604] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.976649] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.976766] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.976791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.976876] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.976902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.977059] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.977095] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.977221] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.977246] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.977410] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.977448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.977596] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.977623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.977730] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.977757] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.977842] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.977867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.977968] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.978020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.978208] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.978244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.978375] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.978418] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.978566] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.978592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.978680] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.978706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.978828] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.978853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.978967] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.978993] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.979075] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.979101] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.979259] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.979312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.979462] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.979496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.439 qpair failed and we were unable to recover it. 00:38:46.439 [2024-12-06 17:54:27.979623] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.439 [2024-12-06 17:54:27.979648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.979774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.979800] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.979906] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.979931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.980046] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.980071] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.980250] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.980284] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.980485] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.980521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.980682] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.980727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.980843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.980869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.980964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.980990] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.981078] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.981104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.981286] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.981323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.981469] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.981515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.981716] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.981742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.981829] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.981856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.981944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.981992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.982102] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.982135] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.982283] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.982317] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.982484] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.982516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.982653] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.982695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.982838] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.982864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.982977] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.983013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.983130] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.983174] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.983343] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.983395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.983526] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.983567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.983727] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.983755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.983841] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.983867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.983983] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.984009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.984135] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.984161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.984295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.984321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.984432] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.984457] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.984565] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.984590] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.984680] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.984707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.984841] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.984889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.985024] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.440 [2024-12-06 17:54:27.985062] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.440 qpair failed and we were unable to recover it. 00:38:46.440 [2024-12-06 17:54:27.985232] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.985259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.985378] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.985406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.985516] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.985542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.985683] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.985721] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.985833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.985859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.986001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.986027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.986167] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.986194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.986281] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.986308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.986421] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.986448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.986561] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.986587] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.986679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.986717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.986836] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.986862] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.987005] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.987031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.987147] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.987173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.987300] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.987327] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.987421] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.987447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.987538] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.987565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.987680] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.987707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.987806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.987832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.987946] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.987972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.988089] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.988114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.988227] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.988254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.988394] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.988420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.988538] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.988564] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.988680] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.988707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.988791] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.988816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.988960] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.988985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.989099] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.989129] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.989246] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.989271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.989380] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.989407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.989530] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.989556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.989636] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.989662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.989790] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.989816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.989933] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.989959] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.990045] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.990071] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.990218] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.990244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.990359] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.990385] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.990496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.441 [2024-12-06 17:54:27.990522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.441 qpair failed and we were unable to recover it. 00:38:46.441 [2024-12-06 17:54:27.990636] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.990669] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.990790] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.990817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.990906] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.990932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.991018] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.991045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.991165] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.991191] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.991271] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.991297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.991414] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.991440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.991535] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.991561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.991654] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.991687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.991831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.991857] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.991982] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.992008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.992124] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.992150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.992292] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.992317] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.992436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.992462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.992575] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.992601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.992719] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.992746] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.992837] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.992864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.992952] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.992978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.993060] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.993086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.993205] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.993231] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.993375] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.993401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.993516] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.993541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.993656] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.993688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.993767] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.993794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.993877] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.993902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.994019] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.994046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.994191] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.994217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.994335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.994360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.994470] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.994496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.994584] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.994615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.994776] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.994803] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.994892] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.994918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.995027] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.995054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.995137] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.995163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.995254] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.995280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.995352] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.995378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.995483] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.995509] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.995619] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.442 [2024-12-06 17:54:27.995645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.442 qpair failed and we were unable to recover it. 00:38:46.442 [2024-12-06 17:54:27.995746] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:27.995773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:27.995889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:27.995915] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:27.996007] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:27.996034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:27.996114] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:27.996140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:27.996260] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:27.996286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:27.996411] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:27.996438] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:27.996576] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:27.996601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:27.996740] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:27.996794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:27.996964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:27.997012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:27.997170] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:27.997196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:27.997306] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:27.997332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:27.997442] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:27.997468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:27.997585] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:27.997610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:27.997759] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:27.997808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:27.997905] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:27.997931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:27.998024] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:27.998050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:27.998164] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:27.998192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:27.998304] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:27.998331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:27.998453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:27.998492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:27.998612] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:27.998639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:27.998780] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:27.998819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:27.998976] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:27.999013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:27.999163] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:27.999200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:27.999357] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:27.999407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:27.999526] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:27.999551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:27.999674] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:27.999700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:27.999815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:27.999841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:27.999986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:28.000024] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:28.000147] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:28.000184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:28.000342] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:28.000380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:28.000492] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:28.000519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:28.000608] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:28.000638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:28.000747] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:28.000774] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:28.000857] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:28.000883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:28.000962] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:28.000988] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:28.001103] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:28.001129] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:28.001247] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:28.001273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:28.001415] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.443 [2024-12-06 17:54:28.001441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.443 qpair failed and we were unable to recover it. 00:38:46.443 [2024-12-06 17:54:28.001562] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.001588] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.001700] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.001730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.001812] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.001838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.001991] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.002017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.002105] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.002131] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.002288] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.002314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.002427] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.002452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.002555] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.002581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.002693] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.002719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.002804] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.002830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.002982] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.003008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.003119] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.003145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.003290] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.003315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.003469] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.003495] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.003615] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.003640] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.003754] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.003789] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.003948] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.003984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.004131] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.004166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.004332] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.004357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.004499] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.004525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.004607] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.004633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.004781] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.004816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.004925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.004972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.005076] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.005125] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.005261] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.005287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.005400] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.005426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.005507] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.005533] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.005679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.005705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.005822] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.444 [2024-12-06 17:54:28.005848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.444 qpair failed and we were unable to recover it. 00:38:46.444 [2024-12-06 17:54:28.005973] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.005998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.006077] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.006102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.006197] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.006223] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.006349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.006374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.006467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.006497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.006582] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.006608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.006722] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.006749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.006864] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.006890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.006995] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.007021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.007133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.007159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.007296] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.007322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.007405] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.007431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.007570] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.007596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.007694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.007721] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.007869] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.007894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.008035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.008065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.008168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.008195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.008275] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.008301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.008414] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.008441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.008558] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.008583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.008689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.008732] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.008883] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.008911] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.009060] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.009088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.009214] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.009240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.009374] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.009400] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.009480] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.009505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.009594] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.009619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.009708] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.009750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.009844] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.009870] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.009984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.010025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.010156] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.010182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.010279] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.010305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.010442] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.010468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.010606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.010633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.010752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.010778] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.010857] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.010883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.011031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.011057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.011196] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.445 [2024-12-06 17:54:28.011222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.445 qpair failed and we were unable to recover it. 00:38:46.445 [2024-12-06 17:54:28.011338] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.011364] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.011446] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.011472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.011585] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.011612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.011731] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.011758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.011872] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.011898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.012045] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.012071] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.012187] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.012217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.012334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.012360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.012442] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.012468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.012559] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.012585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.012733] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.012759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.012881] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.012907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.013035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.013061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.013149] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.013176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.013262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.013287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.013382] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.013408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.013619] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.013659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.013833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.013862] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.013974] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.014003] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.014160] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.014187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.014313] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.014340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.014461] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.014488] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.014611] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.014637] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.014783] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.014812] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.014907] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.014941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.015054] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.015081] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.015168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.015195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.015312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.015338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.015530] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.015556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.015680] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.015707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.015827] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.015854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.016044] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.016070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.016164] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.016190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.016335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.016373] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.016474] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.016503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.016647] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.016683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.016780] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.016807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.446 [2024-12-06 17:54:28.016925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.446 [2024-12-06 17:54:28.016953] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.446 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.017038] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.017064] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.017207] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.017238] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.017428] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.017459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.017595] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.017627] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.017807] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.017833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.017948] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.017973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.018089] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.018132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.018245] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.018276] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.018380] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.018406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.018547] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.018580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.018721] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.018747] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.018833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.018859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.018977] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.019004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.019149] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.019181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.019314] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.019357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.019459] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.019492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.019628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.019660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.019800] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.019826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.019946] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.019971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.020111] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.020142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.020262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.020287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.020404] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.020436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.020572] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.020604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.020749] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.020776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.020894] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.020920] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.021020] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.021052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.021160] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.021186] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.021367] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.021399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.021533] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.021566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.021685] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.021712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.021827] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.021852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.021959] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.021991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.022139] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.022164] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.022259] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.022286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.022431] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.022463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.022580] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.022610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.022743] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.022770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.447 [2024-12-06 17:54:28.022880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.447 [2024-12-06 17:54:28.022906] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.447 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.023049] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.023075] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.023189] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.023215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.023391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.023422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.023554] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.023600] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.023751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.023778] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.023896] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.023921] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.024033] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.024058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.024171] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.024198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.024334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.024366] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.024475] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.024515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.024657] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.024709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.024827] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.024853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.024992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.025018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.025132] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.025178] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.025312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.025344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.025542] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.025574] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.025722] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.025748] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.025866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.025892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.025972] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.025997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.026089] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.026114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.026240] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.026273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.026470] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.026502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.026604] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.026636] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.026791] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.026831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.027001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.027040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.027185] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.027231] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.027399] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.027431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.027587] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.027613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.027740] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.027767] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.027857] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.027883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.028052] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.028099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.028238] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.028271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.028428] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.028472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.028579] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.028604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.028761] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.028807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.028920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.028971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.029091] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.029117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.448 [2024-12-06 17:54:28.029242] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.448 [2024-12-06 17:54:28.029267] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.448 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.029388] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.029415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.029498] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.029524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.029674] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.029700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.029779] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.029805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.029896] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.029922] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.030008] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.030034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.030151] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.030179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.030294] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.030321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.030412] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.030438] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.030529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.030554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.030642] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.030675] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.030796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.030822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.030939] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.030965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.031135] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.031175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.031295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.031323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.031442] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.031470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.031561] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.031588] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.031705] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.031734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.031857] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.031884] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.031965] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.031992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.032132] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.032178] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.032301] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.032347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.032467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.032493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.032613] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.032639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.032756] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.032804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.032917] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.032943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.033060] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.033086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.033175] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.033201] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.033312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.033338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.033453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.033478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.033583] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.033608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.033727] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.033753] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.033872] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.033898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.034014] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.449 [2024-12-06 17:54:28.034039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.449 qpair failed and we were unable to recover it. 00:38:46.449 [2024-12-06 17:54:28.034150] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.034176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.034290] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.034315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.034399] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.034425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.034502] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.034528] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.034637] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.034674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.034770] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.034796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.034942] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.034971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.035074] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.035101] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.035188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.035215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.035301] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.035328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.035442] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.035469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.035558] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.035584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.035734] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.035762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.035906] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.035933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.036046] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.036073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.036207] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.036240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.036345] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.036379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.036564] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.036611] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.036733] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.036762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.036909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.036961] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.037108] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.037140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.037311] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.037343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.037450] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.037483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.037617] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.037648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.037778] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.037805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.037903] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.037928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.038019] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.038045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.038210] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.038242] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.038350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.038383] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.038518] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.038553] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.038672] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.038699] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.038787] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.038814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.038928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.038972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.039118] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.039152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.039321] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.039354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.039483] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.039516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.039623] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.039650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.450 [2024-12-06 17:54:28.039773] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.450 [2024-12-06 17:54:28.039800] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.450 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.039897] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.039925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.040051] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.040077] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.040171] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.040197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.040359] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.040393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.040509] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.040553] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.040662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.040718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.040837] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.040864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.040988] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.041015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.041143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.041190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.041364] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.041397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.041509] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.041536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.041695] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.041740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.041830] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.041856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.041946] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.041972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.042114] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.042141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.042291] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.042324] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.042485] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.042518] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.042709] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.042736] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.042854] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.042880] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.042999] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.043026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.043172] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.043219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.043355] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.043389] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.043606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.043639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.043775] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.043802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.043884] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.043911] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.044057] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.044084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.044199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.044226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.044393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.044426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.044590] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.044623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.044796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.044823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.044951] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.044983] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.045127] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.045154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.045304] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.045351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.045489] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.045522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.045649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.045686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.045814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.045841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.045979] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.451 [2024-12-06 17:54:28.046004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.451 qpair failed and we were unable to recover it. 00:38:46.451 [2024-12-06 17:54:28.046116] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.046142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.046228] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.046255] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.046424] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.046457] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.046590] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.046618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.046732] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.046759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.046901] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.046929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.047045] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.047071] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.047188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.047214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.047364] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.047397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.047584] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.047616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.047763] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.047790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.047909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.047941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.048072] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.048099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.048243] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.048269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.048491] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.048555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.048755] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.048782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.048905] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.048950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.049085] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.049118] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.049259] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.049307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.049474] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.049507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.049644] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.049684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.049824] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.049850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.049972] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.049998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.050153] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.050187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.050384] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.050416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.050563] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.050596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.050739] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.050766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.050877] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.050903] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.050998] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.051024] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.051162] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.051187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.051283] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.051313] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.051446] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.051478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.051606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.051637] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.051828] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.051867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.051972] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.052000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.052165] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.052211] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.052312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.052343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.452 [2024-12-06 17:54:28.052531] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.452 [2024-12-06 17:54:28.052579] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.452 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.052701] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.052728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.052817] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.052843] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.052958] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.052984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.053065] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.053091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.053206] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.053231] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.053323] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.053350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.053475] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.053514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.053605] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.053633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.053777] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.053805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.053923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.053949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.054027] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.054053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.054169] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.054196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.054333] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.054379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.054459] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.054489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.054583] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.054609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.054728] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.054761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.054908] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.054953] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.055043] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.055069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.055155] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.055180] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.055292] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.055318] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.055407] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.055432] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.055518] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.055546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.055637] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.055672] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.055819] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.055845] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.055935] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.055962] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.056156] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.056183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.056324] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.056350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.056441] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.056467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.056557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.056583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.056678] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.056704] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.056807] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.056839] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.056987] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.057033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.057149] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.057174] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.453 [2024-12-06 17:54:28.057256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.453 [2024-12-06 17:54:28.057280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.453 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.057360] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.057385] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.057470] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.057495] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.057631] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.057657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.057753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.057779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.057873] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.057898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.057986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.058013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.058130] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.058155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.058237] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.058262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.058349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.058375] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.058483] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.058510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.058600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.058627] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.058725] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.058750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.058829] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.058854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.058926] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.058951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.059062] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.059088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.059199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.059226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.059300] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.059326] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.059472] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.059498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.059579] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.059604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.059678] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.059703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.059824] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.059852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.059946] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.059973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.060087] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.060113] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.060196] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.060222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.060307] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.060332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.060438] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.060476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.060576] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.060606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.060698] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.060726] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.060867] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.060894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.061037] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.061063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.061160] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.061186] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.061319] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.061366] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.061482] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.061508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.061656] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.061689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.061776] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.061802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.061906] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.061939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.062121] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.062168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.454 qpair failed and we were unable to recover it. 00:38:46.454 [2024-12-06 17:54:28.062258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.454 [2024-12-06 17:54:28.062284] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.062363] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.062388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.062501] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.062527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.062621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.062652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.062801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.062831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.062963] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.062992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.063112] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.063140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.063238] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.063270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.063375] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.063404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.063540] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.063567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.063716] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.063742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.063861] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.063905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.063993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.064019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.064117] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.064145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.064246] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.064272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.064379] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.064405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.064525] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.064551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.064640] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.064671] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.064755] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.064782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.064932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.064957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.065066] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.065092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.065209] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.065235] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.065349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.065375] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.065457] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.065483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.065570] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.065596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.065674] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.065700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.065815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.065841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.065949] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.065975] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.066052] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.066077] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.066159] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.066186] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.066272] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.066298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.066380] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.066406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.066549] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.066575] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.066685] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.066711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.066815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.066841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.066918] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.066944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.067032] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.067059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.067169] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.067199] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.455 qpair failed and we were unable to recover it. 00:38:46.455 [2024-12-06 17:54:28.067319] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.455 [2024-12-06 17:54:28.067345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.067489] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.067517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.067631] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.067657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.067753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.067779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.067888] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.067917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.068035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.068061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.068155] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.068181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.068344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.068372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.068519] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.068548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.068643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.068683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.068847] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.068879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.069028] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.069060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.069202] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.069234] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.069398] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.069444] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.069537] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.069563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.069648] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.069686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.069764] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.069790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.069895] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.069927] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.070056] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.070082] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.070162] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.070187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.070306] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.070332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.070480] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.070506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.070593] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.070623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.070745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.070772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.070891] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.070917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.071019] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.071048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.071135] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.071170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.071256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.071284] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.071396] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.071424] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.071518] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.071547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.071688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.071730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.071825] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.071850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.071948] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.071974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.072105] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.072134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.072258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.072287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.072411] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.072439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.072539] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.072568] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.072716] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.456 [2024-12-06 17:54:28.072742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.456 qpair failed and we were unable to recover it. 00:38:46.456 [2024-12-06 17:54:28.072866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.072892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.073007] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.073036] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.073190] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.073219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.073373] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.073401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.073529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.073558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.073697] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.073724] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.073819] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.073844] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.074003] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.074032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.074156] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.074200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.074317] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.074346] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.074476] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.074504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.074631] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.074656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.074750] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.074776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.074868] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.074895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.075039] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.075065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.075162] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.075188] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.075302] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.075331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.075449] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.075490] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.075610] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.075638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.075805] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.075831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.075974] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.076000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.076114] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.076159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.076311] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.076340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.076518] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.076547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.076676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.076725] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.076846] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.076872] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.076983] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.077009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.077131] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.077157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.077312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.077345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.077532] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.077561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.077732] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.077759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.077840] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.077865] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.077947] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.077973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.078084] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.078112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.078250] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.078278] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.078437] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.078465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.078597] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.078626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.078749] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.457 [2024-12-06 17:54:28.078775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.457 qpair failed and we were unable to recover it. 00:38:46.457 [2024-12-06 17:54:28.078869] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.078895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.078979] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.079005] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.079134] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.079163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.079255] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.079284] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.079413] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.079441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.079538] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.079567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.079678] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.079705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.079852] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.079878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.080010] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.080038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.080191] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.080220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.080351] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.080379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.080508] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.080537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.080689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.080732] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.080852] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.080877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.081014] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.081043] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.081224] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.081253] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.081347] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.081375] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.081537] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.081569] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.081709] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.081736] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.081848] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.081873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.081978] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.082007] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.082143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.082169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.082283] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.082309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.082450] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.082480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.082625] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.082650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.082773] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.082798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.082890] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.082916] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.083044] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.083070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.083194] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.083220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.083360] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.083390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.083521] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.083556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.083695] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.083738] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.083881] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.083907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.458 [2024-12-06 17:54:28.084025] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.458 [2024-12-06 17:54:28.084051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.458 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.084167] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.084193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.084303] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.084335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.084435] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.084463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.084611] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.084639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.084775] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.084801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.084944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.084970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.085056] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.085082] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.085191] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.085219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.085367] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.085395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.085527] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.085555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.085657] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.085715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.085829] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.085855] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.085973] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.085999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.086139] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.086168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.086300] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.086329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.086419] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.086487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.086620] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.086650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.086813] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.086841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.086998] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.087026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.087130] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.087159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.087248] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.087277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.087387] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.087415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.087518] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.087547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.087676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.087705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.087837] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.087865] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.087965] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.087996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.088117] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.088145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.088305] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.088333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.088488] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.088519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.088678] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.088709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.088837] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.088867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.088980] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.089010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.089161] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.089190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.089321] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.089351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.089480] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.089510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.089632] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.089661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.459 [2024-12-06 17:54:28.089851] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.459 [2024-12-06 17:54:28.089886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.459 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.089991] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.090021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.090126] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.090155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.090275] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.090305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.090457] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.090491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.090610] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.090710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.090877] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.090907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.091066] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.091096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.091219] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.091248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.091406] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.091435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.091535] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.091566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.091655] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.091690] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.091843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.091875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.092009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.092039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.092140] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.092169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.092266] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.092297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.092460] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.092493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.092618] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.092648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.092782] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.092811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.092938] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.092968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.093106] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.093135] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.093253] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.093283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.093416] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.093446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.093577] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.093606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.093735] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.093765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.093903] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.093933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.094091] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.094121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.094253] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.094283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.094386] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.094416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.094548] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.094581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.094684] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.094714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.094874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.094903] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.095062] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.095093] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.095188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.095218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.095318] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.095348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.095471] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.095500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.095652] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.095707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.095806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.095837] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.460 [2024-12-06 17:54:28.095940] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.460 [2024-12-06 17:54:28.095969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.460 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.096089] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.096119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.096255] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.096291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.096390] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.096420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.096559] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.096589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.096724] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.096755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.096895] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.096925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.097021] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.097050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.097146] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.097177] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.097332] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.097361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.097463] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.097492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.097599] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.097630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.097763] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.097793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.097883] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.097913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.098084] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.098114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.098210] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.098241] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.098369] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.098402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.098598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.098650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.098814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.098844] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.098981] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.099011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.099112] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.099142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.099290] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.099319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.099444] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.099474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.099611] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.099643] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.099816] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.099848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.099955] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.099985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.100120] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.100150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.100254] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.100283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.100386] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.100417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.100559] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.100589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.100719] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.100749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.100856] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.100888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.101024] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.101054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.101155] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.101184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.101313] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.101343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.101474] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.101507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.101603] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.101634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.101779] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.101809] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.461 qpair failed and we were unable to recover it. 00:38:46.461 [2024-12-06 17:54:28.101940] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.461 [2024-12-06 17:54:28.101969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.102123] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.102153] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.102288] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.102319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.102525] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.102559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.102708] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.102743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.102872] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.102901] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.103060] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.103090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.103223] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.103253] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.103353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.103384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.103522] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.103551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.103679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.103710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.103865] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.103895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.103997] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.104027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.104152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.104181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.104312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.104341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.104445] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.104475] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.104570] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.104599] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.104760] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.104790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.104929] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.104960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.105068] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.105098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.105259] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.105288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.105414] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.105444] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.105560] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.105609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.105766] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.105796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.105932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.105962] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.106108] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.106140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.106281] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.106314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.106452] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.106483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.106633] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.106672] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.106790] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.106820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.106957] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.106988] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.107099] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.107134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.107309] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.107356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.107562] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.107610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.107791] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.107860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.107990] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.108052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.108272] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.108341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.108485] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.108533] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.462 qpair failed and we were unable to recover it. 00:38:46.462 [2024-12-06 17:54:28.108662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.462 [2024-12-06 17:54:28.108705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.108873] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.108907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.109007] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.109037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.109171] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.109201] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.109342] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.109373] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.109502] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.109533] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.109635] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.109681] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.109818] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.109851] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.110013] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.110043] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.110138] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.110169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.110302] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.110333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.110469] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.110499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.110629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.110660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.110801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.110832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.110970] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.111000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.111126] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.111156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.111288] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.111319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.111416] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.111447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.111581] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.111611] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.111764] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.111796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.111906] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.111936] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.112022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.112052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.112161] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.112193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.112331] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.112361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.112453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.112483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.112646] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.112686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.112786] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.112817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.112956] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.112987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.113090] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.113121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.113262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.113293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.113394] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.113425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.113574] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.113604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.113740] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.113772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.113907] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.463 [2024-12-06 17:54:28.113961] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.463 qpair failed and we were unable to recover it. 00:38:46.463 [2024-12-06 17:54:28.114134] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.114180] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.114344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.114390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.114494] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.114526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.114631] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.114662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.114832] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.114863] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.114985] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.115018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.115125] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.115156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.115335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.115366] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.115521] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.115555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.115705] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.115737] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.115844] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.115876] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.116007] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.116040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.116176] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.116208] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.116437] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.116473] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.116642] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.116682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.116809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.116840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.116946] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.116976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.117098] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.117132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.117266] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.117297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.117413] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.117447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.117547] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.117577] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.117694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.117725] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.117835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.117868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.117975] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.118005] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.118137] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.118168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.118263] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.118293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.118399] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.118430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.118561] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.118592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.118753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.118784] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.118878] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.118909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.119039] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.119070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.119177] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.119208] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.119336] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.119367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.119506] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.119538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.119635] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.119673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.119832] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.119864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.120025] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.120057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.120189] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.464 [2024-12-06 17:54:28.120220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.464 qpair failed and we were unable to recover it. 00:38:46.464 [2024-12-06 17:54:28.120376] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.120407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.120540] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.120579] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.120743] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.120776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.120910] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.120941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.121124] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.121173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.121346] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.121380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.121564] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.121598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.121769] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.121819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.121999] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.122032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.122161] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.122201] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.122408] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.122444] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.122615] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.122648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.122842] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.122908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.123082] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.123148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.123316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.123346] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.123482] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.123513] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.123652] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.123699] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.123833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.123863] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.123965] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.123998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.124143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.124175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.124352] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.124383] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.124519] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.124551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.124690] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.124723] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.124857] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.124890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.125027] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.125059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.125181] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.125214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.125352] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.125385] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.125512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.125544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.125689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.125722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.125861] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.125893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.125996] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.126031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.126152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.126184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.126323] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.126354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.126512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.126546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.126702] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.126749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.126858] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.126890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.127022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.465 [2024-12-06 17:54:28.127053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.465 qpair failed and we were unable to recover it. 00:38:46.465 [2024-12-06 17:54:28.127191] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.127223] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.127358] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.127392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.127518] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.127550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.127700] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.127733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.127869] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.127905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.128047] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.128080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.128221] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.128254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.128367] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.128399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.128540] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.128571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.128692] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.128725] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.128862] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.128895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.129030] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.129064] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.129206] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.129239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.129348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.129381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.129514] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.129546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.129689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.129725] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.129833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.129864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.129979] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.130011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.130133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.130166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.130280] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.130312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.130444] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.130475] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.130589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.130621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.130759] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.130795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.130930] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.130962] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.131105] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.131137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.131271] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.131303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.131433] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.131465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.131609] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.131641] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.131758] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.131793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.131931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.131963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.132113] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.132146] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.132266] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.132298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.132456] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.132499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.132672] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.132704] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.466 qpair failed and we were unable to recover it. 00:38:46.466 [2024-12-06 17:54:28.132841] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.466 [2024-12-06 17:54:28.132872] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.133000] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.133033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.133172] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.133208] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.133344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.133377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.133490] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.133522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.133688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.133721] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.133859] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.133892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.134021] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.134052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.134198] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.134230] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.134369] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.134401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.134503] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.134544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.134645] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.134689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.134823] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.134855] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.134964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.134996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.135110] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.135142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.135244] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.135277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.135383] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.135417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.135563] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.135595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.135726] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.135759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.135927] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.135960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.136103] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.136135] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.136267] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.136298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.136432] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.136468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.136585] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.136617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.136752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.136784] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.136894] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.136926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.137094] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.137126] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.137268] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.137301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.137416] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.137448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.137558] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.137591] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.137702] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.137751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.137927] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.137967] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.138100] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.138135] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.138291] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.467 [2024-12-06 17:54:28.138325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.467 qpair failed and we were unable to recover it. 00:38:46.467 [2024-12-06 17:54:28.138500] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.138542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.138720] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.138753] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.138890] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.138921] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.139034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.139066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.139171] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.139204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.139352] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.139384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.139521] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.139552] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.139693] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.139727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.139867] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.139903] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.140071] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.140104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.140239] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.140271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.140404] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.140437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.140541] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.140573] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.140720] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.140753] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.140898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.140932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.141068] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.141104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.141268] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.141310] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.141452] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.141485] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.141653] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.141693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.141836] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.141869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.142005] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.142039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.142170] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.142202] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.142334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.142365] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.142531] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.142563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.142686] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.142720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.142862] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.142894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.143033] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.143066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.143201] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.143234] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.143381] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.143415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.143556] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.143588] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.143706] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.143739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.143881] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.143915] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.468 [2024-12-06 17:54:28.144056] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.468 [2024-12-06 17:54:28.144092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.468 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.144232] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.144265] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.144414] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.144446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.144566] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.144604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.144777] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.144811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.144914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.144950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.145124] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.145157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.145292] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.145336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.145555] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.145612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.145811] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.145856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.146013] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.146058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.146279] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.146348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.146601] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.146661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.146873] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.146933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.147112] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.147153] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.147323] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.147356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.147498] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.147541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.147687] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.147733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.147939] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.147983] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.148152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.148197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.148358] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.148403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.148607] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.148653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.148838] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.148875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.148998] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.149032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.149184] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.149225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.149335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.149369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.149512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.149546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.149660] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.149705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.149835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.149867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.150004] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.150038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.150188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.150221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.150369] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.150403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.150568] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.150608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.150808] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.150842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.469 qpair failed and we were unable to recover it. 00:38:46.469 [2024-12-06 17:54:28.150958] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.469 [2024-12-06 17:54:28.150991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.151136] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.151169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.151302] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.151335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.151432] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.151464] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.151579] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.151612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.151796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.151831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.151976] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.152009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.152153] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.152187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.152378] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.152419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.152589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.152629] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.152821] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.152862] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.153044] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.153085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.153263] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.153303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.153470] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.153510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.153681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.153737] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.153887] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.153927] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.154116] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.154149] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.154302] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.154337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.154510] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.154543] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.154694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.154728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.154898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.154931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.155068] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.155101] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.155242] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.155275] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.155423] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.155456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.155556] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.155590] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.155735] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.155769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.155916] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.155950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.156053] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.156087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.156223] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.156257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.156401] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.156434] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.156581] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.156620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.156768] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.156802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.156942] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.156974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.157107] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.157139] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.157272] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.470 [2024-12-06 17:54:28.157304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.470 qpair failed and we were unable to recover it. 00:38:46.470 [2024-12-06 17:54:28.157498] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.157538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.157691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.157726] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.157866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.157899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.158071] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.158104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.158275] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.158308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.158453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.158485] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.158602] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.158636] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.158783] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.158816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.158957] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.158989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.159130] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.159164] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.159299] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.159332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.159472] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.159504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.159649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.159699] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.159832] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.159865] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.160071] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.160104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.160273] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.160306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.160451] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.160485] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.160594] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.160629] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.160773] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.160807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.160944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.160978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.161112] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.161145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.161292] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.161325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.161499] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.161533] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.161722] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.161758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.161889] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.161924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.162071] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.162105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.162315] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.162350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.162455] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.162490] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.162628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.162669] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.162848] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.162884] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.163033] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.163068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.163247] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.163283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.163426] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.163461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.163610] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.163645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.163808] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.163842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.471 qpair failed and we were unable to recover it. 00:38:46.471 [2024-12-06 17:54:28.164026] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.471 [2024-12-06 17:54:28.164067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.164265] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.164297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.164428] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.164460] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.164662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.164701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.164879] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.164912] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.165134] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.165168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.165345] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.165380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.165488] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.165523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.165626] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.165660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.165791] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.165827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.166003] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.166039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.166217] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.166251] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.166398] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.166432] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.166557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.166594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.166699] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.166735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.166906] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.166941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.167121] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.167155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.167356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.167387] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.167554] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.167586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.167751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.167783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.167920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.167952] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.168089] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.168121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.168259] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.168309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.168483] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.168518] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.168621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.168656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.168812] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.168846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.168961] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.169016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.169182] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.169218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.169413] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.169453] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.169649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.169697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.169836] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.169871] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.472 qpair failed and we were unable to recover it. 00:38:46.472 [2024-12-06 17:54:28.170019] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.472 [2024-12-06 17:54:28.170053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.170170] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.170205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.170316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.170351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.170492] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.170527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.170710] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.170746] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.170905] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.170938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.171038] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.171072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.171215] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.171248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.171423] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.171454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.171594] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.171631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.171838] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.171873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.171984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.172020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.172169] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.172205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.172322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.172357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.172475] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.172510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.172673] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.172708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.172863] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.172898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.173005] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.173039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.173187] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.173222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.173337] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.173371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.173508] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.173543] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.173703] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.173738] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.173845] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.173878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.174042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.174076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.174184] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.174218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.174400] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.174435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.174566] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.174600] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.174753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.174788] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.174933] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.174968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.175127] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.175161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.175333] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.175367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.175502] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.175535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.175686] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.175721] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.175908] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.175944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.176059] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.176096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.473 qpair failed and we were unable to recover it. 00:38:46.473 [2024-12-06 17:54:28.176236] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.473 [2024-12-06 17:54:28.176271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.176436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.176471] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.176682] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.176717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.176905] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.176939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.177116] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.177149] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.177356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.177388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.177514] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.177546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.177711] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.177743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.177904] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.177951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.178056] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.178091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.178236] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.178272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.178416] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.178451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.178566] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.178602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.178722] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.178759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.178904] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.178948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.179096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.179130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.179302] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.179337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.179523] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.179558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.179726] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.179759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.179923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.179955] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.180093] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.180143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.180279] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.180312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.180511] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.180551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.180736] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.180773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.180955] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.180991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.181224] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.181257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.181449] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.181490] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.181613] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.181654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.181841] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.181877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.182014] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.182051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.182201] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.182238] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.182382] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.182418] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.182589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.182631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.182813] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.182851] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.183012] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.183050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.183225] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.183265] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.183393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.183435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.474 [2024-12-06 17:54:28.183643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.474 [2024-12-06 17:54:28.183707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.474 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.183858] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.183900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.184070] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.184109] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.184230] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.184268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.184499] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.184540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.184722] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.184762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.184950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.184987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.185200] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.185232] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.185370] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.185402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.185586] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.185626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.185818] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.185852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.185967] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.186001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.186165] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.186198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.186337] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.186370] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.186499] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.186532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.186675] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.186709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.186885] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.186926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.187136] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.187174] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.187314] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.187347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.187528] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.187569] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.187705] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.187747] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.187906] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.187946] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.188148] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.188188] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.188356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.188396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.188540] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.188580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.188762] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.188803] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.188966] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.189005] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.189159] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.189197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.189379] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.189412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.189548] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.189580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.189745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.189777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.189895] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.189927] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.190099] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.190139] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.190255] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.190295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.190512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.190546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.190650] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.190690] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.190858] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.190898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.191063] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.191104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.475 [2024-12-06 17:54:28.191262] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.475 [2024-12-06 17:54:28.191303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.475 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.191468] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.191507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.191640] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.191703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.191832] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.191872] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.192002] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.192043] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.192210] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.192251] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.192443] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.192475] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.192573] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.192606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.192745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.192778] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.192948] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.192979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.193161] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.193194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.193337] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.193378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.193567] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.193606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.193756] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.193798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.193958] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.194000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.194152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.194192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.194348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.194388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.194565] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.194605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.194773] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.194814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.195015] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.195054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.195174] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.195207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.195401] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.195433] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.195558] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.195590] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.195755] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.195796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.195959] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.196001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.196165] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.196206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.196350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.196390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.196520] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.196561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.196688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.196730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.196870] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.196911] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.197109] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.197149] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.197304] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.197336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.197471] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.197505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.197715] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.197748] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.197881] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.197914] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.198041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.198081] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.198241] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.198282] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.198413] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.198452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.198620] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.476 [2024-12-06 17:54:28.198660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.476 qpair failed and we were unable to recover it. 00:38:46.476 [2024-12-06 17:54:28.198878] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.198911] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.199041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.199074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.199257] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.199298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.199459] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.199500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.199673] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.199714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.199839] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.199879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.200073] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.200114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.200290] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.200330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.200463] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.200504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.200615] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.200656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.200855] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.200886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.201028] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.201060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.201155] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.201187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.201294] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.201326] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.201547] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.201581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.201727] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.201760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.201956] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.201996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.202194] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.202235] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.202386] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.202427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.202595] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.202628] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.202780] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.202840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.202993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.203045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.203173] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.203205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.203339] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.203371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.203503] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.203534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.203692] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.203734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.203887] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.203927] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.204089] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.204130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.204304] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.204344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.204505] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.204545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.204745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.204786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.204946] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.204987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.205142] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.205182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.477 qpair failed and we were unable to recover it. 00:38:46.477 [2024-12-06 17:54:28.205349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.477 [2024-12-06 17:54:28.205389] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.205591] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.205632] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.205761] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.205801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.205992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.206032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.206191] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.206231] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.206427] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.206467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.206598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.206639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.206809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.206849] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.206992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.207035] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.207230] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.207270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.207466] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.207505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.207702] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.207743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.207901] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.207941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.208099] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.208139] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.208327] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.208361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.208502] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.208535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.208680] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.208731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.208894] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.208950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.209115] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.209157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.209329] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.209368] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.209529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.209570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.209734] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.209775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.209939] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.209978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.210146] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.210186] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.210345] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.210385] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.210549] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.210589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.210753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.210794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.210986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.211026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.211203] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.211244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.211414] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.211447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.211578] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.211611] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.211791] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.211832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.212016] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.212056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.212228] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.212268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.212423] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.212464] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.212605] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.212646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.212821] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.212862] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.213025] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.213067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.213227] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.213266] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.213427] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.213467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.213646] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.478 [2024-12-06 17:54:28.213689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.478 qpair failed and we were unable to recover it. 00:38:46.478 [2024-12-06 17:54:28.213831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.213885] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.214042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.214081] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.214271] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.214311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.214506] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.214546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.214708] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.214750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.214928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.214962] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.215094] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.215127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.215300] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.215341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.215547] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.215586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.215756] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.215797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.215972] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.216012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.216120] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.216160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.216300] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.216342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.216499] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.216546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.216711] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.216752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.216931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.216972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.217122] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.217154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.217261] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.217292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.217501] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.217534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.217679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.217713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.217870] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.217925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.218058] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.218099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.218226] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.218266] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.218459] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.218499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.218661] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.218713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.218906] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.218937] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.219102] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.219134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.219281] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.219315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.219442] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.219483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.219606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.219646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.219823] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.219863] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.220016] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.220056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.220217] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.220257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.220423] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.220464] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.220624] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.220681] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.220817] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.220858] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.221034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.221074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.221229] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.221269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.221428] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.221467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.221632] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.221682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.221862] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.479 [2024-12-06 17:54:28.221896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.479 qpair failed and we were unable to recover it. 00:38:46.479 [2024-12-06 17:54:28.222032] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.222065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.222235] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.222275] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.222465] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.222505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.222641] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.222692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.222845] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.222877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.223085] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.223124] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.223321] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.223362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.223487] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.223527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.223739] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.223773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.223914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.223947] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.224123] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.224163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.224358] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.224398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.224525] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.224574] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.224776] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.224817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.224979] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.225020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.225165] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.225206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.225371] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.225412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.225604] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.225644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.225772] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.225811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.226017] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.226058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.226217] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.226257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.226378] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.226417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.226586] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.226626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.226768] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.226808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.226967] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.227006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.227207] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.227247] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.227405] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.227445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.227602] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.227642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.227829] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.227870] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.228029] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.228069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.228256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.228296] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.228447] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.228487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.228653] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.228718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.228880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.228920] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.229077] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.229118] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.229309] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.229349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.229466] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.229506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.229707] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.229748] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.229870] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.229910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.230088] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.230128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.230328] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.230369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.480 qpair failed and we were unable to recover it. 00:38:46.480 [2024-12-06 17:54:28.230502] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.480 [2024-12-06 17:54:28.230542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.230661] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.230720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.230849] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.230890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.231087] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.231128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.231322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.231362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.231530] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.231571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.231766] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.231808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.231928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.231967] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.232135] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.232176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.232350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.232384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.232528] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.232561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.232746] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.232786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.232923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.232956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.233155] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.233188] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.233328] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.233362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.233526] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.233566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.233724] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.233766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.233915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.233955] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.234118] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.234157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.234350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.234390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.234541] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.234582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.234798] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.234831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.234968] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.235001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.235174] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.235214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.235372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.235412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.235579] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.235619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.235808] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.235850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.236000] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.236040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.236190] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.236229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.236429] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.236470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.236621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.236662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.236851] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.236892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.237093] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.237133] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.237283] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.237323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.237473] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.237513] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.237713] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.237755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.237875] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.237916] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.238033] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.238073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.238279] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.238320] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.481 qpair failed and we were unable to recover it. 00:38:46.481 [2024-12-06 17:54:28.238486] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.481 [2024-12-06 17:54:28.238526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.482 qpair failed and we were unable to recover it. 00:38:46.482 [2024-12-06 17:54:28.238641] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.482 [2024-12-06 17:54:28.238690] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.482 qpair failed and we were unable to recover it. 00:38:46.482 [2024-12-06 17:54:28.238857] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.482 [2024-12-06 17:54:28.238898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.482 qpair failed and we were unable to recover it. 00:38:46.482 [2024-12-06 17:54:28.239094] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.482 [2024-12-06 17:54:28.239133] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.482 qpair failed and we were unable to recover it. 00:38:46.482 [2024-12-06 17:54:28.239295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.482 [2024-12-06 17:54:28.239335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.482 qpair failed and we were unable to recover it. 00:38:46.482 [2024-12-06 17:54:28.239542] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.482 [2024-12-06 17:54:28.239576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.482 qpair failed and we were unable to recover it. 00:38:46.482 [2024-12-06 17:54:28.239690] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.482 [2024-12-06 17:54:28.239724] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.482 qpair failed and we were unable to recover it. 00:38:46.482 [2024-12-06 17:54:28.239865] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.482 [2024-12-06 17:54:28.239898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.482 qpair failed and we were unable to recover it. 00:38:46.482 [2024-12-06 17:54:28.240059] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.482 [2024-12-06 17:54:28.240099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.482 qpair failed and we were unable to recover it. 00:38:46.482 [2024-12-06 17:54:28.240259] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.482 [2024-12-06 17:54:28.240299] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.482 qpair failed and we were unable to recover it. 00:38:46.482 [2024-12-06 17:54:28.240489] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.482 [2024-12-06 17:54:28.240530] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.482 qpair failed and we were unable to recover it. 00:38:46.482 [2024-12-06 17:54:28.240705] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.482 [2024-12-06 17:54:28.240739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.482 qpair failed and we were unable to recover it. 00:38:46.482 [2024-12-06 17:54:28.240941] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.482 [2024-12-06 17:54:28.240988] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.482 qpair failed and we were unable to recover it. 00:38:46.482 [2024-12-06 17:54:28.241153] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.482 [2024-12-06 17:54:28.241193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.482 qpair failed and we were unable to recover it. 00:38:46.482 [2024-12-06 17:54:28.241397] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.482 [2024-12-06 17:54:28.241438] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.482 qpair failed and we were unable to recover it. 00:38:46.482 [2024-12-06 17:54:28.241602] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.482 [2024-12-06 17:54:28.241643] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.482 qpair failed and we were unable to recover it. 00:38:46.482 [2024-12-06 17:54:28.241800] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.482 [2024-12-06 17:54:28.241840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.482 qpair failed and we were unable to recover it. 00:38:46.482 [2024-12-06 17:54:28.242042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.482 [2024-12-06 17:54:28.242083] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.482 qpair failed and we were unable to recover it. 00:38:46.482 [2024-12-06 17:54:28.242253] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.482 [2024-12-06 17:54:28.242294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.482 qpair failed and we were unable to recover it. 00:38:46.482 [2024-12-06 17:54:28.242460] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.482 [2024-12-06 17:54:28.242499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.482 qpair failed and we were unable to recover it. 00:38:46.482 [2024-12-06 17:54:28.242703] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.482 [2024-12-06 17:54:28.242745] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.482 qpair failed and we were unable to recover it. 00:38:46.482 [2024-12-06 17:54:28.242901] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.482 [2024-12-06 17:54:28.242941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.482 qpair failed and we were unable to recover it. 00:38:46.482 [2024-12-06 17:54:28.243140] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.482 [2024-12-06 17:54:28.243180] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.482 qpair failed and we were unable to recover it. 00:38:46.482 [2024-12-06 17:54:28.243351] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.482 [2024-12-06 17:54:28.243401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.482 qpair failed and we were unable to recover it. 00:38:46.482 [2024-12-06 17:54:28.243555] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.482 [2024-12-06 17:54:28.243590] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.482 qpair failed and we were unable to recover it. 00:38:46.482 [2024-12-06 17:54:28.243778] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.482 [2024-12-06 17:54:28.243819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.482 qpair failed and we were unable to recover it. 00:38:46.482 [2024-12-06 17:54:28.243962] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.482 [2024-12-06 17:54:28.244003] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.482 qpair failed and we were unable to recover it. 00:38:46.482 [2024-12-06 17:54:28.244176] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.482 [2024-12-06 17:54:28.244217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.482 qpair failed and we were unable to recover it. 00:38:46.482 [2024-12-06 17:54:28.244375] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.483 [2024-12-06 17:54:28.244408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.483 qpair failed and we were unable to recover it. 00:38:46.483 [2024-12-06 17:54:28.244585] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.483 [2024-12-06 17:54:28.244625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.483 qpair failed and we were unable to recover it. 00:38:46.483 [2024-12-06 17:54:28.244816] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.483 [2024-12-06 17:54:28.244851] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.483 qpair failed and we were unable to recover it. 00:38:46.483 [2024-12-06 17:54:28.244984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.483 [2024-12-06 17:54:28.245018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.483 qpair failed and we were unable to recover it. 00:38:46.483 [2024-12-06 17:54:28.245160] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.483 [2024-12-06 17:54:28.245194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.483 qpair failed and we were unable to recover it. 00:38:46.483 [2024-12-06 17:54:28.245364] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.483 [2024-12-06 17:54:28.245397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.483 qpair failed and we were unable to recover it. 00:38:46.483 [2024-12-06 17:54:28.245515] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.483 [2024-12-06 17:54:28.245548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.483 qpair failed and we were unable to recover it. 00:38:46.483 [2024-12-06 17:54:28.245685] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.483 [2024-12-06 17:54:28.245720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.483 qpair failed and we were unable to recover it. 00:38:46.483 [2024-12-06 17:54:28.245846] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.483 [2024-12-06 17:54:28.245880] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.483 qpair failed and we were unable to recover it. 00:38:46.483 [2024-12-06 17:54:28.245978] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.483 [2024-12-06 17:54:28.246011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.483 qpair failed and we were unable to recover it. 00:38:46.483 [2024-12-06 17:54:28.246158] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.483 [2024-12-06 17:54:28.246193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.483 qpair failed and we were unable to recover it. 00:38:46.483 [2024-12-06 17:54:28.246367] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.483 [2024-12-06 17:54:28.246400] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.483 qpair failed and we were unable to recover it. 00:38:46.483 [2024-12-06 17:54:28.246538] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.483 [2024-12-06 17:54:28.246571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.483 qpair failed and we were unable to recover it. 00:38:46.483 [2024-12-06 17:54:28.246679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.483 [2024-12-06 17:54:28.246713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.483 qpair failed and we were unable to recover it. 00:38:46.483 [2024-12-06 17:54:28.246849] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.483 [2024-12-06 17:54:28.246883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.483 qpair failed and we were unable to recover it. 00:38:46.483 [2024-12-06 17:54:28.247022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.483 [2024-12-06 17:54:28.247055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.483 qpair failed and we were unable to recover it. 00:38:46.483 [2024-12-06 17:54:28.247234] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.483 [2024-12-06 17:54:28.247267] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.483 qpair failed and we were unable to recover it. 00:38:46.483 [2024-12-06 17:54:28.247399] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.483 [2024-12-06 17:54:28.247432] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.483 qpair failed and we were unable to recover it. 00:38:46.483 [2024-12-06 17:54:28.247548] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.483 [2024-12-06 17:54:28.247583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.483 qpair failed and we were unable to recover it. 00:38:46.483 [2024-12-06 17:54:28.247757] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.483 [2024-12-06 17:54:28.247792] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.483 qpair failed and we were unable to recover it. 00:38:46.781 [2024-12-06 17:54:28.247896] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.781 [2024-12-06 17:54:28.247929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.781 qpair failed and we were unable to recover it. 00:38:46.781 [2024-12-06 17:54:28.248077] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.781 [2024-12-06 17:54:28.248110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.781 qpair failed and we were unable to recover it. 00:38:46.781 [2024-12-06 17:54:28.248214] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.781 [2024-12-06 17:54:28.248248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.781 qpair failed and we were unable to recover it. 00:38:46.781 [2024-12-06 17:54:28.248393] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.781 [2024-12-06 17:54:28.248426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.781 qpair failed and we were unable to recover it. 00:38:46.781 [2024-12-06 17:54:28.248566] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.781 [2024-12-06 17:54:28.248605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.781 qpair failed and we were unable to recover it. 00:38:46.781 [2024-12-06 17:54:28.248760] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.781 [2024-12-06 17:54:28.248794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.248937] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.248970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.249145] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.249179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.249283] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.249316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.249448] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.249481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.249600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.249635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.249783] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.249817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.249961] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.249994] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.250098] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.250132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.250249] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.250284] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.250418] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.250451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.250587] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.250620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.250746] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.250780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.250927] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.250960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.251066] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.251099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.251270] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.251303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.251418] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.251451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.251581] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.251614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.251736] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.251771] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.251900] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.251934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.252072] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.252107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.252254] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.252287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.252419] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.252452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.252587] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.252620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.252814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.252848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.252993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.253026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.253173] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.253206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.253350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.253384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.253491] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.253524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.253630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.253672] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.253854] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.253888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.254034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.254067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.254188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.254221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.254365] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.254399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.782 [2024-12-06 17:54:28.254527] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.782 [2024-12-06 17:54:28.254560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.782 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.254692] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.254733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.254931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.254972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.255132] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.255173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.255303] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.255342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.255479] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.255526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.255696] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.255737] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.255932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.255972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.256115] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.256156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.256282] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.256322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.256517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.256558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.256692] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.256733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.256861] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.256900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.257052] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.257092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.257229] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.257271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.257442] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.257482] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.257630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.257670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.257785] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.257819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.257965] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.257998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.258167] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.258207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.258382] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.258416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.258558] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.258592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.258753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.258794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.258960] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.259000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.259205] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.259245] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.259409] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.259449] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.259572] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.259613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.259777] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.259819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.259989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.260030] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.260188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.260229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.260423] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.260463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.260621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.260662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.260849] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.260882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.783 [2024-12-06 17:54:28.261064] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.783 [2024-12-06 17:54:28.261097] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.783 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.261271] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.261312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.261479] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.261520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.261687] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.261728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.261884] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.261925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.262113] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.262154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.262312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.262352] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.262550] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.262591] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.262738] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.262779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.262908] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.262948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.263096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.263137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.263297] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.263337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.263497] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.263544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.263708] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.263749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.263937] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.263978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.264138] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.264178] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.264376] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.264408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.264557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.264607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.264824] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.264865] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.265054] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.265095] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.265268] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.265308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.265463] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.265503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.265631] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.265682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.265823] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.265864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.266044] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.266084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.266237] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.266277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.266450] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.266491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.266657] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.266700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.266802] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.266837] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.267038] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.267079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.267233] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.267274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.784 [2024-12-06 17:54:28.267403] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.784 [2024-12-06 17:54:28.267443] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.784 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.267619] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.267661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.267835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.267876] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.268042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.268081] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.268284] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.268318] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.268484] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.268539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.268709] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.268751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.268892] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.268934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.269134] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.269174] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.269336] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.269375] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.269574] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.269615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.269786] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.269827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.269957] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.269997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.270196] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.270236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.270410] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.270442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.270609] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.270642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.270808] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.270850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.271014] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.271055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.271216] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.271257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.271431] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.271472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.271635] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.271683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.271810] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.271857] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.272075] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.272108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.272259] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.272292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.272419] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.272461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.272629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.272679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.272872] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.272913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.273049] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.273091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.273254] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.273294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.273426] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.785 [2024-12-06 17:54:28.273466] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.785 qpair failed and we were unable to recover it. 00:38:46.785 [2024-12-06 17:54:28.273619] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.273659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.273799] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.273839] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.273992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.274032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.274156] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.274197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.274354] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.274394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.274554] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.274595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.274740] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.274781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.274909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.274951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.275107] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.275147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.275344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.275384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.275573] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.275613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.275784] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.275825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.275989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.276028] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.276230] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.276271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.276385] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.276425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.276561] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.276601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.276783] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.276826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.277032] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.277073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.277245] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.277286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.277448] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.277489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.277612] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.277652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.277827] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.277867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.278073] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.278113] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.278283] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.278323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.278484] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.278524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.278657] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.278708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.278895] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.278955] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.279136] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.279193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.279334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.279375] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.279500] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.279542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.279675] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.279718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.279887] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.786 [2024-12-06 17:54:28.279935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.786 qpair failed and we were unable to recover it. 00:38:46.786 [2024-12-06 17:54:28.280137] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.787 [2024-12-06 17:54:28.280178] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.787 qpair failed and we were unable to recover it. 00:38:46.787 [2024-12-06 17:54:28.280295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.787 [2024-12-06 17:54:28.280335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.787 qpair failed and we were unable to recover it. 00:38:46.787 [2024-12-06 17:54:28.280508] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.787 [2024-12-06 17:54:28.280549] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.787 qpair failed and we were unable to recover it. 00:38:46.787 [2024-12-06 17:54:28.280758] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.787 [2024-12-06 17:54:28.280792] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.787 qpair failed and we were unable to recover it. 00:38:46.787 [2024-12-06 17:54:28.280903] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.787 [2024-12-06 17:54:28.280936] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.787 qpair failed and we were unable to recover it. 00:38:46.787 [2024-12-06 17:54:28.281160] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.787 [2024-12-06 17:54:28.281194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.787 qpair failed and we were unable to recover it. 00:38:46.787 [2024-12-06 17:54:28.281332] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.787 [2024-12-06 17:54:28.281365] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.787 qpair failed and we were unable to recover it. 00:38:46.787 [2024-12-06 17:54:28.281510] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.787 [2024-12-06 17:54:28.281550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.787 qpair failed and we were unable to recover it. 00:38:46.787 [2024-12-06 17:54:28.281706] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.787 [2024-12-06 17:54:28.281748] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.787 qpair failed and we were unable to recover it. 00:38:46.787 [2024-12-06 17:54:28.281912] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.787 [2024-12-06 17:54:28.281953] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.787 qpair failed and we were unable to recover it. 00:38:46.787 [2024-12-06 17:54:28.282111] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.787 [2024-12-06 17:54:28.282151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.787 qpair failed and we were unable to recover it. 00:38:46.787 [2024-12-06 17:54:28.282313] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.787 [2024-12-06 17:54:28.282353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.787 qpair failed and we were unable to recover it. 00:38:46.787 [2024-12-06 17:54:28.282513] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.787 [2024-12-06 17:54:28.282553] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.787 qpair failed and we were unable to recover it. 00:38:46.787 [2024-12-06 17:54:28.282729] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.787 [2024-12-06 17:54:28.282771] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.787 qpair failed and we were unable to recover it. 00:38:46.787 [2024-12-06 17:54:28.282911] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.787 [2024-12-06 17:54:28.282951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.787 qpair failed and we were unable to recover it. 00:38:46.787 [2024-12-06 17:54:28.283144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.787 [2024-12-06 17:54:28.283184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.787 qpair failed and we were unable to recover it. 00:38:46.787 [2024-12-06 17:54:28.283310] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.787 [2024-12-06 17:54:28.283351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.787 qpair failed and we were unable to recover it. 00:38:46.787 [2024-12-06 17:54:28.283549] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.787 [2024-12-06 17:54:28.283589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.787 qpair failed and we were unable to recover it. 00:38:46.787 [2024-12-06 17:54:28.283773] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.787 [2024-12-06 17:54:28.283815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.787 qpair failed and we were unable to recover it. 00:38:46.787 [2024-12-06 17:54:28.283976] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.787 [2024-12-06 17:54:28.284017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.787 qpair failed and we were unable to recover it. 00:38:46.787 [2024-12-06 17:54:28.284180] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.787 [2024-12-06 17:54:28.284220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.787 qpair failed and we were unable to recover it. 00:38:46.787 [2024-12-06 17:54:28.284385] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.787 [2024-12-06 17:54:28.284425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.787 qpair failed and we were unable to recover it. 00:38:46.787 [2024-12-06 17:54:28.284616] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.787 [2024-12-06 17:54:28.284657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.787 qpair failed and we were unable to recover it. 00:38:46.787 [2024-12-06 17:54:28.284841] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.787 [2024-12-06 17:54:28.284882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.787 qpair failed and we were unable to recover it. 00:38:46.787 [2024-12-06 17:54:28.285046] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.285086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.285248] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.285288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.285435] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.285477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.285681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.285723] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.285854] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.285894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.286094] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.286135] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.286327] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.286367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.286530] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.286570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.286744] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.286785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.286945] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.286985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.287178] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.287218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.287370] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.287411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.287531] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.287571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.287773] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.287813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.287987] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.288027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.288150] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.288197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.288392] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.288431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.288565] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.288606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.288752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.288793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.288987] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.289027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.289202] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.289243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.289375] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.289416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.289542] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.289582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.289791] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.289833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.289961] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.290001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.290156] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.290196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.290328] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.290369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.290557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.290596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.290763] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.290804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.290977] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.291018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.291219] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.291259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.291424] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.291465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.788 qpair failed and we were unable to recover it. 00:38:46.788 [2024-12-06 17:54:28.291634] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.788 [2024-12-06 17:54:28.291686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.789 qpair failed and we were unable to recover it. 00:38:46.789 [2024-12-06 17:54:28.291826] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.789 [2024-12-06 17:54:28.291866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.789 qpair failed and we were unable to recover it. 00:38:46.789 [2024-12-06 17:54:28.292073] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.789 [2024-12-06 17:54:28.292113] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.789 qpair failed and we were unable to recover it. 00:38:46.789 [2024-12-06 17:54:28.292259] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.789 [2024-12-06 17:54:28.292300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.789 qpair failed and we were unable to recover it. 00:38:46.789 [2024-12-06 17:54:28.292428] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.789 [2024-12-06 17:54:28.292470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.789 qpair failed and we were unable to recover it. 00:38:46.789 [2024-12-06 17:54:28.292632] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.789 [2024-12-06 17:54:28.292693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.789 qpair failed and we were unable to recover it. 00:38:46.789 [2024-12-06 17:54:28.292815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.789 [2024-12-06 17:54:28.292856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.789 qpair failed and we were unable to recover it. 00:38:46.789 [2024-12-06 17:54:28.293024] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.789 [2024-12-06 17:54:28.293064] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.789 qpair failed and we were unable to recover it. 00:38:46.789 [2024-12-06 17:54:28.293223] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.789 [2024-12-06 17:54:28.293264] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.789 qpair failed and we were unable to recover it. 00:38:46.789 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/target_disconnect.sh: line 36: 414723 Killed "${NVMF_APP[@]}" "$@" 00:38:46.789 [2024-12-06 17:54:28.293425] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.789 [2024-12-06 17:54:28.293465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.789 qpair failed and we were unable to recover it. 00:38:46.789 [2024-12-06 17:54:28.293639] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.789 [2024-12-06 17:54:28.293692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.789 qpair failed and we were unable to recover it. 00:38:46.789 [2024-12-06 17:54:28.293849] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.789 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@48 -- # disconnect_init 10.0.0.2 00:38:46.789 [2024-12-06 17:54:28.293890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.789 qpair failed and we were unable to recover it. 00:38:46.789 [2024-12-06 17:54:28.294099] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.789 [2024-12-06 17:54:28.294132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.789 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@17 -- # nvmfappstart -m 0xF0 00:38:46.789 qpair failed and we were unable to recover it. 00:38:46.789 [2024-12-06 17:54:28.294237] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.789 [2024-12-06 17:54:28.294271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.789 qpair failed and we were unable to recover it. 00:38:46.789 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:38:46.789 [2024-12-06 17:54:28.294410] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.789 [2024-12-06 17:54:28.294444] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.789 qpair failed and we were unable to recover it. 00:38:46.789 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@726 -- # xtrace_disable 00:38:46.789 [2024-12-06 17:54:28.294614] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.789 [2024-12-06 17:54:28.294655] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.789 qpair failed and we were unable to recover it. 00:38:46.789 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:38:46.789 [2024-12-06 17:54:28.294795] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.789 [2024-12-06 17:54:28.294836] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.789 qpair failed and we were unable to recover it. 00:38:46.789 [2024-12-06 17:54:28.295038] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.789 [2024-12-06 17:54:28.295072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.789 qpair failed and we were unable to recover it. 00:38:46.789 [2024-12-06 17:54:28.295231] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.789 [2024-12-06 17:54:28.295283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.789 qpair failed and we were unable to recover it. 00:38:46.789 [2024-12-06 17:54:28.295412] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.789 [2024-12-06 17:54:28.295452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.789 qpair failed and we were unable to recover it. 00:38:46.789 [2024-12-06 17:54:28.295657] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.789 [2024-12-06 17:54:28.295717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.789 qpair failed and we were unable to recover it. 00:38:46.789 [2024-12-06 17:54:28.295932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.789 [2024-12-06 17:54:28.295973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.789 qpair failed and we were unable to recover it. 00:38:46.789 [2024-12-06 17:54:28.296153] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.789 [2024-12-06 17:54:28.296194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.789 qpair failed and we were unable to recover it. 00:38:46.789 [2024-12-06 17:54:28.296376] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.789 [2024-12-06 17:54:28.296417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.790 qpair failed and we were unable to recover it. 00:38:46.790 [2024-12-06 17:54:28.296620] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.790 [2024-12-06 17:54:28.296661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.790 qpair failed and we were unable to recover it. 00:38:46.790 [2024-12-06 17:54:28.296809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.790 [2024-12-06 17:54:28.296850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.790 qpair failed and we were unable to recover it. 00:38:46.790 [2024-12-06 17:54:28.297013] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.790 [2024-12-06 17:54:28.297058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.790 qpair failed and we were unable to recover it. 00:38:46.790 [2024-12-06 17:54:28.297257] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.790 [2024-12-06 17:54:28.297298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.790 qpair failed and we were unable to recover it. 00:38:46.790 [2024-12-06 17:54:28.297450] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.790 [2024-12-06 17:54:28.297490] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.790 qpair failed and we were unable to recover it. 00:38:46.790 [2024-12-06 17:54:28.297622] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.790 [2024-12-06 17:54:28.297673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.790 qpair failed and we were unable to recover it. 00:38:46.790 [2024-12-06 17:54:28.297814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.790 [2024-12-06 17:54:28.297855] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.790 qpair failed and we were unable to recover it. 00:38:46.790 [2024-12-06 17:54:28.298015] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.790 [2024-12-06 17:54:28.298055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.790 qpair failed and we were unable to recover it. 00:38:46.790 [2024-12-06 17:54:28.298184] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.790 [2024-12-06 17:54:28.298224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.790 qpair failed and we were unable to recover it. 00:38:46.790 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@509 -- # nvmfpid=415275 00:38:46.790 [2024-12-06 17:54:28.298425] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.790 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF0 00:38:46.790 [2024-12-06 17:54:28.298467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.790 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@510 -- # waitforlisten 415275 00:38:46.790 qpair failed and we were unable to recover it. 00:38:46.790 [2024-12-06 17:54:28.298640] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.790 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@835 -- # '[' -z 415275 ']' 00:38:46.790 [2024-12-06 17:54:28.298703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.790 qpair failed and we were unable to recover it. 00:38:46.790 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:38:46.790 [2024-12-06 17:54:28.298902] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.790 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@840 -- # local max_retries=100 00:38:46.790 [2024-12-06 17:54:28.298943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.790 qpair failed and we were unable to recover it. 00:38:46.790 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:38:46.790 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:38:46.790 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@844 -- # xtrace_disable 00:38:46.790 [2024-12-06 17:54:28.299163] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.790 [2024-12-06 17:54:28.299197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.790 qpair failed and we were unable to recover it. 00:38:46.790 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:38:46.790 [2024-12-06 17:54:28.299311] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.790 [2024-12-06 17:54:28.299345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.790 qpair failed and we were unable to recover it. 00:38:46.790 [2024-12-06 17:54:28.299984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.790 [2024-12-06 17:54:28.300029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.790 qpair failed and we were unable to recover it. 00:38:46.790 [2024-12-06 17:54:28.300242] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.790 [2024-12-06 17:54:28.300277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.790 qpair failed and we were unable to recover it. 00:38:46.790 [2024-12-06 17:54:28.300406] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.790 [2024-12-06 17:54:28.300431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.790 qpair failed and we were unable to recover it. 00:38:46.790 [2024-12-06 17:54:28.300548] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.790 [2024-12-06 17:54:28.300573] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.790 qpair failed and we were unable to recover it. 00:38:46.790 [2024-12-06 17:54:28.300699] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.790 [2024-12-06 17:54:28.300725] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.790 qpair failed and we were unable to recover it. 00:38:46.790 [2024-12-06 17:54:28.300876] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.790 [2024-12-06 17:54:28.300900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.790 qpair failed and we were unable to recover it. 00:38:46.790 [2024-12-06 17:54:28.301010] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.790 [2024-12-06 17:54:28.301043] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.790 qpair failed and we were unable to recover it. 00:38:46.790 [2024-12-06 17:54:28.301159] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.790 [2024-12-06 17:54:28.301189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.790 qpair failed and we were unable to recover it. 00:38:46.790 [2024-12-06 17:54:28.301311] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.790 [2024-12-06 17:54:28.301337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.790 qpair failed and we were unable to recover it. 00:38:46.790 [2024-12-06 17:54:28.301454] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.790 [2024-12-06 17:54:28.301480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.790 qpair failed and we were unable to recover it. 00:38:46.790 [2024-12-06 17:54:28.301567] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.790 [2024-12-06 17:54:28.301593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.790 qpair failed and we were unable to recover it. 00:38:46.790 [2024-12-06 17:54:28.301777] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.790 [2024-12-06 17:54:28.301834] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.790 qpair failed and we were unable to recover it. 00:38:46.790 [2024-12-06 17:54:28.301974] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.791 [2024-12-06 17:54:28.302004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.791 qpair failed and we were unable to recover it. 00:38:46.791 [2024-12-06 17:54:28.302170] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.791 [2024-12-06 17:54:28.302206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.791 qpair failed and we were unable to recover it. 00:38:46.791 [2024-12-06 17:54:28.302309] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.791 [2024-12-06 17:54:28.302345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.791 qpair failed and we were unable to recover it. 00:38:46.791 [2024-12-06 17:54:28.302481] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.791 [2024-12-06 17:54:28.302517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.791 qpair failed and we were unable to recover it. 00:38:46.791 [2024-12-06 17:54:28.302629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.791 [2024-12-06 17:54:28.302671] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.791 qpair failed and we were unable to recover it. 00:38:46.791 [2024-12-06 17:54:28.302847] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.791 [2024-12-06 17:54:28.302906] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.791 qpair failed and we were unable to recover it. 00:38:46.791 [2024-12-06 17:54:28.303091] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.791 [2024-12-06 17:54:28.303147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.791 qpair failed and we were unable to recover it. 00:38:46.791 [2024-12-06 17:54:28.303250] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.791 [2024-12-06 17:54:28.303278] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.791 qpair failed and we were unable to recover it. 00:38:46.791 [2024-12-06 17:54:28.303373] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.791 [2024-12-06 17:54:28.303399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.791 qpair failed and we were unable to recover it. 00:38:46.791 [2024-12-06 17:54:28.303544] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.791 [2024-12-06 17:54:28.303571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.791 qpair failed and we were unable to recover it. 00:38:46.791 [2024-12-06 17:54:28.303689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.791 [2024-12-06 17:54:28.303724] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.791 qpair failed and we were unable to recover it. 00:38:46.791 [2024-12-06 17:54:28.303842] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.791 [2024-12-06 17:54:28.303878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.791 qpair failed and we were unable to recover it. 00:38:46.791 [2024-12-06 17:54:28.304023] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.791 [2024-12-06 17:54:28.304063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.791 qpair failed and we were unable to recover it. 00:38:46.791 [2024-12-06 17:54:28.304212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.791 [2024-12-06 17:54:28.304239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.791 qpair failed and we were unable to recover it. 00:38:46.791 [2024-12-06 17:54:28.304363] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.791 [2024-12-06 17:54:28.304390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.791 qpair failed and we were unable to recover it. 00:38:46.791 [2024-12-06 17:54:28.304481] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.791 [2024-12-06 17:54:28.304508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.791 qpair failed and we were unable to recover it. 00:38:46.791 [2024-12-06 17:54:28.304589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.791 [2024-12-06 17:54:28.304617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.791 qpair failed and we were unable to recover it. 00:38:46.791 [2024-12-06 17:54:28.304783] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.791 [2024-12-06 17:54:28.304844] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.791 qpair failed and we were unable to recover it. 00:38:46.791 [2024-12-06 17:54:28.305088] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.791 [2024-12-06 17:54:28.305150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.791 qpair failed and we were unable to recover it. 00:38:46.791 [2024-12-06 17:54:28.305300] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.791 [2024-12-06 17:54:28.305369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.791 qpair failed and we were unable to recover it. 00:38:46.791 [2024-12-06 17:54:28.305531] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.791 [2024-12-06 17:54:28.305572] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.791 qpair failed and we were unable to recover it. 00:38:46.791 [2024-12-06 17:54:28.305721] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.791 [2024-12-06 17:54:28.305747] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.791 qpair failed and we were unable to recover it. 00:38:46.791 [2024-12-06 17:54:28.305861] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.791 [2024-12-06 17:54:28.305887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.791 qpair failed and we were unable to recover it. 00:38:46.791 [2024-12-06 17:54:28.305980] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.791 [2024-12-06 17:54:28.306007] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.791 qpair failed and we were unable to recover it. 00:38:46.791 [2024-12-06 17:54:28.306128] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.791 [2024-12-06 17:54:28.306169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.791 qpair failed and we were unable to recover it. 00:38:46.791 [2024-12-06 17:54:28.306333] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.791 [2024-12-06 17:54:28.306395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.791 qpair failed and we were unable to recover it. 00:38:46.791 [2024-12-06 17:54:28.306574] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.791 [2024-12-06 17:54:28.306610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.791 qpair failed and we were unable to recover it. 00:38:46.791 [2024-12-06 17:54:28.306767] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.791 [2024-12-06 17:54:28.306796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.791 qpair failed and we were unable to recover it. 00:38:46.791 [2024-12-06 17:54:28.306881] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.306908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.307012] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.307038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.307179] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.307220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.307446] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.307505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.307716] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.307763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.307880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.307907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.307993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.308019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.308114] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.308142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.308309] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.308370] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.308536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.308579] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.308744] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.308772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.308859] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.308885] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.309003] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.309039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.309128] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.309155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.309263] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.309308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.309460] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.309500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.309627] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.309681] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.309802] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.309828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.309934] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.309966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.310053] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.310079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.310231] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.310277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.310538] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.310583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.310765] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.310792] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.310884] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.310910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.311035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.311062] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.311182] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.311244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.311507] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.311548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.311686] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.311737] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.311854] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.311880] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.311989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.792 [2024-12-06 17:54:28.312015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.792 qpair failed and we were unable to recover it. 00:38:46.792 [2024-12-06 17:54:28.312133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.793 [2024-12-06 17:54:28.312158] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.793 qpair failed and we were unable to recover it. 00:38:46.793 [2024-12-06 17:54:28.312276] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.793 [2024-12-06 17:54:28.312328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.793 qpair failed and we were unable to recover it. 00:38:46.793 [2024-12-06 17:54:28.312504] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.793 [2024-12-06 17:54:28.312545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.793 qpair failed and we were unable to recover it. 00:38:46.793 [2024-12-06 17:54:28.312720] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.793 [2024-12-06 17:54:28.312748] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.793 qpair failed and we were unable to recover it. 00:38:46.793 [2024-12-06 17:54:28.312865] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.793 [2024-12-06 17:54:28.312892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.793 qpair failed and we were unable to recover it. 00:38:46.793 [2024-12-06 17:54:28.312989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.793 [2024-12-06 17:54:28.313027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.793 qpair failed and we were unable to recover it. 00:38:46.793 [2024-12-06 17:54:28.313142] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.793 [2024-12-06 17:54:28.313168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.793 qpair failed and we were unable to recover it. 00:38:46.793 [2024-12-06 17:54:28.313259] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.793 [2024-12-06 17:54:28.313286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.793 qpair failed and we were unable to recover it. 00:38:46.793 [2024-12-06 17:54:28.313428] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.793 [2024-12-06 17:54:28.313469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.793 qpair failed and we were unable to recover it. 00:38:46.793 [2024-12-06 17:54:28.313637] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.793 [2024-12-06 17:54:28.313695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.793 qpair failed and we were unable to recover it. 00:38:46.793 [2024-12-06 17:54:28.313808] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.793 [2024-12-06 17:54:28.313834] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.793 qpair failed and we were unable to recover it. 00:38:46.793 [2024-12-06 17:54:28.313950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.793 [2024-12-06 17:54:28.313978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.793 qpair failed and we were unable to recover it. 00:38:46.793 [2024-12-06 17:54:28.314070] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.793 [2024-12-06 17:54:28.314096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.793 qpair failed and we were unable to recover it. 00:38:46.793 [2024-12-06 17:54:28.314180] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.793 [2024-12-06 17:54:28.314206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.793 qpair failed and we were unable to recover it. 00:38:46.793 [2024-12-06 17:54:28.314352] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.793 [2024-12-06 17:54:28.314396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.793 qpair failed and we were unable to recover it. 00:38:46.793 [2024-12-06 17:54:28.314570] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.793 [2024-12-06 17:54:28.314622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.793 qpair failed and we were unable to recover it. 00:38:46.793 [2024-12-06 17:54:28.314791] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.793 [2024-12-06 17:54:28.314818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.793 qpair failed and we were unable to recover it. 00:38:46.793 [2024-12-06 17:54:28.314964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.793 [2024-12-06 17:54:28.314990] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.793 qpair failed and we were unable to recover it. 00:38:46.793 [2024-12-06 17:54:28.315100] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.793 [2024-12-06 17:54:28.315148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.793 qpair failed and we were unable to recover it. 00:38:46.793 [2024-12-06 17:54:28.315296] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.793 [2024-12-06 17:54:28.315340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.793 qpair failed and we were unable to recover it. 00:38:46.793 [2024-12-06 17:54:28.315561] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.793 [2024-12-06 17:54:28.315603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.793 qpair failed and we were unable to recover it. 00:38:46.793 [2024-12-06 17:54:28.315802] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.793 [2024-12-06 17:54:28.315829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.793 qpair failed and we were unable to recover it. 00:38:46.793 [2024-12-06 17:54:28.315950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.793 [2024-12-06 17:54:28.315977] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.316055] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.316081] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.316202] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.316228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.316350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.316376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.316504] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.316546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.316735] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.316762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.316853] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.316879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.317007] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.317032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.317151] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.317176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.317273] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.317299] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.317429] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.317472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.317610] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.317652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.317791] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.317818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.317931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.317957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.318043] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.318068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.318146] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.318172] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.318311] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.318354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.318569] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.318612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.318753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.318780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.318870] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.318896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.319019] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.319045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.319140] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.319166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.319277] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.319319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.319481] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.319521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.319646] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.319714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.319822] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.319848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.319935] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.319961] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.320071] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.320097] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.320211] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.320254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.320471] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.320511] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.320647] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.320715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.320858] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.320884] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.320989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.321027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.321128] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.794 [2024-12-06 17:54:28.321156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.794 qpair failed and we were unable to recover it. 00:38:46.794 [2024-12-06 17:54:28.321246] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.321311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.321483] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.321525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.321722] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.321751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.321866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.321892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.321992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.322018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.322136] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.322163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.322306] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.322347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.322472] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.322514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.322725] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.322768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.322935] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.322976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.323103] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.323144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.323286] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.323331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.323475] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.323516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.323652] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.323710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.323866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.323907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.324061] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.324103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.324272] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.324316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.324477] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.324533] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.324747] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.324794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.324939] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.324998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.325153] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.325201] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.325341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.325386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.325534] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.325570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.325690] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.325725] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.325833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.325867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.325974] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.326031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.326177] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.326211] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.326353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.326403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.326564] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.326598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.326752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.795 [2024-12-06 17:54:28.326804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.795 qpair failed and we were unable to recover it. 00:38:46.795 [2024-12-06 17:54:28.326949] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.326989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.327137] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.327172] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.327362] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.327387] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.327496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.327521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.327679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.327706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.327797] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.327823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.327934] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.327960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.328076] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.328102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.328182] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.328208] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.328340] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.328379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.328468] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.328500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.328625] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.328653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.328797] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.328832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.328934] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.328983] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.329129] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.329155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.329242] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.329269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.329355] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.329381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.329493] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.329519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.329597] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.329622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.329726] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.329753] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.329844] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.329870] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.329983] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.330009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.330115] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.330141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.330255] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.330282] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.330399] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.330426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.330538] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.330563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.330644] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.330677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.330774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.330826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.330941] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.330966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.331091] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.331117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.331256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.331282] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.331364] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.331389] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.331511] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.331537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.331646] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.796 [2024-12-06 17:54:28.331679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.796 qpair failed and we were unable to recover it. 00:38:46.796 [2024-12-06 17:54:28.331791] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.331825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.331930] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.331977] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.332082] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.332107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.332248] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.332278] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.332366] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.332392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.332508] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.332534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.332641] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.332672] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.332766] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.332793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.332909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.332934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.333058] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.333084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.333203] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.333229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.333349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.333378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.333456] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.333482] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.333628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.333654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.333756] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.333782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.333857] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.333883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.333981] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.334006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.334096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.334123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.334197] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.334222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.334297] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.334323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.334435] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.334461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.334572] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.334597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.334703] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.334729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.334869] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.334895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.335010] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.335035] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.335124] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.335150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.335265] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.335291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.335401] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.335427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.335518] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.335544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.335671] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.335718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.335852] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.335878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.336004] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.336034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.336177] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.336203] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.336323] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.797 [2024-12-06 17:54:28.336349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.797 qpair failed and we were unable to recover it. 00:38:46.797 [2024-12-06 17:54:28.336435] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.336461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.336557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.336583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.336679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.336727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.336845] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.336891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.337030] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.337055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.337166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.337192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.337281] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.337307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.337413] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.337438] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.337522] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.337548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.337630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.337656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.337773] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.337804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.337915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.337940] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.338029] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.338056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.338169] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.338195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.338308] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.338334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.338428] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.338454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.338594] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.338619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.338715] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.338741] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.338855] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.338881] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.339001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.339027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.339107] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.339137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.339253] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.339278] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.339422] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.339449] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.339529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.339559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.339652] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.339691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.339780] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.339805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.339887] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.339913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.340015] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.340041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.340161] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.798 [2024-12-06 17:54:28.340189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.798 qpair failed and we were unable to recover it. 00:38:46.798 [2024-12-06 17:54:28.340284] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.340310] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.340450] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.340476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.340568] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.340594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.340714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.340740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.340858] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.340885] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.341006] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.341031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.341116] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.341142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.341266] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.341292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.341408] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.341435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.341522] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.341548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.341679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.341705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.341826] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.341852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.341932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.341958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.342051] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.342077] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.342188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.342215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.342300] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.342327] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.342413] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.342439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.342555] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.342581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.342695] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.342722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.342875] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.342901] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.343022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.343048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.343131] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.343160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.343251] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.343278] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.343362] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.343388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.343525] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.343551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.343668] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.343694] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.343781] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.343807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.343916] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.343941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.344031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.344058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.344142] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.344169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.344248] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.344274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.344419] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.344445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.344563] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.344588] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.344674] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.799 [2024-12-06 17:54:28.344701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.799 qpair failed and we were unable to recover it. 00:38:46.799 [2024-12-06 17:54:28.344786] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.344813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.344937] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.344962] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.345073] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.345099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.345188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.345214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.345333] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.345360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.345440] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.345465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.345578] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.345604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.345695] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.345722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.345866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.345892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.345985] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.346011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.346125] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.346119] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:38:46.800 [2024-12-06 17:54:28.346154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.346195] [ DPDK EAL parameters: nvmf -c 0xF0 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:38:46.800 [2024-12-06 17:54:28.346243] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.346268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.346392] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.346418] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.346537] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.346564] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.346653] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.346689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.346803] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.346829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.346944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.347019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.347131] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.347156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.347247] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.347273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.347417] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.347443] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.347554] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.347581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.347680] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.347707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.347797] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.347823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.347907] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.347934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.348013] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.348039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.348130] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.348157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.348271] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.348297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.348390] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.348417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.348511] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.348537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.348616] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.348646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.348755] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.348782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.348899] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.348926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.349030] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.349056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.349168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.349194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.800 [2024-12-06 17:54:28.349295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.800 [2024-12-06 17:54:28.349321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.800 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.349432] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.349459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.349551] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.349577] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.349676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.349729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.349851] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.349877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.349988] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.350015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.350135] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.350161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.350252] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.350278] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.350415] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.350442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.350527] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.350553] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.350635] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.350662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.350776] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.350803] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.350896] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.350923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.351005] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.351031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.351146] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.351173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.351286] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.351312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.351428] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.351456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.351569] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.351596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.351688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.351715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.351829] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.351856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.351950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.351976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.352119] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.352145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.352253] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.352279] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.352391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.352417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.352508] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.352534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.352655] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.352688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.352775] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.352801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.352917] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.352943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.353060] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.353086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.353165] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.353190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.353280] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.353307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.353391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.353421] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.353515] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.353540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.353654] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.353694] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.801 [2024-12-06 17:54:28.353779] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.801 [2024-12-06 17:54:28.353805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.801 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.353922] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.353947] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.354031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.354057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.354173] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.354199] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.354310] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.354335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.354455] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.354481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.354596] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.354623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.354748] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.354805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.354949] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.354975] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.355119] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.355145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.355289] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.355315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.355408] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.355434] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.355544] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.355571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.355671] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.355699] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.355816] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.355843] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.355923] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.355949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.356065] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.356091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.356232] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.356258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.356376] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.356401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.356512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.356538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.356628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.356653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.356744] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.356800] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.357317] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.357346] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.357481] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.357508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.357601] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.357626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.357737] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.357763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.357882] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.357907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.358036] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.358063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.358195] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.358221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.358326] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.358351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.358469] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.358511] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.358626] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.358652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.358799] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.358825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.358938] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.358964] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.359082] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.359107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.359221] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.359264] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.802 qpair failed and we were unable to recover it. 00:38:46.802 [2024-12-06 17:54:28.359349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.802 [2024-12-06 17:54:28.359380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.359464] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.359489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.359576] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.359605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.359716] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.359746] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.359861] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.359887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.360027] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.360054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.360186] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.360212] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.360324] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.360350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.360498] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.360525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.360634] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.360658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.360807] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.360832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.360949] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.360999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.361165] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.361191] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.361305] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.361347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.361497] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.361523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.361650] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.361692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.361805] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.361830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.361940] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.361967] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.362052] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.362079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.362217] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.362243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.362401] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.362428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.362591] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.362618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.362753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.362780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.362925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.362966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.363070] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.363097] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.363214] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.363240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.363348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.363374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.363493] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.363520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.363615] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.363642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.363773] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.803 [2024-12-06 17:54:28.363800] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.803 qpair failed and we were unable to recover it. 00:38:46.803 [2024-12-06 17:54:28.363924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.363951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.364069] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.364094] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.364214] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.364241] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.364331] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.364356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.364460] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.364487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.364577] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.364604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.364699] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.364727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.364828] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.364858] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.365016] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.365046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.365161] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.365187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.365305] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.365337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.365466] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.365496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.365616] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.365644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.365818] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.365853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.365984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.366010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.366147] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.366172] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.366350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.366376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.366479] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.366504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.366612] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.366639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.366790] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.366817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.366937] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.366963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.367080] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.367106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.367201] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.367227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.367350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.367377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.367489] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.367516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.367599] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.367626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.367753] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.367780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.367898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.367925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.368011] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.368040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.368186] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.368213] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.368330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.368357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.368475] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.368502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.804 [2024-12-06 17:54:28.368626] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.804 [2024-12-06 17:54:28.368652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.804 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.368800] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.368826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.368924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.368952] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.369083] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.369110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.369231] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.369257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.369371] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.369398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.369498] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.369524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.369637] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.369678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.369822] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.369848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.369961] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.369987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.370090] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.370119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.370232] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.370257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.370345] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.370372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.370498] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.370524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.370605] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.370631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.370760] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.370787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.370878] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.370905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.370994] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.371019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.371159] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.371185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.371271] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.371297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.371388] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.371417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.371505] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.371535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.371625] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.371651] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.371778] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.371804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.371885] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.371910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.372026] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.372052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.372140] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.372165] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.372281] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.372307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.372426] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.372455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.372542] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.372568] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.372709] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.372735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.372830] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.372856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.372998] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.373027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.373141] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.373166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.373257] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.373283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.373400] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.373427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.373508] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.373534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.805 qpair failed and we were unable to recover it. 00:38:46.805 [2024-12-06 17:54:28.373648] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.805 [2024-12-06 17:54:28.373678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.373793] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.373817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.373910] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.373935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.374050] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.374074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.374174] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.374200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.374334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.374359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.374494] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.374519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.374626] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.374651] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.374780] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.374805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.374921] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.374945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.375030] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.375055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.375145] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.375170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.375310] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.375334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.375447] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.375472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.375566] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.375591] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.375676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.375703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.375841] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.375866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.375975] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.376000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.376134] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.376160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.376264] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.376289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.376374] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.376400] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.376477] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.376501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.376585] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.376610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.376733] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.376760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.376882] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.376913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.377005] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.377031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.377146] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.377174] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.377256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.377282] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.377368] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.377396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.377513] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.377538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.377682] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.377712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.377805] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.377830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.377948] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.806 [2024-12-06 17:54:28.377974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.806 qpair failed and we were unable to recover it. 00:38:46.806 [2024-12-06 17:54:28.378082] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.378108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.378216] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.378242] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.378369] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.378397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.378519] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.378545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.378688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.378714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.378843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.378869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.378955] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.378982] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.379067] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.379092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.379202] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.379228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.379372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.379399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.379531] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.379557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.379675] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.379701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.379788] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.379814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.379925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.379951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.380047] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.380073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.380151] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.380177] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.380269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.380295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.380380] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.380406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.380502] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.380527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.380617] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.380642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.380779] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.380806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.380920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.380945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.381029] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.381055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.381140] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.381166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.381259] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.381284] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.381397] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.381423] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.381510] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.381536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.381621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.381646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.381766] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.381792] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.381873] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.381899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.382014] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.382039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.382135] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.382166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.382263] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.382290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.382399] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.382425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.382562] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.807 [2024-12-06 17:54:28.382588] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.807 qpair failed and we were unable to recover it. 00:38:46.807 [2024-12-06 17:54:28.382706] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.382732] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.382816] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.382842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.382948] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.382975] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.383060] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.383087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.383182] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.383208] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.383283] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.383309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.383424] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.383451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.383598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.383625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.383747] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.383773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.383905] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.383932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.384078] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.384106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.384221] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.384247] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.384337] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.384362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.384474] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.384500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.384574] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.384600] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.384689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.384716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.384833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.384859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.384951] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.384977] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.385078] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.385104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.385216] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.385242] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.385364] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.385390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.385520] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.385546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.385668] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.385695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.385790] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.385815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.385932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.385957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.386071] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.386096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.386210] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.386236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.386352] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.386378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.386462] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.808 [2024-12-06 17:54:28.386487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.808 qpair failed and we were unable to recover it. 00:38:46.808 [2024-12-06 17:54:28.386630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.386658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.386831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.386860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.386975] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.387000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.387118] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.387143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.387285] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.387311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.387402] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.387427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.387509] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.387536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.387633] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.387670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.387816] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.387842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.387966] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.387992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.388137] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.388163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.388269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.388295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.388406] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.388431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.388522] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.388548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.388629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.388655] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.388777] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.388802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.388920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.388945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.389085] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.389110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.389233] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.389258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.389371] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.389397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.389511] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.389537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.389625] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.389651] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.389770] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.389795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.389884] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.389910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.390011] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.390039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.390132] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.390158] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.390237] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.390262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.390375] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.390401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.390516] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.390542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.390622] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.390647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.390760] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.390786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.390876] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.390902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.391040] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.391065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.809 qpair failed and we were unable to recover it. 00:38:46.809 [2024-12-06 17:54:28.391204] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.809 [2024-12-06 17:54:28.391229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.391357] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.391383] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.391495] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.391521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.391631] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.391656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.391778] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.391804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.391896] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.391921] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.392037] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.392062] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.392147] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.392172] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.392286] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.392312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.392392] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.392417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.392505] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.392531] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.392643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.392674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.392814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.392840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.392952] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.392978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.393094] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.393124] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.393237] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.393262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.393377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.393403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.393520] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.393546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.393621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.393646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.393762] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.393788] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.393866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.393891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.394009] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.394034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.394123] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.394148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.394255] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.394280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.394421] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.394447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.394531] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.394557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.394688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.394714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.394797] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.394822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.394913] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.394939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.395058] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.395083] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.395168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.395193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.395307] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.395333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.395441] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.395466] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.810 [2024-12-06 17:54:28.395605] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.810 [2024-12-06 17:54:28.395631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.810 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.395784] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.395811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.395900] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.395926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.396062] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.396087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.396199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.396225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.396370] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.396396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.396535] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.396560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.396654] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.396686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.396801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.396827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.396909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.396935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.397034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.397060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.397134] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.397159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.397235] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.397260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.397348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.397373] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.397487] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.397512] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.397653] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.397683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.397827] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.397853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.397933] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.397958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.398096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.398122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.398234] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.398260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.398397] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.398423] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.398516] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.398546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.398636] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.398661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.398778] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.398804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.398882] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.398908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.399007] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.399032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.399139] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.399165] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.399254] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.399280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.399401] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.399427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.399507] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.399532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.399643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.399676] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.399767] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.399792] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.399870] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.399895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.400017] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.811 [2024-12-06 17:54:28.400044] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.811 qpair failed and we were unable to recover it. 00:38:46.811 [2024-12-06 17:54:28.400163] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.400189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.400314] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.400340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.400452] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.400477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.400587] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.400614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.400742] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.400768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.400880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.400905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.401007] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.401032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.401116] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.401143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.401258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.401284] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.401400] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.401426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.401512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.401538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.401645] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.401684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.401800] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.401825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.401936] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.401962] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.402112] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.402137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.402251] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.402277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.402395] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.402420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.402497] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.402522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.402609] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.402634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.402756] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.402782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.402895] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.402920] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.403036] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.403062] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.403175] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.403201] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.403319] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.403344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.403437] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.403462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.403601] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.403626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.403715] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.403740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.403879] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.403909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.403986] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.404012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.404124] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.404149] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.404268] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.404294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.404379] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.404405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.404484] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.812 [2024-12-06 17:54:28.404510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.812 qpair failed and we were unable to recover it. 00:38:46.812 [2024-12-06 17:54:28.404632] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.404657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.404810] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.404835] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.404948] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.404973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.405088] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.405114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.405195] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.405220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.405357] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.405382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.405461] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.405487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.405634] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.405659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.405788] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.405813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.405930] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.405955] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.406038] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.406064] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.406173] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.406198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.406291] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.406316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.406403] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.406428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.406517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.406543] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.406657] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.406689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.406832] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.406857] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.406934] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.406960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.407057] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.407084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.407171] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.407196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.407335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.407360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.407472] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.407508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.407676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.407705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.407827] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.407854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.407976] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.408005] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.408094] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.408122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.408217] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.408245] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.408358] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.408384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.408496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.813 [2024-12-06 17:54:28.408523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.813 qpair failed and we were unable to recover it. 00:38:46.813 [2024-12-06 17:54:28.408670] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.408698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.408811] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.408838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.408958] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.408996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.409114] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.409141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.409278] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.409305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.409406] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.409439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.409521] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.409548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.409671] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.409699] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.409814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.409841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.409958] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.409986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.410127] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.410154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.410296] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.410323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.410440] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.410468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.410577] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.410612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.410739] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.410766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.410880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.410907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.410985] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.411014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.411153] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.411180] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.411267] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.411294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.411437] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.411463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.411576] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.411602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.411712] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.411739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.411859] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.411888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.412082] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.412110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.412195] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.412223] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.412342] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.412369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.412455] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.412483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.412621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.412660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.412860] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.412888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.412999] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.413026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.413112] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.413139] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.413223] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.413250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.413370] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.413397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.814 qpair failed and we were unable to recover it. 00:38:46.814 [2024-12-06 17:54:28.413482] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.814 [2024-12-06 17:54:28.413509] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.413600] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.413628] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.413743] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.413771] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.413911] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.413938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.414027] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.414053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.414187] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.414213] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.414327] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.414353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.414430] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.414456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.414545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.414571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.414649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.414681] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.414792] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.414818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.414929] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.414955] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.415037] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.415068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.415184] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.415212] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.415288] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.415315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.415410] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.415437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.415549] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.415576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.415652] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.415688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.415770] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.415797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.415906] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.415933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.416019] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.416046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.416132] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.416161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.416278] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.416306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.416416] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.416442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.416527] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.416553] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.416635] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.416674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.416798] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.416824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.416903] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.416930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.417078] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.417106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.417225] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.417253] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.417335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.417362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.417483] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.417510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.815 [2024-12-06 17:54:28.417601] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.815 [2024-12-06 17:54:28.417630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.815 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.417755] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.417783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.417896] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.417924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.418035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.418062] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.418140] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.418167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.418277] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.418305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.418417] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.418443] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.418589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.418615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.418701] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.418728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.418836] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.418862] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.418977] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.419004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.419084] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.419111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.419255] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.419281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.419390] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.419416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.419556] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.419582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.419696] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.419725] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.419864] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.419890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.420005] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.420031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.420169] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.420196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.420337] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.420363] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.420459] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.420493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.420638] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.420676] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.420823] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.420850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.420993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.421020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.421135] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.421163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.421277] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.421304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.421397] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.421425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.421539] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.421565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.421707] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.421733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.421851] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.421877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.421955] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.421981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.422057] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.422083] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.422216] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.422244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.422380] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.422407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.422527] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.422554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.422632] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.422659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.816 qpair failed and we were unable to recover it. 00:38:46.816 [2024-12-06 17:54:28.422806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.816 [2024-12-06 17:54:28.422833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.422922] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.422949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.423063] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.423090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.423209] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.423236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.423319] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.423346] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.423426] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.423454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.423559] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.423585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.423676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.423703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.423793] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.423819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.423899] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.423925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.424035] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.424062] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.424146] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.424174] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.424316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.424343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.424463] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.424490] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.424580] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.424607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.424699] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.424726] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.424833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.424860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.424944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.424979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.425065] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.425091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.425174] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.425202] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.425309] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.425336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.425442] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.425470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.425550] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.425578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.425722] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.425750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.425860] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.425886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.425973] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.426000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.426117] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.426143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.426229] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.426255] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.426325] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:38:46.817 [2024-12-06 17:54:28.426371] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.426395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.426481] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.426509] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.426591] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.426618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.817 qpair failed and we were unable to recover it. 00:38:46.817 [2024-12-06 17:54:28.426735] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.817 [2024-12-06 17:54:28.426763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.426860] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.426887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.427015] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.427042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.427127] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.427154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.427260] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.427287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.427400] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.427427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.427509] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.427537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.427690] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.427719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.427834] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.427861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.427952] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.427979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.428117] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.428143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.428258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.428284] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.428372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.428399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.428506] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.428532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.428640] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.428675] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.428765] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.428791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.428913] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.428942] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.429062] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.429090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.429184] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.429211] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.429319] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.429346] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.429466] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.429493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.429607] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.429635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.429724] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.429751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.429833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.429860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.429979] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.430006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.430102] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.430130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.430243] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.430270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.430356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.430383] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.430496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.430524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.430610] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.430637] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.430789] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.430815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.430904] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.430930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.431054] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.431080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.431166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.431197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.431304] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.431331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.818 [2024-12-06 17:54:28.431437] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.818 [2024-12-06 17:54:28.431464] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.818 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.431551] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.431577] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.431728] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.431757] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.431871] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.431899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.432020] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.432047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.432163] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.432190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.432310] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.432337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.432452] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.432479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.432593] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.432621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.432745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.432773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.432859] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.432887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.433081] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.433108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.433226] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.433254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.433374] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.433401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.433494] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.433521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.433642] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.433685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.433802] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.433830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.433917] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.433944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.434087] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.434113] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.434232] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.434260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.434416] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.434445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.434534] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.434560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.434675] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.434703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.434820] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.434847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.434959] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.434985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.435100] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.435127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.435237] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.435264] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.435356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.435382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.435474] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.435500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.435583] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.435611] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.435725] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.435753] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.819 [2024-12-06 17:54:28.435837] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.819 [2024-12-06 17:54:28.435864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.819 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.435949] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.435976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.436083] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.436110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.436202] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.436229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.436321] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.436348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.436493] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.436520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.436634] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.436677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.436784] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.436814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.436928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.436955] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.437049] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.437075] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.437166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.437192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.437277] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.437304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.437389] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.437416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.437491] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.437517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.437640] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.437680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.437794] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.437820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.437968] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.437994] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.438111] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.438136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.438247] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.438272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.438396] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.438425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.438544] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.438571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.438771] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.438800] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.438912] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.438940] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.439054] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.439082] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.439200] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.439228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.439342] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.439369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.439483] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.439510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.439632] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.439659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.439786] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.439813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.439935] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.439963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.440076] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.440104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.440221] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.440250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.440370] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.440397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.820 qpair failed and we were unable to recover it. 00:38:46.820 [2024-12-06 17:54:28.440536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.820 [2024-12-06 17:54:28.440562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.440680] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.440708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.440822] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.440849] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.440961] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.440991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.441138] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.441164] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.441283] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.441310] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.441423] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.441450] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.441529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.441555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.441694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.441721] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.441835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.441862] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.441953] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.441981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.442094] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.442120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.442264] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.442290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.442366] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.442393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.442475] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.442506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.442647] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.442685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.442767] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.442794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.442871] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.442898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.443021] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.443047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.443167] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.443193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.443307] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.443334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.443421] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.443448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.443564] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.443590] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.443707] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.443734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.443821] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.443848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.443990] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.444016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.444159] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.444185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.444295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.444323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.444414] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.444441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.444531] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.444558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.444674] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.444701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.444803] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.444845] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.445021] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.445063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.445201] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.445229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.821 [2024-12-06 17:54:28.445322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.821 [2024-12-06 17:54:28.445349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.821 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.445492] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.445518] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.445633] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.445660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.445759] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.445786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.445906] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.445933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.446045] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.446072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.446215] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.446242] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.446364] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.446395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.446479] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.446507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.446650] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.446692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.446818] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.446846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.446929] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.446957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.447066] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.447094] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.447211] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.447239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.447347] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.447374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.447453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.447481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.447587] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.447614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.447731] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.447759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.447891] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.447918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.448041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.448068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.448157] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.448184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.448275] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.448304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.448391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.448418] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.448514] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.448541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.448634] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.448674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.448788] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.448815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.448941] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.448979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.449105] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.449133] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.449225] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.449252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.449336] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.449363] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.449504] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.449532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.449675] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.449702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.449815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.449842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.449942] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.449969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.450058] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.450085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.822 [2024-12-06 17:54:28.450168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.822 [2024-12-06 17:54:28.450196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.822 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.450295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.450324] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.450473] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.450500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.450606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.450632] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.450749] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.450777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.450860] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.450887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.450971] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.450999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.451151] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.451179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.451300] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.451328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.451422] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.451449] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.451557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.451584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.451710] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.451751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.451842] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.451872] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.452082] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.452111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.452232] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.452260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.452351] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.452379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.452464] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.452492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.452636] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.452675] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.452770] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.452797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.452897] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.452924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.453040] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.453067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.453144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.453171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.453292] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.453321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.453408] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.453437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.453598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.453626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.453761] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.453790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.453940] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.453969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.454094] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.454121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.454232] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.454259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.454341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.454369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.454490] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.454517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.454599] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.454627] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.454725] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.454754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.454851] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.454878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.455071] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.823 [2024-12-06 17:54:28.455099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.823 qpair failed and we were unable to recover it. 00:38:46.823 [2024-12-06 17:54:28.455253] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.455280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.455401] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.455428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.455543] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.455571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.455658] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.455696] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.455808] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.455840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.455951] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.455979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.456093] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.456120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.456257] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.456284] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.456368] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.456395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.456536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.456563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.456660] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.456696] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.456796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.456823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.456911] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.456938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.457078] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.457119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.457211] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.457240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.457329] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.457357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.457467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.457494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.457648] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.457702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.457833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.457862] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.457947] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.457975] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.458084] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.458111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.458227] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.458254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.458337] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.458364] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.458441] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.458467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.458614] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.458644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.458752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.458780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.458867] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.458897] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.458985] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.459012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.459150] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.459177] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.459287] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.459314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.459409] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.459437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.459528] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.459563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.459673] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.459701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.459786] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.459812] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.824 qpair failed and we were unable to recover it. 00:38:46.824 [2024-12-06 17:54:28.459900] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.824 [2024-12-06 17:54:28.459927] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.460038] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.460064] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.460146] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.460174] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.460292] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.460319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.460410] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.460439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.460534] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.460562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.460639] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.460685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.460799] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.460826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.460969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.460995] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.461076] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.461103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.461247] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.461275] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.461377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.461405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.461556] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.461583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.461701] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.461729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.461846] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.461874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.461987] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.462014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.462099] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.462127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.462214] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.462242] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.462356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.462383] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.462477] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.462503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.462616] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.462644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.462729] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.462757] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.462868] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.462894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.462975] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.463002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.463128] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.463155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.463249] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.463278] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.463425] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.463454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.463571] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.463598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.463718] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.463746] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.463824] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.463851] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.463963] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.463990] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.464116] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.464143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.464257] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.464285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.464373] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.464402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.464515] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.464544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.464630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.464670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.464763] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.825 [2024-12-06 17:54:28.464791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.825 qpair failed and we were unable to recover it. 00:38:46.825 [2024-12-06 17:54:28.464901] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.464934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.465049] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.465076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.465219] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.465246] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.465367] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.465395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.465486] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.465514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.465630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.465657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.465788] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.465815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.465904] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.465931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.466049] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.466076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.466191] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.466219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.466370] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.466398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.466509] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.466536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.466650] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.466684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.466769] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.466797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.466946] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.466973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.467090] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.467117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.467229] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.467256] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.467344] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.467379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.467465] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.467494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.467607] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.467634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.467779] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.467807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.468738] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.468772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.468898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.468925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.469057] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.469084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.469195] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.469222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.469361] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.469387] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.469501] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.469527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.469647] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.469691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.469811] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.469839] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.469916] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.469943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.470027] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.470053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.470143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.470170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.470283] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.470310] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.470403] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.470430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.470511] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.470539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.470650] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.470685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.470797] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.470823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.470938] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.470976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.826 qpair failed and we were unable to recover it. 00:38:46.826 [2024-12-06 17:54:28.471092] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.826 [2024-12-06 17:54:28.471119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.471261] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.471288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.471376] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.471408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.471501] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.471528] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.471640] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.471689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.471790] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.471817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.471932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.471968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.472079] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.472106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.472222] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.472249] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.472390] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.472416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.472524] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.472552] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.472679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.472707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.472828] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.472855] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.472984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.473011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.473127] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.473154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.473272] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.473300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.473385] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.473413] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.473528] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.473555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.473685] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.473712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.473803] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.473829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.473920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.473947] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.474031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.474057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.474138] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.474164] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.474250] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.474277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.474377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.474418] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.474510] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.474538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.474615] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.474642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.474749] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.474776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.474860] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.474887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.474987] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.475014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.475095] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.475122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.475208] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.475239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.475333] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.475373] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.475458] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.475486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.475601] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.475627] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.827 [2024-12-06 17:54:28.475622] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.475673] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:38:46.827 [2024-12-06 17:54:28.475691] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:38:46.827 [2024-12-06 17:54:28.475704] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:38:46.827 [2024-12-06 17:54:28.475715] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:38:46.827 [2024-12-06 17:54:28.475723] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.475748] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.475835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.475860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.475980] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.476008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.476124] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.476150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.476270] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.476299] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.827 [2024-12-06 17:54:28.476391] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.827 [2024-12-06 17:54:28.476431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.827 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.476526] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.476554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.476680] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.476708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.476798] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.476826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.476914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.476941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.477102] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.477128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.477264] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.477306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.477260] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 5 00:38:46.828 [2024-12-06 17:54:28.477351] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 7 00:38:46.828 [2024-12-06 17:54:28.477411] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.477437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.477308] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 6 00:38:46.828 [2024-12-06 17:54:28.477354] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:38:46.828 [2024-12-06 17:54:28.477555] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.477580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.477683] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.477710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.477801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.477828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.477914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.477940] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.478031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.478062] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.478179] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.478206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.478304] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.478333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.478420] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.478447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.478583] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.478623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.478733] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.478761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.478851] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.478878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.478974] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.479001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.479078] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.479104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.479186] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.479213] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.479303] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.479332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.479445] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.479484] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.479583] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.479623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.479714] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.479744] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.479835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.479863] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.479948] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.479979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.480053] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.480080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.480197] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.480225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.480314] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.480342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.480453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.480480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.480609] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.480636] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.480775] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.480805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.480886] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.480913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.481047] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.481074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.828 qpair failed and we were unable to recover it. 00:38:46.828 [2024-12-06 17:54:28.481152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.828 [2024-12-06 17:54:28.481179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.481258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.481287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.481370] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.481399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.481485] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.481518] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.481636] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.481679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.481757] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.481785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.481902] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.481929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.482042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.482069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.482196] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.482224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.482310] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.482337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.482419] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.482446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.482556] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.482583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.482672] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.482700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.482812] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.482839] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.482917] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.482944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.483031] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.483058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.483175] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.483216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.483321] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.483349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.483426] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.483454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.483533] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.483561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.483636] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.483677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.483796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.483824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.483938] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.483968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.484052] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.484080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.484169] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.484196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.484303] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.484330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.484414] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.484441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.484540] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.484579] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.484679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.484708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.484798] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.484826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.484917] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.484945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.485053] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.485079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.485181] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.485221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.829 qpair failed and we were unable to recover it. 00:38:46.829 [2024-12-06 17:54:28.485337] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.829 [2024-12-06 17:54:28.485365] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.485479] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.485509] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.485594] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.485621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.485724] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.485752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.485835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.485863] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.485990] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.486018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.486101] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.486128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.486207] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.486234] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.486316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.486343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.486424] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.486453] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.486566] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.486598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.486702] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.486731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.486843] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.486871] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.486969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.486996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.487106] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.487133] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.487255] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.487282] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.487365] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.487393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.487484] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.487512] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.487592] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.487619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.487758] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.487785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.487867] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.487894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.487981] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.488008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.488088] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.488114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.488231] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.488258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.488355] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.488385] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.488487] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.488527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.488619] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.488647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.488797] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.488824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.488938] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.488975] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.489065] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.489092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.489212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.489240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.489329] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.489358] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.489449] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.489489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.489613] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.489641] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.489742] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.489770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.489854] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.489881] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.489977] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.490003] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.490097] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.490127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.490247] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.490275] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.490357] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.490384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.490467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.830 [2024-12-06 17:54:28.490494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.830 qpair failed and we were unable to recover it. 00:38:46.830 [2024-12-06 17:54:28.490585] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.490613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.490730] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.490760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.490845] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.490872] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.490987] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.491014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.491102] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.491130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.491220] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.491249] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.491332] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.491359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.491475] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.491502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.491586] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.491613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.491699] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.491727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.491814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.491841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.491922] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.491960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.492042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.492070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.492148] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.492175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.492303] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.492331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.492442] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.492469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.492551] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.492579] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.492670] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.492698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.492780] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.492808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.492921] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.492959] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.493034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.493061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.493141] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.493169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.493286] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.493313] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.493404] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.493433] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.493552] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.493581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.493677] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.493706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.493792] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.493819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.493897] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.493924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.494021] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.494049] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.494165] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.494194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.494314] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.494342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.494427] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.494456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.494536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.494563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.494681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.494708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.494791] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.494818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.494928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.494966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.495045] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.495076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.495185] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.831 [2024-12-06 17:54:28.495212] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.831 qpair failed and we were unable to recover it. 00:38:46.831 [2024-12-06 17:54:28.495297] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.495325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.495406] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.495432] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.495535] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.495564] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.495675] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.495703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.495787] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.495815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.495898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.495925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.496046] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.496073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.496166] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.496195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.496281] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.496309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.496405] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.496445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.496536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.496566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.496648] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.496696] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.496785] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.496814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.496890] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.496917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.497042] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.497069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.497160] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.497186] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.497263] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.497289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.497374] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.497401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.497476] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.497502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.497597] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.497625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.497725] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.497755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.497840] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.497868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.497946] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.497980] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.498124] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.498152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.498231] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.498258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.498366] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.498394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.498487] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.498527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.498609] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.498638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.498735] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.498763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.498850] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.498878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.498965] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.498992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.499137] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.499164] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.499245] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.499272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.499353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.499379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.499487] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.499514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.499591] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.499619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.499718] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.499749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.499836] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.499864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.499948] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.499975] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.832 [2024-12-06 17:54:28.500070] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.832 [2024-12-06 17:54:28.500097] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.832 qpair failed and we were unable to recover it. 00:38:46.833 [2024-12-06 17:54:28.500178] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.833 [2024-12-06 17:54:28.500205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.833 qpair failed and we were unable to recover it. 00:38:46.833 [2024-12-06 17:54:28.500293] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.833 [2024-12-06 17:54:28.500319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.833 qpair failed and we were unable to recover it. 00:38:46.833 [2024-12-06 17:54:28.500429] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.833 [2024-12-06 17:54:28.500456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.833 qpair failed and we were unable to recover it. 00:38:46.833 [2024-12-06 17:54:28.500551] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.833 [2024-12-06 17:54:28.500580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.833 qpair failed and we were unable to recover it. 00:38:46.833 [2024-12-06 17:54:28.500659] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.833 [2024-12-06 17:54:28.500693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.833 qpair failed and we were unable to recover it. 00:38:46.833 [2024-12-06 17:54:28.500801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.833 [2024-12-06 17:54:28.500828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.833 qpair failed and we were unable to recover it. 00:38:46.833 [2024-12-06 17:54:28.500912] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.833 [2024-12-06 17:54:28.500941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.833 qpair failed and we were unable to recover it. 00:38:46.833 [2024-12-06 17:54:28.501027] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.833 [2024-12-06 17:54:28.501056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.833 qpair failed and we were unable to recover it. 00:38:46.833 [2024-12-06 17:54:28.501146] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.833 [2024-12-06 17:54:28.501185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.833 qpair failed and we were unable to recover it. 00:38:46.833 [2024-12-06 17:54:28.501278] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.833 [2024-12-06 17:54:28.501307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.833 qpair failed and we were unable to recover it. 00:38:46.833 [2024-12-06 17:54:28.501413] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.833 [2024-12-06 17:54:28.501440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.833 qpair failed and we were unable to recover it. 00:38:46.833 [2024-12-06 17:54:28.501529] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.833 [2024-12-06 17:54:28.501555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.833 qpair failed and we were unable to recover it. 00:38:46.833 [2024-12-06 17:54:28.501687] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.833 [2024-12-06 17:54:28.501717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.833 qpair failed and we were unable to recover it. 00:38:46.833 [2024-12-06 17:54:28.501848] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.833 [2024-12-06 17:54:28.501875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.833 qpair failed and we were unable to recover it. 00:38:46.833 [2024-12-06 17:54:28.501951] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.833 [2024-12-06 17:54:28.501987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.833 qpair failed and we were unable to recover it. 00:38:46.833 [2024-12-06 17:54:28.502068] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.833 [2024-12-06 17:54:28.502095] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.833 qpair failed and we were unable to recover it. 00:38:46.833 [2024-12-06 17:54:28.502207] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.833 [2024-12-06 17:54:28.502233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.833 qpair failed and we were unable to recover it. 00:38:46.833 [2024-12-06 17:54:28.502320] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.833 [2024-12-06 17:54:28.502347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.833 qpair failed and we were unable to recover it. 00:38:46.833 [2024-12-06 17:54:28.502434] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.833 [2024-12-06 17:54:28.502461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.833 qpair failed and we were unable to recover it. 00:38:46.833 [2024-12-06 17:54:28.502571] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.833 [2024-12-06 17:54:28.502599] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.833 qpair failed and we were unable to recover it. 00:38:46.833 [2024-12-06 17:54:28.502695] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.833 [2024-12-06 17:54:28.502724] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.833 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.502837] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.502864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.502941] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.502979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.503094] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.503121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.503232] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.503260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.503348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.503381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.503502] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.503528] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.503609] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.503635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.503729] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.503757] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.503838] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.503865] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.503945] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.503971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.504059] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.504086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.504163] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.504189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.504271] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.504298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.504384] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.504412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.504531] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.504559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.504637] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.504671] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.504752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.504779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.504893] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.504920] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.505012] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.505040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.505134] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.505161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.505278] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.505305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.505384] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.505411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.505493] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.505522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.505634] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.505675] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.505760] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.505786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.505875] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.505901] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.505995] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.506022] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.506142] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.506169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.506261] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.506287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.506371] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.506399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.506478] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.506504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.506612] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.506640] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.506773] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.506814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.506951] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.506992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.507115] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.507145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.507233] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.507260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.507345] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.507372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.507458] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.507485] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.507568] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.507595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.507700] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.834 [2024-12-06 17:54:28.507727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.834 qpair failed and we were unable to recover it. 00:38:46.834 [2024-12-06 17:54:28.507809] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.507835] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.507915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.507943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.508074] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.508101] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.508192] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.508221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.508304] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.508334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.508420] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.508447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.508554] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.508582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.508698] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.508727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.508807] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.508833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.508922] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.508948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.509034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.509061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.509140] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.509166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.509277] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.509305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.509422] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.509451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.509574] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.509614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.509710] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.509738] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.509844] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.509871] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.509953] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.509980] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.510093] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.510120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.510212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.510240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.510327] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.510355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.510440] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.510470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.510584] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.510612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.510723] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.510751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.510829] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.510856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.510937] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.510965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.511049] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.511076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.511156] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.511183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.511294] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.511322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.511402] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.511432] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.511519] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.511555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.511670] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.511704] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.511818] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.511845] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.511931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.511958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.512033] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.512060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.512165] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.512192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.512284] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.512312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.835 qpair failed and we were unable to recover it. 00:38:46.835 [2024-12-06 17:54:28.512431] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.835 [2024-12-06 17:54:28.512459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.512536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.512563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.512646] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.512687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.512788] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.512815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.512905] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.512932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.513016] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.513048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.513172] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.513201] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.513320] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.513348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.513436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.513464] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.513544] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.513571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.513653] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.513688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.513774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.513801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.513915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.513942] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.514051] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.514078] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.514192] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.514219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.514337] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.514366] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.514448] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.514477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.514561] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.514589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.514705] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.514732] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.514823] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.514849] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.514934] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.514961] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.515076] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.515104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.515183] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.515210] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.515321] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.515349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.515463] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.515491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.515577] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.515605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.515713] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.515753] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.515840] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.515867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.515995] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.516023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.516105] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.516132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.516245] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.516271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.516348] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.516377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.516497] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.516532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.516621] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.516652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.516747] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.516775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.516860] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.516887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.517008] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.517034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.517121] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.517148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.517234] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.517261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.517410] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.517437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.517526] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.836 [2024-12-06 17:54:28.517566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.836 qpair failed and we were unable to recover it. 00:38:46.836 [2024-12-06 17:54:28.517654] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.517689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.517807] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.517836] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.517924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.517951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.518030] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.518057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.518144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.518172] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.518288] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.518320] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.518405] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.518433] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.518550] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.518590] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.518681] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.518709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.518811] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.518842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.518965] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.518993] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.519080] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.519107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.519181] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.519207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.519321] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.519348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.519463] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.519489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.519606] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.519633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.519734] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.519762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.519850] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.519878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.519964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.519990] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.520079] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.520112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.520222] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.520253] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.520342] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.520370] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.520481] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.520508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.520586] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.520612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.520702] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.520732] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.520824] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.520866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.521008] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.521050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.521171] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.521200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.521322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.521350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.521441] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.521469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.521588] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.521615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.521710] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.521738] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.521829] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.521857] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.521941] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.521968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.522092] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.522119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.522212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.522240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.522336] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.522364] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.522478] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.837 [2024-12-06 17:54:28.522506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.837 qpair failed and we were unable to recover it. 00:38:46.837 [2024-12-06 17:54:28.522588] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.522616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.522713] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.522743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.522836] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.522863] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.522956] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.522996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.523122] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.523151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.523234] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.523261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.523347] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.523374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.523452] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.523478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.523558] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.523585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.523682] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.523716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.523831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.523858] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.523953] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.524000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.524120] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.524148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.524239] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.524268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.524421] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.524448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.524528] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.524556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.524640] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.524673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.524763] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.524790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.524866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.524893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.524975] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.525009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.525088] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.525121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.525214] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.525243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.525364] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.525393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.525487] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.525514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.525628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.525654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.525756] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.525783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.525865] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.525891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.525979] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.526006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.526119] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.526146] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.526232] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.526260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.526381] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.526410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.526528] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.526556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.526675] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.526703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.526788] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.526816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.526894] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.526921] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.527020] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.527048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.527138] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.527166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.527254] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.838 [2024-12-06 17:54:28.527283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.838 qpair failed and we were unable to recover it. 00:38:46.838 [2024-12-06 17:54:28.527402] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.527430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.527540] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.527567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.527648] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.527689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.527774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.527802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.527911] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.527938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.528027] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.528053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.528140] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.528169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.528322] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.528351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.528443] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.528482] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.528579] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.528607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.528693] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.528721] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.528832] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.528864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.528949] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.528976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.529062] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.529089] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.529169] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.529195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.529307] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.529336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.529421] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.529449] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.529539] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.529566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.529655] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.529689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.529800] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.529827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.529913] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.529940] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.530028] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.530055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.530139] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.530168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.530277] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.530304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.530392] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.530421] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.530508] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.530535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.530613] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.530640] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.530731] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.530760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.530846] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.530874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.531002] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.531042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.531159] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.531187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.531301] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.531328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.531412] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.531440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.531517] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.531545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.531628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.531656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.531757] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.531785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.531863] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.531891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.532023] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.532050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.532132] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.532169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.532308] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.532336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.532478] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.839 [2024-12-06 17:54:28.532505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.839 qpair failed and we were unable to recover it. 00:38:46.839 [2024-12-06 17:54:28.532589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.532616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.532752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.532780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.532874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.532901] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.533028] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.533055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.533139] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.533166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.533249] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.533286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.533376] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.533416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.533559] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.533588] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.533688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.533728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.533817] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.533846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.533937] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.533975] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.534095] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.534122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.534198] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.534224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.534321] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.534348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.534438] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.534478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.534570] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.534598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.534691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.534721] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.534814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.534841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.534932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.534968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.535083] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.535110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.535249] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.535285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.535380] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.535410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.535527] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.535555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.535633] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.535662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.535777] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.535804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.535885] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.535911] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.536004] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.536031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.536109] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.536135] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.536253] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.536290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.536398] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.536426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.536505] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.536531] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.536633] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.536672] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.536760] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.536787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.536877] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.536906] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.537036] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.537064] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.537147] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.537175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.537258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.537286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.537385] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.537421] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.537532] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.537560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.537689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.537718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.537833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.537860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.537945] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.537973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.840 qpair failed and we were unable to recover it. 00:38:46.840 [2024-12-06 17:54:28.538097] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.840 [2024-12-06 17:54:28.538124] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.538231] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.538258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.538368] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.538394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.538480] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.538508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.538592] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.538620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.538760] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.538789] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.538870] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.538898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.538981] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.539010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.539090] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.539118] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.539238] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.539266] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.539361] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.539389] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.539504] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.539532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.539615] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.539642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.539738] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.539765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.539863] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.539903] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.540022] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.540050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.540137] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.540164] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.540246] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.540273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.540351] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.540378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.540463] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.540491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.540572] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.540599] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.540712] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.540739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.540816] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.540848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.540943] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.540971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.541089] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.541116] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.541232] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.541258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.541339] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.541366] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.541449] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.541476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.541557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.541583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.541694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.541721] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.541806] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.541834] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.541915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.541942] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.542091] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.542118] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.542208] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.542237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.542320] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.542348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.542490] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.542517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.542668] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.542696] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.542772] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.542799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.542876] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.542904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.543038] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.543065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.841 qpair failed and we were unable to recover it. 00:38:46.841 [2024-12-06 17:54:28.543143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.841 [2024-12-06 17:54:28.543170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.543258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.543286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.543397] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.543424] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.543553] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.543580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.543680] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.543721] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.543828] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.543868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.544005] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.544045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.544129] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.544157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.544248] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.544275] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.544390] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.544417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.544500] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.544528] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.544612] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.544640] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.544755] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.544783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.544863] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.544891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.544992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.545019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.545100] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.545127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.545209] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.545236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.545355] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.545382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.545469] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.545496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.545581] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.545608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.545724] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.545752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.545834] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.545861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.545954] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.545987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.546081] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.546109] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.546190] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.546217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.546330] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.546357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.546449] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.546489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.546593] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.546633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.546752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.546781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.546866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.546893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.547015] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.547042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.547175] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.547223] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.547308] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.547336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.547432] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.547461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.547583] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.547610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.547693] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.547721] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.547810] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.842 [2024-12-06 17:54:28.547837] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.842 qpair failed and we were unable to recover it. 00:38:46.842 [2024-12-06 17:54:28.547918] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.547945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.548028] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.548055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.548168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.548195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.548272] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.548298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.548404] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.548444] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.548530] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.548558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.548696] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.548725] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.548811] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.548838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.548916] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.548942] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.549034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.549061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.549174] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.549202] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.549280] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.549307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.549426] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.549454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.549539] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.549566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.549677] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.549709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.549819] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.549846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.549926] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.549965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.550060] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.550086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.550203] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.550232] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.550311] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.550338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.550428] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.550457] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.550547] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.550574] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.550705] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.550734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.550817] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.550844] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.550929] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.550957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.551087] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.551120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.551218] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.551246] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.551333] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.551360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.551472] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.551499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.551612] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.551639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.551732] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.551761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.551847] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.551876] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.551988] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.552016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.552124] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.552152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.552235] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.552263] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.552353] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.552381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.552470] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.552498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.552610] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.552658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.552766] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.552794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.552890] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.552918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.553008] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.553035] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.553116] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.843 [2024-12-06 17:54:28.553142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.843 qpair failed and we were unable to recover it. 00:38:46.843 [2024-12-06 17:54:28.553232] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.553261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.553374] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.553403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.553486] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.553519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.553649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.553691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.553804] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.553832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.553931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.553974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.554077] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.554104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.554188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.554216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.554332] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.554359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.554456] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.554496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.554592] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.554626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.554722] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.554751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.554835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.554861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.554942] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.554980] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.555075] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.555102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.555211] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.555238] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.555347] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.555374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.555465] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.555505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.555630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.555681] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.555771] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.555799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.555882] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.555909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.556002] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.556031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.556121] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.556148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.556236] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.556264] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.556357] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.556384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.556465] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.556493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.556569] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.556597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.556727] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.556755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.556842] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.556868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.556950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.556984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.557069] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.557096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.557207] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.557234] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.557314] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.557341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.557431] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.557459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.557539] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.557566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.557652] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.557688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.557778] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.557805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.557902] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.557934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.844 qpair failed and we were unable to recover it. 00:38:46.844 [2024-12-06 17:54:28.558040] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.844 [2024-12-06 17:54:28.558067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.558151] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.558178] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.558288] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.558315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.558400] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.558429] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.558537] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.558564] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.558656] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.558698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.558788] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.558817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.558930] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.558969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.559052] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.559080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.559167] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.559195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.559283] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.559309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.559397] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.559425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.559505] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.559532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.559634] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.559672] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.559752] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.559778] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.559855] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.559882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.559995] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.560021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.560109] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.560137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.560241] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.560281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.560380] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.560409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.560499] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.560527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.560641] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.560685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.560775] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.560802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.560891] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.560918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.561002] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.561030] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.561160] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.561187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.561294] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.561322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.561404] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.561431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.561547] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.561579] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.561678] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.561707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.561824] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.561853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.561944] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.561985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.562078] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.562104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.562186] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.562213] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.562298] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.562326] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.562459] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.562487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.562607] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.562635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.562747] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.562775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.562865] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.562892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.563011] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.563043] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.563131] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.563159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.563241] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.845 [2024-12-06 17:54:28.563272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.845 qpair failed and we were unable to recover it. 00:38:46.845 [2024-12-06 17:54:28.563350] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.563377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.563456] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.563483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.563564] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.563591] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.563705] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.563732] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.563815] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.563842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.563928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.563955] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.564034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.564060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.564136] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.564163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.564245] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.564273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.564394] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.564425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.564536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.564565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.564670] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.564698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.564775] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.564802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.564931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.564963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.565051] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.565083] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.565167] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.565193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.565302] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.565329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.565436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.565463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.565541] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.565568] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.565683] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.565710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.565796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.565822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.565898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.565925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.566016] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.566042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.566150] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.566177] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.566251] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.566291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.566384] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.566411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.566502] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.566531] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.566643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.566687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.566778] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.566806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.566917] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.566944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.567025] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.567052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.567141] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.567168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.567273] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.567301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.567415] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.567443] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.567528] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.567555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.567680] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.567707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.567786] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.567813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.846 qpair failed and we were unable to recover it. 00:38:46.846 [2024-12-06 17:54:28.567893] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.846 [2024-12-06 17:54:28.567920] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.568012] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.568042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.568144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.568185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.568274] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.568303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.568431] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.568458] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.568572] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.568598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.568691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.568719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.568831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.568861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.568973] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.569000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.569089] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.569116] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.569198] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.569226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.569306] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.569344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.569432] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.569459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.569565] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.569592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.569712] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.569746] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.569837] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.569864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.569945] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.569976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.570092] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.570120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.570205] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.570233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.570357] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.570386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.570483] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.570510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.570594] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.570621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.570722] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.570749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.570833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.570860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.570937] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.570963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.571045] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.571073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.571152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.571179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.571301] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.571327] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.571452] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.571480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.571569] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.571598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.571692] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.571720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.571833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.571860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.571942] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.571975] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.572060] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.572088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.572170] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.572198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.572285] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.572313] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.572397] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.572424] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.572534] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.572561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.572649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.572686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.572775] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.572802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.572899] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.847 [2024-12-06 17:54:28.572926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.847 qpair failed and we were unable to recover it. 00:38:46.847 [2024-12-06 17:54:28.573025] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.573057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.573146] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.573174] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.573293] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.573321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.573406] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.573435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.573555] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.573582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.573658] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.573691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.573774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.573801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.573919] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.573945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.574029] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.574057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.574143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.574172] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.574293] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.574331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.574420] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.574448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.574536] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.574563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.574643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.574682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.574774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.574801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.574916] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.574943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.575028] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.575056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.575141] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.575168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.575246] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.575273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.575356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.575382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.575470] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.575503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.575615] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.575643] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.575746] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.575775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.575871] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.575899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.575981] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.576009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.576091] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.576117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.576202] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.576230] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.576323] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.576353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.576468] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.576494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.576571] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.576598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.576689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.576716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.576810] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.576838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.576925] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.576951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.577070] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.577097] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.577184] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.577211] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.577299] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.577327] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.577415] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.577442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.577526] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.577554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.577647] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.577690] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.577785] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.577817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.577900] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.577926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.578017] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.848 [2024-12-06 17:54:28.578047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.848 qpair failed and we were unable to recover it. 00:38:46.848 [2024-12-06 17:54:28.578157] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.849 [2024-12-06 17:54:28.578183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.849 qpair failed and we were unable to recover it. 00:38:46.849 [2024-12-06 17:54:28.578270] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.849 [2024-12-06 17:54:28.578297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.849 qpair failed and we were unable to recover it. 00:38:46.849 [2024-12-06 17:54:28.578373] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.849 [2024-12-06 17:54:28.578400] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.849 qpair failed and we were unable to recover it. 00:38:46.849 [2024-12-06 17:54:28.578488] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.849 [2024-12-06 17:54:28.578515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:46.849 qpair failed and we were unable to recover it. 00:38:46.849 [2024-12-06 17:54:28.578602] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.849 [2024-12-06 17:54:28.578630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:46.849 qpair failed and we were unable to recover it. 00:38:46.849 [2024-12-06 17:54:28.578780] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.849 [2024-12-06 17:54:28.578808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.849 qpair failed and we were unable to recover it. 00:38:46.849 [2024-12-06 17:54:28.578890] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.849 [2024-12-06 17:54:28.578916] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.849 qpair failed and we were unable to recover it. 00:38:46.849 [2024-12-06 17:54:28.578998] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.849 [2024-12-06 17:54:28.579025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.849 qpair failed and we were unable to recover it. 00:38:46.849 [2024-12-06 17:54:28.579137] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.849 [2024-12-06 17:54:28.579166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:46.849 qpair failed and we were unable to recover it. 00:38:46.849 [2024-12-06 17:54:28.579258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:46.849 [2024-12-06 17:54:28.579284] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.579380] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.579407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.579493] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.579521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.579636] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.579676] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.579769] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.579796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.579886] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.579913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.580006] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.580033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.580136] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.580165] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.580249] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.580277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.580380] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.580407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.580496] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.580523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.580637] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.580678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.580762] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.580787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.580868] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.580908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.580993] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.581018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.581100] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.581127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.581203] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.581228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.581313] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.581339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.581449] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.581476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.581564] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.581593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.581677] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.581703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.581786] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.581812] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.581898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.581926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.582003] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.582030] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.582156] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.582183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.582295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.582323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.582406] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.582433] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.582548] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.582576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.582674] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.582702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.582797] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.582824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.582917] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.582967] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.583059] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.583086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.583219] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.583251] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.583331] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.583359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.583471] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.133 [2024-12-06 17:54:28.583498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.133 qpair failed and we were unable to recover it. 00:38:47.133 [2024-12-06 17:54:28.583581] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.583607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.583698] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.583728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.583820] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.583847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.583924] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.583962] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.584045] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.584072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.584158] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.584185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.584299] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.584326] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.584438] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.584468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.584549] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.584580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.584679] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.584707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.584793] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.584822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.584935] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.584963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.585048] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.585074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.585189] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.585227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.585311] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.585338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.585425] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.585451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.585548] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.585575] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.585675] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.585705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.585798] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.585825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.585907] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.585934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.586041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.586068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.586155] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.586182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.586267] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.586294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.586397] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.586436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.586543] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.586572] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.586676] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.586704] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.586814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.586841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.586928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.586962] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.587046] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.587072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.587177] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.587206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.587296] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.587328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.587419] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.587445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.587526] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.587553] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.587639] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.587679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.587773] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.587799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.587882] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.587918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.588051] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.588078] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.588161] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.134 [2024-12-06 17:54:28.588187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.134 qpair failed and we were unable to recover it. 00:38:47.134 [2024-12-06 17:54:28.588301] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.588329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.588420] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.588446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.588537] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.588564] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.588649] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.588688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.588769] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.588796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.588885] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.588913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.589004] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.589032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.589114] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.589141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.589255] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.589284] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.589361] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.589390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.589494] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.589535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.589648] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.589693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.589793] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.589822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.589911] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.589938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.590018] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.590055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.590152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.590180] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.590299] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.590328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.590448] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.590479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.590593] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.590621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.590726] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.590755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.590849] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.590876] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.590969] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.590998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.591083] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.591110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.591188] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.591215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.591304] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.591341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.591434] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.591461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.591588] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.591616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.591709] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.591737] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.591823] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.591850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.591945] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.591971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.592057] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.592083] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.592170] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.592197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.592275] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.592304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.592389] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.592419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.592502] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.592530] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.592623] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.592651] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.592749] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.592778] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.592854] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.592885] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.135 qpair failed and we were unable to recover it. 00:38:47.135 [2024-12-06 17:54:28.593008] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.135 [2024-12-06 17:54:28.593034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.593107] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.593134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.593220] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.593247] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.593334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.593363] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.593479] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.593507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.593593] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.593621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.593722] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.593750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.593840] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.593866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.593946] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.593973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.594067] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.594094] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.594178] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.594205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.594318] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.594345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.594436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.594463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.594554] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.594595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.594701] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.594735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.594826] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.594854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.594931] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.594958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.595047] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.595075] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.595186] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.595217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.595299] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.595325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.595407] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.595439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.595525] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.595551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.595627] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.595655] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.595750] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.595791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.595879] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.595909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.596043] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.596071] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.596220] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.596248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.596333] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.596361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.596444] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.596471] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.596560] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.596587] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.596684] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.596711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.596835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.596863] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.596949] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.596987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 [2024-12-06 17:54:28.597097] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.597125] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:38:47.136 [2024-12-06 17:54:28.597216] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.597246] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@868 -- # return 0 00:38:47.136 [2024-12-06 17:54:28.597336] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.597363] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:38:47.136 [2024-12-06 17:54:28.597477] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.136 [2024-12-06 17:54:28.597504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.136 qpair failed and we were unable to recover it. 00:38:47.136 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@732 -- # xtrace_disable 00:38:47.136 [2024-12-06 17:54:28.597617] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.597644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:38:47.137 [2024-12-06 17:54:28.597734] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.597764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.597846] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.597872] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.597951] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.597978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.598056] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.598083] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.598201] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.598229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.598310] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.598337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.598418] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.598447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.598557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.598583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.598671] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.598700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.598791] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.598818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.598936] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.598969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.599065] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.599092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.599205] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.599233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.599318] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.599353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.599433] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.599461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.599545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.599573] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.599689] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.599717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.599797] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.599830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.599911] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.599939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.600033] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.600060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.600139] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.600166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.600261] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.600289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.600387] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.600414] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.600501] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.600527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.600648] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.600683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.600808] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.600834] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.600915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.600942] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.601041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.601070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.601149] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.601175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.601284] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.137 [2024-12-06 17:54:28.601311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.137 qpair failed and we were unable to recover it. 00:38:47.137 [2024-12-06 17:54:28.601421] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.601448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.601535] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.601562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.601653] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.601689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.601773] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.601801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.601915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.601942] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.602056] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.602089] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.602227] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.602254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.602367] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.602394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.602504] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.602532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.602628] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.602657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.602745] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.602777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.602855] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.602883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.602958] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.602985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.603069] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.603106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.603191] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.603217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.603305] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.603331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.603442] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.603469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.603550] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.603577] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.603669] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.603697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.603798] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.603825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.603909] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.603936] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.604049] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.604076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.604162] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.604189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.604270] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.604299] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.604386] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.604415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.604500] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.604527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.604609] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.604640] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.604766] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.604793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.604880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.604907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.604999] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.605026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.605116] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.605142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.605250] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.605277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.605368] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.605394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.605470] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.605496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.605612] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.605639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.605744] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.605771] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.605849] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.605876] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.605964] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.605999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.138 [2024-12-06 17:54:28.606094] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.138 [2024-12-06 17:54:28.606121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.138 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.606237] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.606265] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.606352] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.606379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.606463] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.606491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.606567] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.606604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.606694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.606720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.606803] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.606830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.606926] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.606954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.607034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.607061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.607140] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.607166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.607245] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.607272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.607356] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.607382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.607458] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.607486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.607598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.607626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.607718] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.607746] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.607840] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.607866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.607948] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.607985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.608063] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.608089] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.608198] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.608225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.608310] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.608337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.608417] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.608443] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.608531] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.608559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.608654] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.608707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.608801] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.608830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.608911] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.608939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.609057] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.609084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.609178] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.609212] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.609305] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.609333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.609415] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.609443] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.609520] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.609556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.609675] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.609703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.609793] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.609820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.609911] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.609938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.610036] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.610063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.610145] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.610172] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.610255] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.610289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.610373] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.610401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.610513] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.610541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.610638] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.139 [2024-12-06 17:54:28.610692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe50000b90 with addr=10.0.0.2, port=4420 00:38:47.139 qpair failed and we were unable to recover it. 00:38:47.139 [2024-12-06 17:54:28.610824] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.610864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.610957] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.610994] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.611091] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.611118] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.611207] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.611234] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.611372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.611398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.611486] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.611512] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.611618] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.611645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.611751] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.611781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.611863] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.611891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.612001] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.612028] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.612105] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.612139] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.612257] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.612285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.612372] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.612399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.612476] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.612503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.612593] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.612623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.612756] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.612783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.612866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.612893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.613008] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.613036] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.613177] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.613204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.613294] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.613322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.613438] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.613472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.613547] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.613579] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.613658] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.613693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.613776] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.613804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.613880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.613907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.614052] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.614079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.614170] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.614199] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.614281] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.614309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.614399] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.614427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.614514] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.614542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.614660] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.614707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.614814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.614842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.614926] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.614953] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.615049] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.615088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.615168] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.615195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 [2024-12-06 17:54:28.615321] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.615348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:38:47.140 [2024-12-06 17:54:28.615459] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.140 [2024-12-06 17:54:28.615486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.140 qpair failed and we were unable to recover it. 00:38:47.140 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:38:47.140 [2024-12-06 17:54:28.615576] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.615604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:47.141 [2024-12-06 17:54:28.615701] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.615729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:38:47.141 [2024-12-06 17:54:28.615808] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.615841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.615920] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.615946] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.616029] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.616064] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.616152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.616179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.616269] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.616298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.616417] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.616447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.616535] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.616562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.616688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.616716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.616804] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.616830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.616911] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.616937] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.617020] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.617048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.617146] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.617174] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.617252] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.617280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.617362] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.617395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.617474] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.617501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.617605] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.617632] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.617732] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.617759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.617874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.617901] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.617988] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.618015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.618119] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.618146] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.618230] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.618258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.618341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.618368] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.618487] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.618515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.618594] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.618622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.618715] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.618743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.618825] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.618852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.618936] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.618964] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.619077] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.619103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.619187] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.619216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.619310] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.619339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.619413] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.619439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.619527] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.619554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.619633] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.619679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.619770] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.619799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.619918] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.619944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.141 [2024-12-06 17:54:28.620053] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.141 [2024-12-06 17:54:28.620081] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.141 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.620161] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.620188] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.620301] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.620328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.620415] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.620441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.620521] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.620547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.620632] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.620678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.620787] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.620814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.620898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.620924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.621011] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.621039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.621133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.621161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.621242] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.621269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.621357] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.621384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.621459] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.621485] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.621598] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.621625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.621747] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.621775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.621859] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.621886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.621989] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.622016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.622092] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.622118] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.622198] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.622226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.622316] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.622344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.622465] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.622493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.622578] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.622604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.622726] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.622756] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.622845] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.622872] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.622947] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.622973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.623060] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.623086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.623198] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.623225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.623302] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.623330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.623409] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.623436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.623522] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.623550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.623640] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.623681] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.623796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.623824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.623914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.623942] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.624033] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.624060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.624135] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.624162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.142 qpair failed and we were unable to recover it. 00:38:47.142 [2024-12-06 17:54:28.624246] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.142 [2024-12-06 17:54:28.624272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.624355] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.624383] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.624490] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.624518] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.624607] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.624635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.624738] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.624765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.624874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.624901] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.624996] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.625024] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.625115] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.625142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.625225] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.625252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.625326] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.625353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.625436] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.625462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.625548] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.625575] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.625678] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.625706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.625794] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.625820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.625905] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.625931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.626018] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.626049] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.626133] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.626161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.626276] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.626304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.626390] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.626418] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.626525] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.626552] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.626633] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.626679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.626767] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.626795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.626875] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.626903] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.627010] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.627037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.627152] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.627180] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.627291] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.627318] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.627429] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.627456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.627534] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.627561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.627688] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.627716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.627798] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.627826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.627907] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.627933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.628033] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.628061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.628171] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.628198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.628280] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.628307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.628399] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.628425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.628503] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.628529] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.628613] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.628639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.628732] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.628759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.143 [2024-12-06 17:54:28.628869] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.143 [2024-12-06 17:54:28.628896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.143 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.628988] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.629016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.629105] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.629133] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.629226] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.629252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.629332] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.629358] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.629437] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.629463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.629549] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.629576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.629687] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.629715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.629796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.629823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.629941] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.629979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.630060] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.630086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.630177] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.630204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.630296] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.630322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.630412] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.630438] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.630518] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.630544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.630620] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.630647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.630748] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.630788] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.630871] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.630902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.630996] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.631023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.631131] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.631158] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.631236] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.631264] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.631346] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.631372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.631452] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.631480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.631566] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.631592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.631713] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.631740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.631825] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.631852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.631970] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.632000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.632086] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.632112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.632196] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.632223] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.632302] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.632329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.632411] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.632438] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.632518] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.632545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.632656] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.632688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.632796] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.632823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.632900] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.632926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.633040] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.633066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.633143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.633170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.633246] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.633274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.633378] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.144 [2024-12-06 17:54:28.633418] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.144 qpair failed and we were unable to recover it. 00:38:47.144 [2024-12-06 17:54:28.633548] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.633576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.633677] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.633706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.633791] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.633817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.633895] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.633921] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.634043] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.634069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.634146] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.634172] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.634260] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.634287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.634379] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.634406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.634516] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.634542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.634630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.634674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.634760] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.634786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.634866] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.634892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.634971] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.635003] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.635084] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.635112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.635212] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.635252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.635378] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.635410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.635525] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.635555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.635636] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.635682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.635799] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.635826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.635941] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.635976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.636060] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.636089] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.636178] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.636205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.636291] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.636319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.636411] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.636437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.636550] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.636577] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.636692] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.636720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.636812] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.636840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.636915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.636942] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.637024] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.637051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.637144] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.637172] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.637256] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.637284] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.637365] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.637393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.637509] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.637536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.637623] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.637662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.637757] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.637785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.637891] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.637918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.638010] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.638037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.638119] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.638146] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.145 qpair failed and we were unable to recover it. 00:38:47.145 [2024-12-06 17:54:28.638258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.145 [2024-12-06 17:54:28.638285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.638394] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.638421] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.638533] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.638559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.638635] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.638680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.638795] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.638822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.638914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.638940] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.639025] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.639060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.639181] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.639208] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.639301] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.639329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.639473] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.639502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.639610] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.639637] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.639732] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.639760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.639856] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.639884] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.639976] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.640002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.640116] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.640142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.640239] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.640265] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.640340] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.640367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.640454] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.640481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.640574] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.640600] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.640683] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.640709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.640794] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.640820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.640901] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.640929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.641053] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.641080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.641199] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.641226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.641318] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.641345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.641442] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.641469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.641567] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.641608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.641705] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.641733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.641814] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.641840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.641959] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.641993] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.642077] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.642108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.642185] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.642211] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.642300] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.642327] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.642420] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.642447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.642521] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.146 [2024-12-06 17:54:28.642547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.146 qpair failed and we were unable to recover it. 00:38:47.146 [2024-12-06 17:54:28.642675] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.642702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.642780] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.642807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.642897] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.642923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.643015] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.643044] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.643132] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.643160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.643241] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.643269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.643349] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.643377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.643503] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.643544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.643633] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.643679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.643805] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.643832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.643915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.643941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.644023] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.644050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.644129] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.644156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.644233] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.644259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.644337] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.644364] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.644497] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.644524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.644603] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.644630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.644729] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.644757] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.644847] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.644874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.644998] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.645025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.645107] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.645134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.645249] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.645279] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.645369] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.645398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.645491] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.645518] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.645602] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.645629] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.645767] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.645795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.645876] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.645903] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.645991] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.646017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.646096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.646122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.646227] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.646254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.646343] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.646370] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.646478] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.646505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.646588] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.646614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.646697] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.646741] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.646849] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.646876] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.646962] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.646988] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.647096] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.647122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.647232] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.147 [2024-12-06 17:54:28.647261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.147 qpair failed and we were unable to recover it. 00:38:47.147 [2024-12-06 17:54:28.647376] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.647414] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.647500] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.647527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.647638] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.647678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.647774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.647801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.647880] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.647907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.648004] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.648040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.648129] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.648157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.648242] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.648272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.648389] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.648418] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.648530] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.648556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.648642] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.648681] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.648797] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.648828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.648915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.648941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.649068] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.649096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.649202] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.649229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.649317] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.649345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.649462] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.649489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.649588] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.649616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.649723] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.649750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.649831] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.649859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.649936] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.649967] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.650075] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.650102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.650179] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.650206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.650323] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.650350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.650437] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.650464] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.650570] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.650610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.650743] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.650771] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.650863] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.650890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.650984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.651011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.651091] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.651117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.651232] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.651258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.651339] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.651367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.651453] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.651481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.651579] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.651606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.651721] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.651748] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.651854] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.651880] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.651987] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.652023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.652103] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.148 [2024-12-06 17:54:28.652130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.148 qpair failed and we were unable to recover it. 00:38:47.148 [2024-12-06 17:54:28.652211] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.652240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.652329] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.652357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.652444] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.652472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.652590] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.652617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.652724] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.652751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.652841] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.652867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.652947] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.652974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.653061] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.653088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.653179] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.653207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.653328] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.653356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.653442] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.653468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.653576] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.653602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.653727] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.653754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.653835] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.653862] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.653952] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.653982] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.654074] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.654102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.654189] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.654215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.654299] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.654326] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.654413] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.654439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.654524] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.654551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.654634] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.654662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.654766] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.654793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.654874] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.654900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.654984] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.655010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.655089] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.655116] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.655200] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.655228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.655305] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.655331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.655423] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.655463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.655556] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.655585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.655698] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.655726] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.655812] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.655838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.655928] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.655954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.656085] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.656111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.656206] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.656232] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.656312] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.656338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 Malloc0 00:38:47.149 [2024-12-06 17:54:28.656459] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.656488] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.656604] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.656632] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 [2024-12-06 17:54:28.656754] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.149 [2024-12-06 17:54:28.656781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.149 qpair failed and we were unable to recover it. 00:38:47.149 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:47.149 [2024-12-06 17:54:28.656870] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.656896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@21 -- # rpc_cmd nvmf_create_transport -t tcp -o 00:38:47.150 [2024-12-06 17:54:28.656976] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.657002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.657088] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:47.150 [2024-12-06 17:54:28.657115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.657195] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:38:47.150 [2024-12-06 17:54:28.657224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.657341] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.657367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.657467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.657493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.657577] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.657604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.657694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.657722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.657810] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.657838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.657927] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.657954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.658036] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.658063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.658151] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.658178] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.658300] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.658334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.658427] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.658455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.658545] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.658575] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.658685] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.658712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.658802] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.658831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.658915] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.658943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.659034] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.659061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.659141] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.659168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.659302] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.659329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.659435] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.659462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.659573] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.659601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.659695] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.659722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.659803] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.659829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.659904] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.659930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.659997] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:38:47.150 [2024-12-06 17:54:28.660020] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.660046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.660164] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.660195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.660295] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.660322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.660410] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.660437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.660512] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.660539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.660618] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.660645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.660741] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.660767] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.660852] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.660878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.660998] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.661025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.661101] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.661128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.150 [2024-12-06 17:54:28.661216] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.150 [2024-12-06 17:54:28.661245] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.150 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.661335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.661363] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.661449] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.661476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.661559] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.661586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.661680] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.661707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.661800] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.661827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.661914] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.661941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.662132] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.662159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.662274] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.662300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.662377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.662403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.662523] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.662549] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.662637] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.662669] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.662748] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.662776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.662863] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.662889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.662966] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.662993] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.663101] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.663127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.663214] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.663240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.663357] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.663384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.663481] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.663513] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.663622] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.663659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.663760] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.663787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.663870] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.663896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.663981] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.664008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.664089] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.664115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.664205] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.664232] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.664313] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.664339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.664422] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.664448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.664531] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.664557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.664637] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.664675] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.664766] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.664793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.664904] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.664931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.665004] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.665031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.665118] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.665145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.665225] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.665252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.665379] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.665419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.151 qpair failed and we were unable to recover it. 00:38:47.151 [2024-12-06 17:54:28.665546] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.151 [2024-12-06 17:54:28.665576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.665662] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.665698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.665784] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.665811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.665891] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.665918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.666029] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.666056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.666254] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.666282] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.666375] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.666402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.666478] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.666504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.666694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.666722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.666834] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.666860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.666947] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.666978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.667054] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.667081] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.667155] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.667181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.667260] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.667289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.667380] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.667406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.667516] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.667543] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.667625] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.667652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.667742] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.667768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.667884] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.667911] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.668007] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.668033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.668143] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:47.152 [2024-12-06 17:54:28.668169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.668248] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.668274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:38:47.152 [2024-12-06 17:54:28.668381] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.668408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:47.152 [2024-12-06 17:54:28.668505] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.668534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:38:47.152 [2024-12-06 17:54:28.668617] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.668644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.668736] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.668764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.668850] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.668877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.668963] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.668991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.669088] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.669115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.669223] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.669250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.669334] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.669361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.669464] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.669504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.669597] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.669625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.669838] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.669868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.669955] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.669982] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.670061] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.670092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.152 [2024-12-06 17:54:28.670231] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.152 [2024-12-06 17:54:28.670258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.152 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.670346] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.670372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.670462] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.670502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.670707] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.670736] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.670823] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.670849] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.670932] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.670959] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.671039] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.671066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.671151] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.671177] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.671265] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.671292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.671378] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.671404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.671542] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.671568] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.671643] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.671684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.671770] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.671796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.671911] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.671937] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.672033] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.672062] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.672173] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.672200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.672288] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.672314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.672405] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.672432] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.672521] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.672548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.672630] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.672675] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.672770] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.672808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.672886] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.672913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.672996] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.673023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.673139] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.673166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.673248] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.673275] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.673357] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.673384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.673533] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.673561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.673646] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.673681] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.673765] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.673792] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.673873] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.673899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.673977] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.674003] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.674098] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.674126] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.674206] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.674232] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.674311] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.674337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.674449] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.674476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.674589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.674617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.674719] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.674747] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.674826] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.674854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.153 [2024-12-06 17:54:28.674938] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.153 [2024-12-06 17:54:28.674966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.153 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.675076] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.675103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.675189] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.675216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.675340] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.675367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.675450] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.675477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.675560] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.675589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.675690] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.675717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.675802] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.675828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.675950] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.675979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.676054] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.676080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.676165] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.676192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.154 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.676277] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.676303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:38:47.154 [2024-12-06 17:54:28.676412] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.676438] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:47.154 [2024-12-06 17:54:28.676546] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.676577] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:38:47.154 [2024-12-06 17:54:28.676656] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.676688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.676768] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.676795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.676877] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.676906] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.676983] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.677010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.677126] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.677152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.677228] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.677254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.677335] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.677361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.677467] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.677493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.677577] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.677603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.677710] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.677743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.677829] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.677858] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.677938] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.677966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.678054] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.678087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.678193] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.678220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.678332] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.678360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.678446] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.678473] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.678556] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.678583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.678659] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.678691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.678783] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.678811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.678916] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.678942] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.679025] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.679052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.679165] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.679194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.679315] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.154 [2024-12-06 17:54:28.679342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.154 qpair failed and we were unable to recover it. 00:38:47.154 [2024-12-06 17:54:28.679454] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.679480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.679563] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.679589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.679691] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.679719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.679820] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.679848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.679992] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.680020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.680103] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.680130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.680220] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.680247] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.680325] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.680352] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.680440] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.680467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.680557] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.680584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.680674] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.680703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.680787] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.680814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.680899] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.680926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.681008] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.681036] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.681119] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.681145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.681261] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.681288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.681409] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.681437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.681524] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.681551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.681635] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.681662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.681784] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.681811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.681898] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.681925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.682041] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.682069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.682184] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.682212] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.682296] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.682323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.682430] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.682456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.682533] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.682559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.682654] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.682689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.682774] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.682800] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.682891] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.682920] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.682996] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.683028] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.683117] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.683145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.683258] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.683286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.683369] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.683396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.683505] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.683532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.155 qpair failed and we were unable to recover it. 00:38:47.155 [2024-12-06 17:54:28.683619] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.155 [2024-12-06 17:54:28.683646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.683737] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.683764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.683876] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.683904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.684024] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.684052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.684136] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.684163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.684251] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:47.156 [2024-12-06 17:54:28.684277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.684369] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.684397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:38:47.156 [2024-12-06 17:54:28.684481] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.684508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:47.156 [2024-12-06 17:54:28.684589] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.684615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:38:47.156 [2024-12-06 17:54:28.684694] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.684721] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.684830] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.684857] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.684962] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.684991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.685083] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.685110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.685220] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.685247] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.685327] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.685353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.685441] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.685468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.685575] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.685615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.685723] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.685752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.685833] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.685860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.685953] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.685981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.686064] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.686097] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.686171] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.686198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.686276] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.686302] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.686377] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.686404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.686518] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.686547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.686629] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.686657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.686756] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.686783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.686864] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.686891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.686974] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.687000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.687111] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.687137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe58000b90 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.687246] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.687274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x775730 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.687357] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.687385] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.687465] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.687492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.687573] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.687600] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.687701] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.687730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.687845] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.687872] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.156 qpair failed and we were unable to recover it. 00:38:47.156 [2024-12-06 17:54:28.687991] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.156 [2024-12-06 17:54:28.688018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.157 qpair failed and we were unable to recover it. 00:38:47.157 [2024-12-06 17:54:28.688103] posix.c:1054:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:38:47.157 [2024-12-06 17:54:28.688130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fbe4c000b90 with addr=10.0.0.2, port=4420 00:38:47.157 qpair failed and we were unable to recover it. 00:38:47.157 [2024-12-06 17:54:28.688493] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:38:47.157 [2024-12-06 17:54:28.690954] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.157 [2024-12-06 17:54:28.691085] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.157 [2024-12-06 17:54:28.691116] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.157 [2024-12-06 17:54:28.691133] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.157 [2024-12-06 17:54:28.691146] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.157 [2024-12-06 17:54:28.691182] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.157 qpair failed and we were unable to recover it. 00:38:47.157 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:47.157 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@26 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:38:47.157 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:47.157 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:38:47.157 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:47.157 17:54:28 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@50 -- # wait 414754 00:38:47.157 [2024-12-06 17:54:28.700688] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.157 [2024-12-06 17:54:28.700794] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.157 [2024-12-06 17:54:28.700823] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.157 [2024-12-06 17:54:28.700838] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.157 [2024-12-06 17:54:28.700851] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.157 [2024-12-06 17:54:28.700883] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.157 qpair failed and we were unable to recover it. 00:38:47.157 [2024-12-06 17:54:28.710758] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.157 [2024-12-06 17:54:28.710875] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.157 [2024-12-06 17:54:28.710907] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.157 [2024-12-06 17:54:28.710924] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.157 [2024-12-06 17:54:28.710938] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.157 [2024-12-06 17:54:28.710970] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.157 qpair failed and we were unable to recover it. 00:38:47.157 [2024-12-06 17:54:28.720738] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.157 [2024-12-06 17:54:28.720853] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.157 [2024-12-06 17:54:28.720882] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.157 [2024-12-06 17:54:28.720898] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.157 [2024-12-06 17:54:28.720910] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.157 [2024-12-06 17:54:28.720942] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.157 qpair failed and we were unable to recover it. 00:38:47.157 [2024-12-06 17:54:28.730610] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.157 [2024-12-06 17:54:28.730705] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.157 [2024-12-06 17:54:28.730733] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.157 [2024-12-06 17:54:28.730748] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.157 [2024-12-06 17:54:28.730761] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.157 [2024-12-06 17:54:28.730793] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.157 qpair failed and we were unable to recover it. 00:38:47.157 [2024-12-06 17:54:28.740645] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.157 [2024-12-06 17:54:28.740745] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.157 [2024-12-06 17:54:28.740772] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.157 [2024-12-06 17:54:28.740787] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.157 [2024-12-06 17:54:28.740800] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.157 [2024-12-06 17:54:28.740831] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.157 qpair failed and we were unable to recover it. 00:38:47.157 [2024-12-06 17:54:28.750691] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.157 [2024-12-06 17:54:28.750804] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.157 [2024-12-06 17:54:28.750837] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.157 [2024-12-06 17:54:28.750854] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.157 [2024-12-06 17:54:28.750867] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.157 [2024-12-06 17:54:28.750898] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.157 qpair failed and we were unable to recover it. 00:38:47.157 [2024-12-06 17:54:28.760711] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.157 [2024-12-06 17:54:28.760813] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.157 [2024-12-06 17:54:28.760841] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.157 [2024-12-06 17:54:28.760857] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.157 [2024-12-06 17:54:28.760870] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.157 [2024-12-06 17:54:28.760901] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.157 qpair failed and we were unable to recover it. 00:38:47.157 [2024-12-06 17:54:28.770798] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.157 [2024-12-06 17:54:28.770898] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.157 [2024-12-06 17:54:28.770925] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.157 [2024-12-06 17:54:28.770940] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.157 [2024-12-06 17:54:28.770954] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.157 [2024-12-06 17:54:28.770985] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.157 qpair failed and we were unable to recover it. 00:38:47.157 [2024-12-06 17:54:28.780787] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.157 [2024-12-06 17:54:28.780878] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.157 [2024-12-06 17:54:28.780904] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.157 [2024-12-06 17:54:28.780919] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.157 [2024-12-06 17:54:28.780932] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.157 [2024-12-06 17:54:28.780964] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.157 qpair failed and we were unable to recover it. 00:38:47.157 [2024-12-06 17:54:28.790800] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.157 [2024-12-06 17:54:28.790888] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.157 [2024-12-06 17:54:28.790913] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.157 [2024-12-06 17:54:28.790928] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.157 [2024-12-06 17:54:28.790946] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.157 [2024-12-06 17:54:28.790978] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.157 qpair failed and we were unable to recover it. 00:38:47.157 [2024-12-06 17:54:28.800806] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.157 [2024-12-06 17:54:28.800899] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.157 [2024-12-06 17:54:28.800924] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.157 [2024-12-06 17:54:28.800938] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.157 [2024-12-06 17:54:28.800951] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.158 [2024-12-06 17:54:28.800982] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.158 qpair failed and we were unable to recover it. 00:38:47.158 [2024-12-06 17:54:28.810831] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.158 [2024-12-06 17:54:28.810920] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.158 [2024-12-06 17:54:28.810952] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.158 [2024-12-06 17:54:28.810977] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.158 [2024-12-06 17:54:28.810998] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.158 [2024-12-06 17:54:28.811041] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.158 qpair failed and we were unable to recover it. 00:38:47.158 [2024-12-06 17:54:28.821002] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.158 [2024-12-06 17:54:28.821091] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.158 [2024-12-06 17:54:28.821117] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.158 [2024-12-06 17:54:28.821131] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.158 [2024-12-06 17:54:28.821144] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.158 [2024-12-06 17:54:28.821175] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.158 qpair failed and we were unable to recover it. 00:38:47.158 [2024-12-06 17:54:28.830892] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.158 [2024-12-06 17:54:28.830981] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.158 [2024-12-06 17:54:28.831007] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.158 [2024-12-06 17:54:28.831021] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.158 [2024-12-06 17:54:28.831035] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.158 [2024-12-06 17:54:28.831066] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.158 qpair failed and we were unable to recover it. 00:38:47.158 [2024-12-06 17:54:28.841024] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.158 [2024-12-06 17:54:28.841125] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.158 [2024-12-06 17:54:28.841151] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.158 [2024-12-06 17:54:28.841166] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.158 [2024-12-06 17:54:28.841180] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.158 [2024-12-06 17:54:28.841211] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.158 qpair failed and we were unable to recover it. 00:38:47.158 [2024-12-06 17:54:28.850948] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.158 [2024-12-06 17:54:28.851036] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.158 [2024-12-06 17:54:28.851060] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.158 [2024-12-06 17:54:28.851075] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.158 [2024-12-06 17:54:28.851087] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.158 [2024-12-06 17:54:28.851118] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.158 qpair failed and we were unable to recover it. 00:38:47.158 [2024-12-06 17:54:28.861070] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.158 [2024-12-06 17:54:28.861153] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.158 [2024-12-06 17:54:28.861178] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.158 [2024-12-06 17:54:28.861192] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.158 [2024-12-06 17:54:28.861205] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.158 [2024-12-06 17:54:28.861250] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.158 qpair failed and we were unable to recover it. 00:38:47.158 [2024-12-06 17:54:28.870992] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.158 [2024-12-06 17:54:28.871077] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.158 [2024-12-06 17:54:28.871101] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.158 [2024-12-06 17:54:28.871116] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.158 [2024-12-06 17:54:28.871129] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.158 [2024-12-06 17:54:28.871160] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.158 qpair failed and we were unable to recover it. 00:38:47.158 [2024-12-06 17:54:28.881050] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.158 [2024-12-06 17:54:28.881149] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.158 [2024-12-06 17:54:28.881173] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.158 [2024-12-06 17:54:28.881188] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.158 [2024-12-06 17:54:28.881202] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.158 [2024-12-06 17:54:28.881233] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.158 qpair failed and we were unable to recover it. 00:38:47.158 [2024-12-06 17:54:28.891078] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.158 [2024-12-06 17:54:28.891176] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.158 [2024-12-06 17:54:28.891206] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.158 [2024-12-06 17:54:28.891222] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.158 [2024-12-06 17:54:28.891235] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.158 [2024-12-06 17:54:28.891267] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.158 qpair failed and we were unable to recover it. 00:38:47.158 [2024-12-06 17:54:28.901066] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.158 [2024-12-06 17:54:28.901168] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.158 [2024-12-06 17:54:28.901196] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.158 [2024-12-06 17:54:28.901211] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.158 [2024-12-06 17:54:28.901224] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.158 [2024-12-06 17:54:28.901255] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.158 qpair failed and we were unable to recover it. 00:38:47.158 [2024-12-06 17:54:28.911133] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.158 [2024-12-06 17:54:28.911227] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.158 [2024-12-06 17:54:28.911256] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.158 [2024-12-06 17:54:28.911273] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.158 [2024-12-06 17:54:28.911286] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.158 [2024-12-06 17:54:28.911317] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.158 qpair failed and we were unable to recover it. 00:38:47.158 [2024-12-06 17:54:28.921184] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.158 [2024-12-06 17:54:28.921277] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.158 [2024-12-06 17:54:28.921304] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.158 [2024-12-06 17:54:28.921328] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.158 [2024-12-06 17:54:28.921343] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.158 [2024-12-06 17:54:28.921388] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.158 qpair failed and we were unable to recover it. 00:38:47.158 [2024-12-06 17:54:28.931201] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.158 [2024-12-06 17:54:28.931326] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.158 [2024-12-06 17:54:28.931353] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.158 [2024-12-06 17:54:28.931368] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.158 [2024-12-06 17:54:28.931381] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.158 [2024-12-06 17:54:28.931412] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.159 qpair failed and we were unable to recover it. 00:38:47.159 [2024-12-06 17:54:28.941225] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.159 [2024-12-06 17:54:28.941309] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.159 [2024-12-06 17:54:28.941335] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.159 [2024-12-06 17:54:28.941350] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.159 [2024-12-06 17:54:28.941364] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.159 [2024-12-06 17:54:28.941395] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.159 qpair failed and we were unable to recover it. 00:38:47.418 [2024-12-06 17:54:28.951238] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.418 [2024-12-06 17:54:28.951324] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.418 [2024-12-06 17:54:28.951349] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.418 [2024-12-06 17:54:28.951364] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.418 [2024-12-06 17:54:28.951377] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.418 [2024-12-06 17:54:28.951407] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.418 qpair failed and we were unable to recover it. 00:38:47.418 [2024-12-06 17:54:28.961293] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.418 [2024-12-06 17:54:28.961405] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.418 [2024-12-06 17:54:28.961433] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.418 [2024-12-06 17:54:28.961448] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.418 [2024-12-06 17:54:28.961461] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.418 [2024-12-06 17:54:28.961498] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.418 qpair failed and we were unable to recover it. 00:38:47.418 [2024-12-06 17:54:28.971312] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.418 [2024-12-06 17:54:28.971426] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.418 [2024-12-06 17:54:28.971454] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.418 [2024-12-06 17:54:28.971470] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.418 [2024-12-06 17:54:28.971483] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.418 [2024-12-06 17:54:28.971513] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.418 qpair failed and we were unable to recover it. 00:38:47.418 [2024-12-06 17:54:28.981357] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.418 [2024-12-06 17:54:28.981477] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.418 [2024-12-06 17:54:28.981504] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.418 [2024-12-06 17:54:28.981520] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.418 [2024-12-06 17:54:28.981534] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.418 [2024-12-06 17:54:28.981564] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.418 qpair failed and we were unable to recover it. 00:38:47.418 [2024-12-06 17:54:28.991337] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.418 [2024-12-06 17:54:28.991425] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.418 [2024-12-06 17:54:28.991450] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.418 [2024-12-06 17:54:28.991464] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.418 [2024-12-06 17:54:28.991477] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.418 [2024-12-06 17:54:28.991508] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.418 qpair failed and we were unable to recover it. 00:38:47.418 [2024-12-06 17:54:29.001413] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.418 [2024-12-06 17:54:29.001507] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.418 [2024-12-06 17:54:29.001532] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.418 [2024-12-06 17:54:29.001546] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.418 [2024-12-06 17:54:29.001559] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.418 [2024-12-06 17:54:29.001589] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.418 qpair failed and we were unable to recover it. 00:38:47.418 [2024-12-06 17:54:29.011422] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.418 [2024-12-06 17:54:29.011555] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.419 [2024-12-06 17:54:29.011582] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.419 [2024-12-06 17:54:29.011597] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.419 [2024-12-06 17:54:29.011610] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.419 [2024-12-06 17:54:29.011640] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.419 qpair failed and we were unable to recover it. 00:38:47.419 [2024-12-06 17:54:29.021448] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.419 [2024-12-06 17:54:29.021540] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.419 [2024-12-06 17:54:29.021570] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.419 [2024-12-06 17:54:29.021586] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.419 [2024-12-06 17:54:29.021599] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.419 [2024-12-06 17:54:29.021630] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.419 qpair failed and we were unable to recover it. 00:38:47.419 [2024-12-06 17:54:29.031552] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.419 [2024-12-06 17:54:29.031640] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.419 [2024-12-06 17:54:29.031675] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.419 [2024-12-06 17:54:29.031693] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.419 [2024-12-06 17:54:29.031707] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.419 [2024-12-06 17:54:29.031738] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.419 qpair failed and we were unable to recover it. 00:38:47.419 [2024-12-06 17:54:29.041550] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.419 [2024-12-06 17:54:29.041646] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.419 [2024-12-06 17:54:29.041691] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.419 [2024-12-06 17:54:29.041706] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.419 [2024-12-06 17:54:29.041719] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.419 [2024-12-06 17:54:29.041750] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.419 qpair failed and we were unable to recover it. 00:38:47.419 [2024-12-06 17:54:29.051501] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.419 [2024-12-06 17:54:29.051592] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.419 [2024-12-06 17:54:29.051622] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.419 [2024-12-06 17:54:29.051638] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.419 [2024-12-06 17:54:29.051660] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.419 [2024-12-06 17:54:29.051700] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.419 qpair failed and we were unable to recover it. 00:38:47.419 [2024-12-06 17:54:29.061564] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.419 [2024-12-06 17:54:29.061655] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.419 [2024-12-06 17:54:29.061707] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.419 [2024-12-06 17:54:29.061731] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.419 [2024-12-06 17:54:29.061754] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.419 [2024-12-06 17:54:29.061794] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.419 qpair failed and we were unable to recover it. 00:38:47.419 [2024-12-06 17:54:29.071596] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.419 [2024-12-06 17:54:29.071694] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.419 [2024-12-06 17:54:29.071722] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.419 [2024-12-06 17:54:29.071737] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.419 [2024-12-06 17:54:29.071750] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.419 [2024-12-06 17:54:29.071782] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.419 qpair failed and we were unable to recover it. 00:38:47.419 [2024-12-06 17:54:29.081658] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.419 [2024-12-06 17:54:29.081806] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.419 [2024-12-06 17:54:29.081833] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.419 [2024-12-06 17:54:29.081848] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.419 [2024-12-06 17:54:29.081861] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.419 [2024-12-06 17:54:29.081893] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.419 qpair failed and we were unable to recover it. 00:38:47.419 [2024-12-06 17:54:29.091652] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.419 [2024-12-06 17:54:29.091746] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.419 [2024-12-06 17:54:29.091771] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.419 [2024-12-06 17:54:29.091787] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.419 [2024-12-06 17:54:29.091800] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.419 [2024-12-06 17:54:29.091837] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.419 qpair failed and we were unable to recover it. 00:38:47.419 [2024-12-06 17:54:29.101694] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.419 [2024-12-06 17:54:29.101835] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.419 [2024-12-06 17:54:29.101863] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.419 [2024-12-06 17:54:29.101879] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.419 [2024-12-06 17:54:29.101892] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.419 [2024-12-06 17:54:29.101922] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.419 qpair failed and we were unable to recover it. 00:38:47.419 [2024-12-06 17:54:29.111695] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.419 [2024-12-06 17:54:29.111821] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.419 [2024-12-06 17:54:29.111847] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.419 [2024-12-06 17:54:29.111862] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.419 [2024-12-06 17:54:29.111874] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.419 [2024-12-06 17:54:29.111906] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.419 qpair failed and we were unable to recover it. 00:38:47.419 [2024-12-06 17:54:29.121746] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.419 [2024-12-06 17:54:29.121839] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.419 [2024-12-06 17:54:29.121864] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.419 [2024-12-06 17:54:29.121879] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.419 [2024-12-06 17:54:29.121892] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.419 [2024-12-06 17:54:29.121922] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.419 qpair failed and we were unable to recover it. 00:38:47.419 [2024-12-06 17:54:29.131746] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.419 [2024-12-06 17:54:29.131826] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.419 [2024-12-06 17:54:29.131850] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.419 [2024-12-06 17:54:29.131866] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.419 [2024-12-06 17:54:29.131878] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.419 [2024-12-06 17:54:29.131909] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.419 qpair failed and we were unable to recover it. 00:38:47.419 [2024-12-06 17:54:29.141766] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.419 [2024-12-06 17:54:29.141853] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.420 [2024-12-06 17:54:29.141878] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.420 [2024-12-06 17:54:29.141893] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.420 [2024-12-06 17:54:29.141905] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.420 [2024-12-06 17:54:29.141937] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.420 qpair failed and we were unable to recover it. 00:38:47.420 [2024-12-06 17:54:29.151810] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.420 [2024-12-06 17:54:29.151895] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.420 [2024-12-06 17:54:29.151919] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.420 [2024-12-06 17:54:29.151934] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.420 [2024-12-06 17:54:29.151946] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.420 [2024-12-06 17:54:29.151977] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.420 qpair failed and we were unable to recover it. 00:38:47.420 [2024-12-06 17:54:29.161864] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.420 [2024-12-06 17:54:29.161988] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.420 [2024-12-06 17:54:29.162013] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.420 [2024-12-06 17:54:29.162028] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.420 [2024-12-06 17:54:29.162041] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.420 [2024-12-06 17:54:29.162072] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.420 qpair failed and we were unable to recover it. 00:38:47.420 [2024-12-06 17:54:29.171861] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.420 [2024-12-06 17:54:29.171947] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.420 [2024-12-06 17:54:29.171972] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.420 [2024-12-06 17:54:29.171985] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.420 [2024-12-06 17:54:29.172007] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.420 [2024-12-06 17:54:29.172038] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.420 qpair failed and we were unable to recover it. 00:38:47.420 [2024-12-06 17:54:29.181923] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.420 [2024-12-06 17:54:29.182024] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.420 [2024-12-06 17:54:29.182054] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.420 [2024-12-06 17:54:29.182069] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.420 [2024-12-06 17:54:29.182083] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.420 [2024-12-06 17:54:29.182113] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.420 qpair failed and we were unable to recover it. 00:38:47.420 [2024-12-06 17:54:29.192025] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.420 [2024-12-06 17:54:29.192127] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.420 [2024-12-06 17:54:29.192152] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.420 [2024-12-06 17:54:29.192167] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.420 [2024-12-06 17:54:29.192180] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.420 [2024-12-06 17:54:29.192212] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.420 qpair failed and we were unable to recover it. 00:38:47.420 [2024-12-06 17:54:29.201969] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.420 [2024-12-06 17:54:29.202056] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.420 [2024-12-06 17:54:29.202081] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.420 [2024-12-06 17:54:29.202097] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.420 [2024-12-06 17:54:29.202109] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.420 [2024-12-06 17:54:29.202140] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.420 qpair failed and we were unable to recover it. 00:38:47.420 [2024-12-06 17:54:29.211975] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.420 [2024-12-06 17:54:29.212071] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.420 [2024-12-06 17:54:29.212096] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.420 [2024-12-06 17:54:29.212111] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.420 [2024-12-06 17:54:29.212123] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.420 [2024-12-06 17:54:29.212154] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.420 qpair failed and we were unable to recover it. 00:38:47.420 [2024-12-06 17:54:29.222016] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.420 [2024-12-06 17:54:29.222100] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.420 [2024-12-06 17:54:29.222125] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.420 [2024-12-06 17:54:29.222140] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.420 [2024-12-06 17:54:29.222159] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.420 [2024-12-06 17:54:29.222190] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.420 qpair failed and we were unable to recover it. 00:38:47.420 [2024-12-06 17:54:29.232111] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.420 [2024-12-06 17:54:29.232220] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.420 [2024-12-06 17:54:29.232245] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.420 [2024-12-06 17:54:29.232260] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.420 [2024-12-06 17:54:29.232273] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.420 [2024-12-06 17:54:29.232304] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.420 qpair failed and we were unable to recover it. 00:38:47.420 [2024-12-06 17:54:29.242070] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.420 [2024-12-06 17:54:29.242161] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.420 [2024-12-06 17:54:29.242186] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.420 [2024-12-06 17:54:29.242201] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.420 [2024-12-06 17:54:29.242214] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.420 [2024-12-06 17:54:29.242244] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.420 qpair failed and we were unable to recover it. 00:38:47.420 [2024-12-06 17:54:29.252088] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.420 [2024-12-06 17:54:29.252175] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.420 [2024-12-06 17:54:29.252200] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.420 [2024-12-06 17:54:29.252215] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.420 [2024-12-06 17:54:29.252228] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.420 [2024-12-06 17:54:29.252259] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.420 qpair failed and we were unable to recover it. 00:38:47.680 [2024-12-06 17:54:29.262206] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.680 [2024-12-06 17:54:29.262294] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.680 [2024-12-06 17:54:29.262319] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.680 [2024-12-06 17:54:29.262334] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.680 [2024-12-06 17:54:29.262347] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.680 [2024-12-06 17:54:29.262378] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.680 qpair failed and we were unable to recover it. 00:38:47.680 [2024-12-06 17:54:29.272266] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.680 [2024-12-06 17:54:29.272358] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.680 [2024-12-06 17:54:29.272383] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.680 [2024-12-06 17:54:29.272397] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.680 [2024-12-06 17:54:29.272411] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.680 [2024-12-06 17:54:29.272441] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.680 qpair failed and we were unable to recover it. 00:38:47.680 [2024-12-06 17:54:29.282252] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.680 [2024-12-06 17:54:29.282356] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.680 [2024-12-06 17:54:29.282380] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.680 [2024-12-06 17:54:29.282395] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.680 [2024-12-06 17:54:29.282408] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.680 [2024-12-06 17:54:29.282439] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.680 qpair failed and we were unable to recover it. 00:38:47.680 [2024-12-06 17:54:29.292217] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.680 [2024-12-06 17:54:29.292303] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.680 [2024-12-06 17:54:29.292328] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.680 [2024-12-06 17:54:29.292344] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.680 [2024-12-06 17:54:29.292356] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.680 [2024-12-06 17:54:29.292387] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.680 qpair failed and we were unable to recover it. 00:38:47.680 [2024-12-06 17:54:29.302253] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.680 [2024-12-06 17:54:29.302340] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.680 [2024-12-06 17:54:29.302365] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.680 [2024-12-06 17:54:29.302380] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.680 [2024-12-06 17:54:29.302393] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.680 [2024-12-06 17:54:29.302424] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.680 qpair failed and we were unable to recover it. 00:38:47.680 [2024-12-06 17:54:29.312254] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.680 [2024-12-06 17:54:29.312335] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.680 [2024-12-06 17:54:29.312366] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.680 [2024-12-06 17:54:29.312383] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.680 [2024-12-06 17:54:29.312395] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.680 [2024-12-06 17:54:29.312435] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.680 qpair failed and we were unable to recover it. 00:38:47.680 [2024-12-06 17:54:29.322332] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.680 [2024-12-06 17:54:29.322434] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.680 [2024-12-06 17:54:29.322461] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.680 [2024-12-06 17:54:29.322475] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.680 [2024-12-06 17:54:29.322488] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.680 [2024-12-06 17:54:29.322532] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.680 qpair failed and we were unable to recover it. 00:38:47.680 [2024-12-06 17:54:29.332335] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.680 [2024-12-06 17:54:29.332418] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.680 [2024-12-06 17:54:29.332445] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.680 [2024-12-06 17:54:29.332460] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.680 [2024-12-06 17:54:29.332473] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.680 [2024-12-06 17:54:29.332504] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.680 qpair failed and we were unable to recover it. 00:38:47.680 [2024-12-06 17:54:29.342333] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.680 [2024-12-06 17:54:29.342423] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.680 [2024-12-06 17:54:29.342448] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.680 [2024-12-06 17:54:29.342463] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.680 [2024-12-06 17:54:29.342476] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.680 [2024-12-06 17:54:29.342507] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.680 qpair failed and we were unable to recover it. 00:38:47.680 [2024-12-06 17:54:29.352356] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.680 [2024-12-06 17:54:29.352454] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.680 [2024-12-06 17:54:29.352480] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.680 [2024-12-06 17:54:29.352501] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.680 [2024-12-06 17:54:29.352515] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.680 [2024-12-06 17:54:29.352546] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.680 qpair failed and we were unable to recover it. 00:38:47.680 [2024-12-06 17:54:29.362406] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.680 [2024-12-06 17:54:29.362496] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.680 [2024-12-06 17:54:29.362522] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.680 [2024-12-06 17:54:29.362537] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.680 [2024-12-06 17:54:29.362550] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.680 [2024-12-06 17:54:29.362581] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.680 qpair failed and we were unable to recover it. 00:38:47.680 [2024-12-06 17:54:29.372562] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.680 [2024-12-06 17:54:29.372698] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.680 [2024-12-06 17:54:29.372725] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.680 [2024-12-06 17:54:29.372739] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.680 [2024-12-06 17:54:29.372753] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.680 [2024-12-06 17:54:29.372784] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.680 qpair failed and we were unable to recover it. 00:38:47.680 [2024-12-06 17:54:29.382487] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.680 [2024-12-06 17:54:29.382572] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.680 [2024-12-06 17:54:29.382598] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.680 [2024-12-06 17:54:29.382612] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.680 [2024-12-06 17:54:29.382625] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.681 [2024-12-06 17:54:29.382655] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.681 qpair failed and we were unable to recover it. 00:38:47.681 [2024-12-06 17:54:29.392480] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.681 [2024-12-06 17:54:29.392566] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.681 [2024-12-06 17:54:29.392592] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.681 [2024-12-06 17:54:29.392606] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.681 [2024-12-06 17:54:29.392619] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.681 [2024-12-06 17:54:29.392650] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.681 qpair failed and we were unable to recover it. 00:38:47.681 [2024-12-06 17:54:29.402544] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.681 [2024-12-06 17:54:29.402628] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.681 [2024-12-06 17:54:29.402653] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.681 [2024-12-06 17:54:29.402677] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.681 [2024-12-06 17:54:29.402692] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.681 [2024-12-06 17:54:29.402724] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.681 qpair failed and we were unable to recover it. 00:38:47.681 [2024-12-06 17:54:29.412622] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.681 [2024-12-06 17:54:29.412721] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.681 [2024-12-06 17:54:29.412747] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.681 [2024-12-06 17:54:29.412762] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.681 [2024-12-06 17:54:29.412775] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.681 [2024-12-06 17:54:29.412806] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.681 qpair failed and we were unable to recover it. 00:38:47.681 [2024-12-06 17:54:29.422573] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.681 [2024-12-06 17:54:29.422648] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.681 [2024-12-06 17:54:29.422683] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.681 [2024-12-06 17:54:29.422699] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.681 [2024-12-06 17:54:29.422712] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.681 [2024-12-06 17:54:29.422743] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.681 qpair failed and we were unable to recover it. 00:38:47.681 [2024-12-06 17:54:29.432629] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.681 [2024-12-06 17:54:29.432718] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.681 [2024-12-06 17:54:29.432743] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.681 [2024-12-06 17:54:29.432758] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.681 [2024-12-06 17:54:29.432771] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.681 [2024-12-06 17:54:29.432802] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.681 qpair failed and we were unable to recover it. 00:38:47.681 [2024-12-06 17:54:29.442615] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.681 [2024-12-06 17:54:29.442714] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.681 [2024-12-06 17:54:29.442740] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.681 [2024-12-06 17:54:29.442756] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.681 [2024-12-06 17:54:29.442768] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.681 [2024-12-06 17:54:29.442799] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.681 qpair failed and we were unable to recover it. 00:38:47.681 [2024-12-06 17:54:29.452670] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.681 [2024-12-06 17:54:29.452752] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.681 [2024-12-06 17:54:29.452777] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.681 [2024-12-06 17:54:29.452792] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.681 [2024-12-06 17:54:29.452805] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.681 [2024-12-06 17:54:29.452837] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.681 qpair failed and we were unable to recover it. 00:38:47.681 [2024-12-06 17:54:29.462679] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.681 [2024-12-06 17:54:29.462770] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.681 [2024-12-06 17:54:29.462799] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.681 [2024-12-06 17:54:29.462816] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.681 [2024-12-06 17:54:29.462829] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.681 [2024-12-06 17:54:29.462862] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.681 qpair failed and we were unable to recover it. 00:38:47.681 [2024-12-06 17:54:29.472711] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.681 [2024-12-06 17:54:29.472797] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.681 [2024-12-06 17:54:29.472823] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.681 [2024-12-06 17:54:29.472839] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.681 [2024-12-06 17:54:29.472851] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.681 [2024-12-06 17:54:29.472882] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.681 qpair failed and we were unable to recover it. 00:38:47.681 [2024-12-06 17:54:29.482751] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.681 [2024-12-06 17:54:29.482843] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.681 [2024-12-06 17:54:29.482868] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.681 [2024-12-06 17:54:29.482888] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.681 [2024-12-06 17:54:29.482903] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.681 [2024-12-06 17:54:29.482934] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.681 qpair failed and we were unable to recover it. 00:38:47.681 [2024-12-06 17:54:29.492934] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.681 [2024-12-06 17:54:29.493034] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.681 [2024-12-06 17:54:29.493059] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.681 [2024-12-06 17:54:29.493075] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.681 [2024-12-06 17:54:29.493088] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.681 [2024-12-06 17:54:29.493119] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.681 qpair failed and we were unable to recover it. 00:38:47.681 [2024-12-06 17:54:29.502880] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.681 [2024-12-06 17:54:29.502967] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.681 [2024-12-06 17:54:29.502992] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.681 [2024-12-06 17:54:29.503007] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.681 [2024-12-06 17:54:29.503021] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.681 [2024-12-06 17:54:29.503052] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.681 qpair failed and we were unable to recover it. 00:38:47.681 [2024-12-06 17:54:29.512877] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.681 [2024-12-06 17:54:29.512982] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.681 [2024-12-06 17:54:29.513007] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.681 [2024-12-06 17:54:29.513022] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.681 [2024-12-06 17:54:29.513035] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.682 [2024-12-06 17:54:29.513067] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.682 qpair failed and we were unable to recover it. 00:38:47.939 [2024-12-06 17:54:29.522908] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.939 [2024-12-06 17:54:29.523000] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.939 [2024-12-06 17:54:29.523025] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.939 [2024-12-06 17:54:29.523040] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.939 [2024-12-06 17:54:29.523053] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.939 [2024-12-06 17:54:29.523090] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.939 qpair failed and we were unable to recover it. 00:38:47.939 [2024-12-06 17:54:29.532906] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.939 [2024-12-06 17:54:29.532986] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.939 [2024-12-06 17:54:29.533012] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.939 [2024-12-06 17:54:29.533028] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.939 [2024-12-06 17:54:29.533041] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.939 [2024-12-06 17:54:29.533083] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.939 qpair failed and we were unable to recover it. 00:38:47.939 [2024-12-06 17:54:29.542908] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.939 [2024-12-06 17:54:29.542994] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.939 [2024-12-06 17:54:29.543021] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.939 [2024-12-06 17:54:29.543036] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.939 [2024-12-06 17:54:29.543049] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.939 [2024-12-06 17:54:29.543092] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.939 qpair failed and we were unable to recover it. 00:38:47.939 [2024-12-06 17:54:29.552927] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.939 [2024-12-06 17:54:29.553010] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.939 [2024-12-06 17:54:29.553035] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.940 [2024-12-06 17:54:29.553050] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.940 [2024-12-06 17:54:29.553063] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.940 [2024-12-06 17:54:29.553094] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.940 qpair failed and we were unable to recover it. 00:38:47.940 [2024-12-06 17:54:29.562947] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.940 [2024-12-06 17:54:29.563040] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.940 [2024-12-06 17:54:29.563065] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.940 [2024-12-06 17:54:29.563080] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.940 [2024-12-06 17:54:29.563094] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.940 [2024-12-06 17:54:29.563139] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.940 qpair failed and we were unable to recover it. 00:38:47.940 [2024-12-06 17:54:29.573018] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.940 [2024-12-06 17:54:29.573109] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.940 [2024-12-06 17:54:29.573137] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.940 [2024-12-06 17:54:29.573153] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.940 [2024-12-06 17:54:29.573167] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.940 [2024-12-06 17:54:29.573198] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.940 qpair failed and we were unable to recover it. 00:38:47.940 [2024-12-06 17:54:29.583065] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.940 [2024-12-06 17:54:29.583149] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.940 [2024-12-06 17:54:29.583176] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.940 [2024-12-06 17:54:29.583191] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.940 [2024-12-06 17:54:29.583204] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.940 [2024-12-06 17:54:29.583234] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.940 qpair failed and we were unable to recover it. 00:38:47.940 [2024-12-06 17:54:29.593090] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.940 [2024-12-06 17:54:29.593181] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.940 [2024-12-06 17:54:29.593207] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.940 [2024-12-06 17:54:29.593223] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.940 [2024-12-06 17:54:29.593236] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.940 [2024-12-06 17:54:29.593267] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.940 qpair failed and we were unable to recover it. 00:38:47.940 [2024-12-06 17:54:29.603106] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.940 [2024-12-06 17:54:29.603201] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.940 [2024-12-06 17:54:29.603227] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.940 [2024-12-06 17:54:29.603241] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.940 [2024-12-06 17:54:29.603254] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.940 [2024-12-06 17:54:29.603285] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.940 qpair failed and we were unable to recover it. 00:38:47.940 [2024-12-06 17:54:29.613180] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.940 [2024-12-06 17:54:29.613265] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.940 [2024-12-06 17:54:29.613296] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.940 [2024-12-06 17:54:29.613312] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.940 [2024-12-06 17:54:29.613325] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.940 [2024-12-06 17:54:29.613356] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.940 qpair failed and we were unable to recover it. 00:38:47.940 [2024-12-06 17:54:29.623183] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.940 [2024-12-06 17:54:29.623296] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.940 [2024-12-06 17:54:29.623323] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.940 [2024-12-06 17:54:29.623338] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.940 [2024-12-06 17:54:29.623350] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.940 [2024-12-06 17:54:29.623382] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.940 qpair failed and we were unable to recover it. 00:38:47.940 [2024-12-06 17:54:29.633152] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.940 [2024-12-06 17:54:29.633233] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.940 [2024-12-06 17:54:29.633257] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.940 [2024-12-06 17:54:29.633272] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.940 [2024-12-06 17:54:29.633285] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.940 [2024-12-06 17:54:29.633316] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.940 qpair failed and we were unable to recover it. 00:38:47.940 [2024-12-06 17:54:29.643206] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.940 [2024-12-06 17:54:29.643296] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.940 [2024-12-06 17:54:29.643321] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.940 [2024-12-06 17:54:29.643336] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.940 [2024-12-06 17:54:29.643349] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.940 [2024-12-06 17:54:29.643380] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.940 qpair failed and we were unable to recover it. 00:38:47.940 [2024-12-06 17:54:29.653230] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.940 [2024-12-06 17:54:29.653313] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.940 [2024-12-06 17:54:29.653339] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.940 [2024-12-06 17:54:29.653354] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.940 [2024-12-06 17:54:29.653367] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.940 [2024-12-06 17:54:29.653403] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.940 qpair failed and we were unable to recover it. 00:38:47.940 [2024-12-06 17:54:29.663370] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.940 [2024-12-06 17:54:29.663502] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.940 [2024-12-06 17:54:29.663527] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.940 [2024-12-06 17:54:29.663542] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.940 [2024-12-06 17:54:29.663555] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.940 [2024-12-06 17:54:29.663586] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.940 qpair failed and we were unable to recover it. 00:38:47.940 [2024-12-06 17:54:29.673277] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.940 [2024-12-06 17:54:29.673361] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.940 [2024-12-06 17:54:29.673389] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.940 [2024-12-06 17:54:29.673406] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.940 [2024-12-06 17:54:29.673420] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.940 [2024-12-06 17:54:29.673451] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.940 qpair failed and we were unable to recover it. 00:38:47.940 [2024-12-06 17:54:29.683322] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.940 [2024-12-06 17:54:29.683414] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.940 [2024-12-06 17:54:29.683441] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.940 [2024-12-06 17:54:29.683459] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.941 [2024-12-06 17:54:29.683473] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.941 [2024-12-06 17:54:29.683504] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.941 qpair failed and we were unable to recover it. 00:38:47.941 [2024-12-06 17:54:29.693341] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.941 [2024-12-06 17:54:29.693424] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.941 [2024-12-06 17:54:29.693450] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.941 [2024-12-06 17:54:29.693465] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.941 [2024-12-06 17:54:29.693478] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.941 [2024-12-06 17:54:29.693509] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.941 qpair failed and we were unable to recover it. 00:38:47.941 [2024-12-06 17:54:29.703346] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.941 [2024-12-06 17:54:29.703434] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.941 [2024-12-06 17:54:29.703459] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.941 [2024-12-06 17:54:29.703474] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.941 [2024-12-06 17:54:29.703487] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.941 [2024-12-06 17:54:29.703518] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.941 qpair failed and we were unable to recover it. 00:38:47.941 [2024-12-06 17:54:29.713393] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.941 [2024-12-06 17:54:29.713502] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.941 [2024-12-06 17:54:29.713530] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.941 [2024-12-06 17:54:29.713545] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.941 [2024-12-06 17:54:29.713558] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.941 [2024-12-06 17:54:29.713589] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.941 qpair failed and we were unable to recover it. 00:38:47.941 [2024-12-06 17:54:29.723397] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.941 [2024-12-06 17:54:29.723486] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.941 [2024-12-06 17:54:29.723512] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.941 [2024-12-06 17:54:29.723526] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.941 [2024-12-06 17:54:29.723540] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.941 [2024-12-06 17:54:29.723570] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.941 qpair failed and we were unable to recover it. 00:38:47.941 [2024-12-06 17:54:29.733448] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.941 [2024-12-06 17:54:29.733576] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.941 [2024-12-06 17:54:29.733601] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.941 [2024-12-06 17:54:29.733616] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.941 [2024-12-06 17:54:29.733629] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.941 [2024-12-06 17:54:29.733658] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.941 qpair failed and we were unable to recover it. 00:38:47.941 [2024-12-06 17:54:29.743445] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.941 [2024-12-06 17:54:29.743531] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.941 [2024-12-06 17:54:29.743562] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.941 [2024-12-06 17:54:29.743578] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.941 [2024-12-06 17:54:29.743590] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.941 [2024-12-06 17:54:29.743621] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.941 qpair failed and we were unable to recover it. 00:38:47.941 [2024-12-06 17:54:29.753485] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.941 [2024-12-06 17:54:29.753608] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.941 [2024-12-06 17:54:29.753634] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.941 [2024-12-06 17:54:29.753649] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.941 [2024-12-06 17:54:29.753662] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.941 [2024-12-06 17:54:29.753704] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.941 qpair failed and we were unable to recover it. 00:38:47.941 [2024-12-06 17:54:29.763527] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.941 [2024-12-06 17:54:29.763614] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.941 [2024-12-06 17:54:29.763639] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.941 [2024-12-06 17:54:29.763654] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.941 [2024-12-06 17:54:29.763675] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.941 [2024-12-06 17:54:29.763721] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.941 qpair failed and we were unable to recover it. 00:38:47.941 [2024-12-06 17:54:29.773566] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:47.941 [2024-12-06 17:54:29.773693] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:47.941 [2024-12-06 17:54:29.773718] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:47.941 [2024-12-06 17:54:29.773732] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:47.941 [2024-12-06 17:54:29.773745] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:47.941 [2024-12-06 17:54:29.773777] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:47.941 qpair failed and we were unable to recover it. 00:38:48.199 [2024-12-06 17:54:29.783567] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.199 [2024-12-06 17:54:29.783657] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.199 [2024-12-06 17:54:29.783701] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.199 [2024-12-06 17:54:29.783716] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.199 [2024-12-06 17:54:29.783735] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.199 [2024-12-06 17:54:29.783767] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.199 qpair failed and we were unable to recover it. 00:38:48.199 [2024-12-06 17:54:29.793608] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.199 [2024-12-06 17:54:29.793701] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.199 [2024-12-06 17:54:29.793727] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.199 [2024-12-06 17:54:29.793742] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.199 [2024-12-06 17:54:29.793755] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.199 [2024-12-06 17:54:29.793786] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.199 qpair failed and we were unable to recover it. 00:38:48.199 [2024-12-06 17:54:29.803692] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.199 [2024-12-06 17:54:29.803788] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.199 [2024-12-06 17:54:29.803813] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.199 [2024-12-06 17:54:29.803828] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.199 [2024-12-06 17:54:29.803841] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.199 [2024-12-06 17:54:29.803872] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.199 qpair failed and we were unable to recover it. 00:38:48.199 [2024-12-06 17:54:29.813694] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.199 [2024-12-06 17:54:29.813778] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.199 [2024-12-06 17:54:29.813804] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.199 [2024-12-06 17:54:29.813819] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.199 [2024-12-06 17:54:29.813833] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.199 [2024-12-06 17:54:29.813872] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.199 qpair failed and we were unable to recover it. 00:38:48.199 [2024-12-06 17:54:29.823703] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.199 [2024-12-06 17:54:29.823787] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.199 [2024-12-06 17:54:29.823814] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.199 [2024-12-06 17:54:29.823830] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.199 [2024-12-06 17:54:29.823843] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.199 [2024-12-06 17:54:29.823875] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.199 qpair failed and we were unable to recover it. 00:38:48.199 [2024-12-06 17:54:29.833738] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.199 [2024-12-06 17:54:29.833825] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.199 [2024-12-06 17:54:29.833854] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.199 [2024-12-06 17:54:29.833870] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.199 [2024-12-06 17:54:29.833883] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.199 [2024-12-06 17:54:29.833926] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.199 qpair failed and we were unable to recover it. 00:38:48.199 [2024-12-06 17:54:29.843758] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.199 [2024-12-06 17:54:29.843847] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.199 [2024-12-06 17:54:29.843874] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.199 [2024-12-06 17:54:29.843889] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.199 [2024-12-06 17:54:29.843902] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.199 [2024-12-06 17:54:29.843933] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.199 qpair failed and we were unable to recover it. 00:38:48.199 [2024-12-06 17:54:29.853773] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.199 [2024-12-06 17:54:29.853861] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.199 [2024-12-06 17:54:29.853886] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.199 [2024-12-06 17:54:29.853901] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.199 [2024-12-06 17:54:29.853914] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.199 [2024-12-06 17:54:29.853945] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.199 qpair failed and we were unable to recover it. 00:38:48.199 [2024-12-06 17:54:29.863791] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.199 [2024-12-06 17:54:29.863870] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.199 [2024-12-06 17:54:29.863896] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.199 [2024-12-06 17:54:29.863912] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.199 [2024-12-06 17:54:29.863925] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.199 [2024-12-06 17:54:29.863955] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.199 qpair failed and we were unable to recover it. 00:38:48.199 [2024-12-06 17:54:29.873839] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.199 [2024-12-06 17:54:29.873921] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.200 [2024-12-06 17:54:29.873951] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.200 [2024-12-06 17:54:29.873967] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.200 [2024-12-06 17:54:29.873979] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.200 [2024-12-06 17:54:29.874010] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.200 qpair failed and we were unable to recover it. 00:38:48.200 [2024-12-06 17:54:29.883882] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.200 [2024-12-06 17:54:29.883971] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.200 [2024-12-06 17:54:29.883997] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.200 [2024-12-06 17:54:29.884012] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.200 [2024-12-06 17:54:29.884026] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.200 [2024-12-06 17:54:29.884057] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.200 qpair failed and we were unable to recover it. 00:38:48.200 [2024-12-06 17:54:29.893909] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.200 [2024-12-06 17:54:29.893995] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.200 [2024-12-06 17:54:29.894021] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.200 [2024-12-06 17:54:29.894036] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.200 [2024-12-06 17:54:29.894049] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.200 [2024-12-06 17:54:29.894092] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.200 qpair failed and we were unable to recover it. 00:38:48.200 [2024-12-06 17:54:29.903900] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.200 [2024-12-06 17:54:29.903995] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.200 [2024-12-06 17:54:29.904020] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.200 [2024-12-06 17:54:29.904034] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.200 [2024-12-06 17:54:29.904047] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.200 [2024-12-06 17:54:29.904078] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.200 qpair failed and we were unable to recover it. 00:38:48.200 [2024-12-06 17:54:29.913944] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.200 [2024-12-06 17:54:29.914032] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.200 [2024-12-06 17:54:29.914058] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.200 [2024-12-06 17:54:29.914078] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.200 [2024-12-06 17:54:29.914091] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.200 [2024-12-06 17:54:29.914122] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.200 qpair failed and we were unable to recover it. 00:38:48.200 [2024-12-06 17:54:29.924038] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.200 [2024-12-06 17:54:29.924148] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.200 [2024-12-06 17:54:29.924173] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.200 [2024-12-06 17:54:29.924188] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.200 [2024-12-06 17:54:29.924200] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.200 [2024-12-06 17:54:29.924231] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.200 qpair failed and we were unable to recover it. 00:38:48.200 [2024-12-06 17:54:29.934058] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.200 [2024-12-06 17:54:29.934173] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.200 [2024-12-06 17:54:29.934199] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.200 [2024-12-06 17:54:29.934214] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.200 [2024-12-06 17:54:29.934227] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.200 [2024-12-06 17:54:29.934270] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.200 qpair failed and we were unable to recover it. 00:38:48.200 [2024-12-06 17:54:29.944084] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.200 [2024-12-06 17:54:29.944185] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.200 [2024-12-06 17:54:29.944210] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.200 [2024-12-06 17:54:29.944225] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.200 [2024-12-06 17:54:29.944238] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.200 [2024-12-06 17:54:29.944269] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.200 qpair failed and we were unable to recover it. 00:38:48.200 [2024-12-06 17:54:29.954067] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.200 [2024-12-06 17:54:29.954151] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.200 [2024-12-06 17:54:29.954177] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.200 [2024-12-06 17:54:29.954193] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.200 [2024-12-06 17:54:29.954206] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.200 [2024-12-06 17:54:29.954236] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.200 qpair failed and we were unable to recover it. 00:38:48.200 [2024-12-06 17:54:29.964091] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.200 [2024-12-06 17:54:29.964181] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.200 [2024-12-06 17:54:29.964207] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.200 [2024-12-06 17:54:29.964222] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.200 [2024-12-06 17:54:29.964236] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.200 [2024-12-06 17:54:29.964266] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.200 qpair failed and we were unable to recover it. 00:38:48.200 [2024-12-06 17:54:29.974161] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.200 [2024-12-06 17:54:29.974249] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.200 [2024-12-06 17:54:29.974275] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.200 [2024-12-06 17:54:29.974293] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.200 [2024-12-06 17:54:29.974307] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.201 [2024-12-06 17:54:29.974338] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.201 qpair failed and we were unable to recover it. 00:38:48.201 [2024-12-06 17:54:29.984132] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.201 [2024-12-06 17:54:29.984215] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.201 [2024-12-06 17:54:29.984241] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.201 [2024-12-06 17:54:29.984256] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.201 [2024-12-06 17:54:29.984269] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.201 [2024-12-06 17:54:29.984300] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.201 qpair failed and we were unable to recover it. 00:38:48.201 [2024-12-06 17:54:29.994176] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.201 [2024-12-06 17:54:29.994292] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.201 [2024-12-06 17:54:29.994317] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.201 [2024-12-06 17:54:29.994332] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.201 [2024-12-06 17:54:29.994345] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.201 [2024-12-06 17:54:29.994376] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.201 qpair failed and we were unable to recover it. 00:38:48.201 [2024-12-06 17:54:30.004248] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.201 [2024-12-06 17:54:30.004395] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.201 [2024-12-06 17:54:30.004444] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.201 [2024-12-06 17:54:30.004463] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.201 [2024-12-06 17:54:30.004477] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.201 [2024-12-06 17:54:30.004509] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.201 qpair failed and we were unable to recover it. 00:38:48.201 [2024-12-06 17:54:30.014257] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.201 [2024-12-06 17:54:30.014348] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.201 [2024-12-06 17:54:30.014373] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.201 [2024-12-06 17:54:30.014389] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.201 [2024-12-06 17:54:30.014402] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.201 [2024-12-06 17:54:30.014433] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.201 qpair failed and we were unable to recover it. 00:38:48.201 [2024-12-06 17:54:30.024255] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.201 [2024-12-06 17:54:30.024384] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.201 [2024-12-06 17:54:30.024410] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.201 [2024-12-06 17:54:30.024425] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.201 [2024-12-06 17:54:30.024439] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.201 [2024-12-06 17:54:30.024470] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.201 qpair failed and we were unable to recover it. 00:38:48.201 [2024-12-06 17:54:30.034340] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.201 [2024-12-06 17:54:30.034453] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.201 [2024-12-06 17:54:30.034481] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.201 [2024-12-06 17:54:30.034497] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.201 [2024-12-06 17:54:30.034511] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.201 [2024-12-06 17:54:30.034553] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.201 qpair failed and we were unable to recover it. 00:38:48.461 [2024-12-06 17:54:30.044322] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.461 [2024-12-06 17:54:30.044414] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.461 [2024-12-06 17:54:30.044439] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.461 [2024-12-06 17:54:30.044460] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.461 [2024-12-06 17:54:30.044473] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.461 [2024-12-06 17:54:30.044503] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.461 qpair failed and we were unable to recover it. 00:38:48.461 [2024-12-06 17:54:30.054359] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.461 [2024-12-06 17:54:30.054443] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.461 [2024-12-06 17:54:30.054468] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.461 [2024-12-06 17:54:30.054484] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.461 [2024-12-06 17:54:30.054496] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.461 [2024-12-06 17:54:30.054527] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.461 qpair failed and we were unable to recover it. 00:38:48.461 [2024-12-06 17:54:30.064377] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.461 [2024-12-06 17:54:30.064465] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.461 [2024-12-06 17:54:30.064495] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.461 [2024-12-06 17:54:30.064512] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.461 [2024-12-06 17:54:30.064530] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.461 [2024-12-06 17:54:30.064577] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.461 qpair failed and we were unable to recover it. 00:38:48.461 [2024-12-06 17:54:30.074381] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.461 [2024-12-06 17:54:30.074473] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.462 [2024-12-06 17:54:30.074500] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.462 [2024-12-06 17:54:30.074514] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.462 [2024-12-06 17:54:30.074527] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.462 [2024-12-06 17:54:30.074559] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.462 qpair failed and we were unable to recover it. 00:38:48.462 [2024-12-06 17:54:30.084440] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.462 [2024-12-06 17:54:30.084568] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.462 [2024-12-06 17:54:30.084595] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.462 [2024-12-06 17:54:30.084609] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.462 [2024-12-06 17:54:30.084623] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.462 [2024-12-06 17:54:30.084660] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.462 qpair failed and we were unable to recover it. 00:38:48.462 [2024-12-06 17:54:30.094486] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.462 [2024-12-06 17:54:30.094576] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.462 [2024-12-06 17:54:30.094602] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.462 [2024-12-06 17:54:30.094616] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.462 [2024-12-06 17:54:30.094629] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.462 [2024-12-06 17:54:30.094662] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.462 qpair failed and we were unable to recover it. 00:38:48.462 [2024-12-06 17:54:30.104478] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.462 [2024-12-06 17:54:30.104566] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.462 [2024-12-06 17:54:30.104591] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.462 [2024-12-06 17:54:30.104606] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.462 [2024-12-06 17:54:30.104619] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.462 [2024-12-06 17:54:30.104650] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.462 qpair failed and we were unable to recover it. 00:38:48.462 [2024-12-06 17:54:30.114524] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.462 [2024-12-06 17:54:30.114640] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.462 [2024-12-06 17:54:30.114674] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.462 [2024-12-06 17:54:30.114691] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.462 [2024-12-06 17:54:30.114703] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.462 [2024-12-06 17:54:30.114735] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.462 qpair failed and we were unable to recover it. 00:38:48.462 [2024-12-06 17:54:30.124588] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.462 [2024-12-06 17:54:30.124688] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.462 [2024-12-06 17:54:30.124714] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.462 [2024-12-06 17:54:30.124728] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.462 [2024-12-06 17:54:30.124741] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.462 [2024-12-06 17:54:30.124773] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.462 qpair failed and we were unable to recover it. 00:38:48.462 [2024-12-06 17:54:30.134603] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.462 [2024-12-06 17:54:30.134699] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.462 [2024-12-06 17:54:30.134725] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.462 [2024-12-06 17:54:30.134740] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.462 [2024-12-06 17:54:30.134753] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.462 [2024-12-06 17:54:30.134784] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.462 qpair failed and we were unable to recover it. 00:38:48.462 [2024-12-06 17:54:30.144606] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.462 [2024-12-06 17:54:30.144722] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.462 [2024-12-06 17:54:30.144752] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.462 [2024-12-06 17:54:30.144768] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.462 [2024-12-06 17:54:30.144781] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.462 [2024-12-06 17:54:30.144812] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.462 qpair failed and we were unable to recover it. 00:38:48.462 [2024-12-06 17:54:30.154624] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.462 [2024-12-06 17:54:30.154717] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.462 [2024-12-06 17:54:30.154743] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.462 [2024-12-06 17:54:30.154757] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.462 [2024-12-06 17:54:30.154770] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.462 [2024-12-06 17:54:30.154802] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.462 qpair failed and we were unable to recover it. 00:38:48.462 [2024-12-06 17:54:30.164659] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.462 [2024-12-06 17:54:30.164754] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.462 [2024-12-06 17:54:30.164780] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.462 [2024-12-06 17:54:30.164794] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.462 [2024-12-06 17:54:30.164807] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.462 [2024-12-06 17:54:30.164838] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.462 qpair failed and we were unable to recover it. 00:38:48.462 [2024-12-06 17:54:30.174763] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.462 [2024-12-06 17:54:30.174850] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.462 [2024-12-06 17:54:30.174883] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.462 [2024-12-06 17:54:30.174899] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.462 [2024-12-06 17:54:30.174912] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.462 [2024-12-06 17:54:30.174943] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.462 qpair failed and we were unable to recover it. 00:38:48.462 [2024-12-06 17:54:30.184710] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.462 [2024-12-06 17:54:30.184797] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.462 [2024-12-06 17:54:30.184826] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.462 [2024-12-06 17:54:30.184843] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.462 [2024-12-06 17:54:30.184855] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.462 [2024-12-06 17:54:30.184887] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.462 qpair failed and we were unable to recover it. 00:38:48.462 [2024-12-06 17:54:30.194722] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.462 [2024-12-06 17:54:30.194815] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.462 [2024-12-06 17:54:30.194840] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.463 [2024-12-06 17:54:30.194854] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.463 [2024-12-06 17:54:30.194867] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.463 [2024-12-06 17:54:30.194899] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.463 qpair failed and we were unable to recover it. 00:38:48.463 [2024-12-06 17:54:30.204807] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.463 [2024-12-06 17:54:30.204902] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.463 [2024-12-06 17:54:30.204927] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.463 [2024-12-06 17:54:30.204942] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.463 [2024-12-06 17:54:30.204955] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.463 [2024-12-06 17:54:30.204986] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.463 qpair failed and we were unable to recover it. 00:38:48.463 [2024-12-06 17:54:30.214831] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.463 [2024-12-06 17:54:30.214943] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.463 [2024-12-06 17:54:30.214970] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.463 [2024-12-06 17:54:30.214986] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.463 [2024-12-06 17:54:30.215005] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.463 [2024-12-06 17:54:30.215036] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.463 qpair failed and we were unable to recover it. 00:38:48.463 [2024-12-06 17:54:30.224805] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.463 [2024-12-06 17:54:30.224886] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.463 [2024-12-06 17:54:30.224911] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.463 [2024-12-06 17:54:30.224925] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.463 [2024-12-06 17:54:30.224938] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.463 [2024-12-06 17:54:30.224969] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.463 qpair failed and we were unable to recover it. 00:38:48.463 [2024-12-06 17:54:30.234857] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.463 [2024-12-06 17:54:30.234940] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.463 [2024-12-06 17:54:30.234965] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.463 [2024-12-06 17:54:30.234979] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.463 [2024-12-06 17:54:30.234992] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.463 [2024-12-06 17:54:30.235036] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.463 qpair failed and we were unable to recover it. 00:38:48.463 [2024-12-06 17:54:30.244886] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.463 [2024-12-06 17:54:30.244975] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.463 [2024-12-06 17:54:30.245001] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.463 [2024-12-06 17:54:30.245015] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.463 [2024-12-06 17:54:30.245028] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.463 [2024-12-06 17:54:30.245058] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.463 qpair failed and we were unable to recover it. 00:38:48.463 [2024-12-06 17:54:30.254915] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.463 [2024-12-06 17:54:30.255002] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.463 [2024-12-06 17:54:30.255027] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.463 [2024-12-06 17:54:30.255042] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.463 [2024-12-06 17:54:30.255055] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.463 [2024-12-06 17:54:30.255085] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.463 qpair failed and we were unable to recover it. 00:38:48.463 [2024-12-06 17:54:30.264985] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.463 [2024-12-06 17:54:30.265077] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.463 [2024-12-06 17:54:30.265101] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.463 [2024-12-06 17:54:30.265116] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.463 [2024-12-06 17:54:30.265129] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.463 [2024-12-06 17:54:30.265159] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.463 qpair failed and we were unable to recover it. 00:38:48.463 [2024-12-06 17:54:30.275004] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.463 [2024-12-06 17:54:30.275127] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.463 [2024-12-06 17:54:30.275153] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.463 [2024-12-06 17:54:30.275169] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.463 [2024-12-06 17:54:30.275182] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.463 [2024-12-06 17:54:30.275213] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.463 qpair failed and we were unable to recover it. 00:38:48.463 [2024-12-06 17:54:30.285011] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.463 [2024-12-06 17:54:30.285100] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.463 [2024-12-06 17:54:30.285125] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.463 [2024-12-06 17:54:30.285139] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.463 [2024-12-06 17:54:30.285153] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.463 [2024-12-06 17:54:30.285184] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.463 qpair failed and we were unable to recover it. 00:38:48.463 [2024-12-06 17:54:30.295077] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.463 [2024-12-06 17:54:30.295195] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.463 [2024-12-06 17:54:30.295221] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.463 [2024-12-06 17:54:30.295237] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.463 [2024-12-06 17:54:30.295249] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.463 [2024-12-06 17:54:30.295281] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.463 qpair failed and we were unable to recover it. 00:38:48.721 [2024-12-06 17:54:30.305098] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.721 [2024-12-06 17:54:30.305184] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.721 [2024-12-06 17:54:30.305215] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.721 [2024-12-06 17:54:30.305230] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.721 [2024-12-06 17:54:30.305244] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.721 [2024-12-06 17:54:30.305275] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.721 qpair failed and we were unable to recover it. 00:38:48.721 [2024-12-06 17:54:30.315096] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.721 [2024-12-06 17:54:30.315182] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.721 [2024-12-06 17:54:30.315214] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.721 [2024-12-06 17:54:30.315239] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.721 [2024-12-06 17:54:30.315261] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.721 [2024-12-06 17:54:30.315315] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.722 qpair failed and we were unable to recover it. 00:38:48.722 [2024-12-06 17:54:30.325173] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.722 [2024-12-06 17:54:30.325294] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.722 [2024-12-06 17:54:30.325322] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.722 [2024-12-06 17:54:30.325337] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.722 [2024-12-06 17:54:30.325350] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.722 [2024-12-06 17:54:30.325381] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.722 qpair failed and we were unable to recover it. 00:38:48.722 [2024-12-06 17:54:30.335205] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.722 [2024-12-06 17:54:30.335299] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.722 [2024-12-06 17:54:30.335324] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.722 [2024-12-06 17:54:30.335339] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.722 [2024-12-06 17:54:30.335352] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.722 [2024-12-06 17:54:30.335383] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.722 qpair failed and we were unable to recover it. 00:38:48.722 [2024-12-06 17:54:30.345263] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.722 [2024-12-06 17:54:30.345356] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.722 [2024-12-06 17:54:30.345381] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.722 [2024-12-06 17:54:30.345395] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.722 [2024-12-06 17:54:30.345414] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.722 [2024-12-06 17:54:30.345445] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.722 qpair failed and we were unable to recover it. 00:38:48.722 [2024-12-06 17:54:30.355235] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.722 [2024-12-06 17:54:30.355366] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.722 [2024-12-06 17:54:30.355392] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.722 [2024-12-06 17:54:30.355407] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.722 [2024-12-06 17:54:30.355419] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.722 [2024-12-06 17:54:30.355450] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.722 qpair failed and we were unable to recover it. 00:38:48.722 [2024-12-06 17:54:30.365241] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.722 [2024-12-06 17:54:30.365332] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.722 [2024-12-06 17:54:30.365357] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.722 [2024-12-06 17:54:30.365371] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.722 [2024-12-06 17:54:30.365384] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.722 [2024-12-06 17:54:30.365415] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.722 qpair failed and we were unable to recover it. 00:38:48.722 [2024-12-06 17:54:30.375267] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.722 [2024-12-06 17:54:30.375354] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.722 [2024-12-06 17:54:30.375380] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.722 [2024-12-06 17:54:30.375395] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.722 [2024-12-06 17:54:30.375408] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.722 [2024-12-06 17:54:30.375450] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.722 qpair failed and we were unable to recover it. 00:38:48.722 [2024-12-06 17:54:30.385309] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.722 [2024-12-06 17:54:30.385407] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.722 [2024-12-06 17:54:30.385433] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.722 [2024-12-06 17:54:30.385448] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.722 [2024-12-06 17:54:30.385462] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.722 [2024-12-06 17:54:30.385492] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.722 qpair failed and we were unable to recover it. 00:38:48.722 [2024-12-06 17:54:30.395312] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.722 [2024-12-06 17:54:30.395397] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.722 [2024-12-06 17:54:30.395422] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.722 [2024-12-06 17:54:30.395437] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.722 [2024-12-06 17:54:30.395450] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.722 [2024-12-06 17:54:30.395480] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.722 qpair failed and we were unable to recover it. 00:38:48.722 [2024-12-06 17:54:30.405353] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.722 [2024-12-06 17:54:30.405445] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.722 [2024-12-06 17:54:30.405470] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.722 [2024-12-06 17:54:30.405485] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.722 [2024-12-06 17:54:30.405498] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.722 [2024-12-06 17:54:30.405528] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.722 qpair failed and we were unable to recover it. 00:38:48.722 [2024-12-06 17:54:30.415402] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.722 [2024-12-06 17:54:30.415487] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.722 [2024-12-06 17:54:30.415515] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.722 [2024-12-06 17:54:30.415532] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.722 [2024-12-06 17:54:30.415546] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.722 [2024-12-06 17:54:30.415577] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.723 qpair failed and we were unable to recover it. 00:38:48.723 [2024-12-06 17:54:30.425388] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.723 [2024-12-06 17:54:30.425515] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.723 [2024-12-06 17:54:30.425543] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.723 [2024-12-06 17:54:30.425558] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.723 [2024-12-06 17:54:30.425571] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.723 [2024-12-06 17:54:30.425602] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.723 qpair failed and we were unable to recover it. 00:38:48.723 [2024-12-06 17:54:30.435445] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.723 [2024-12-06 17:54:30.435526] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.723 [2024-12-06 17:54:30.435557] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.723 [2024-12-06 17:54:30.435572] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.723 [2024-12-06 17:54:30.435585] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.723 [2024-12-06 17:54:30.435617] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.723 qpair failed and we were unable to recover it. 00:38:48.723 [2024-12-06 17:54:30.445488] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.723 [2024-12-06 17:54:30.445597] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.723 [2024-12-06 17:54:30.445623] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.723 [2024-12-06 17:54:30.445638] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.723 [2024-12-06 17:54:30.445651] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.723 [2024-12-06 17:54:30.445690] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.723 qpair failed and we were unable to recover it. 00:38:48.723 [2024-12-06 17:54:30.455520] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.723 [2024-12-06 17:54:30.455653] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.723 [2024-12-06 17:54:30.455689] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.723 [2024-12-06 17:54:30.455705] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.723 [2024-12-06 17:54:30.455718] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.723 [2024-12-06 17:54:30.455749] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.723 qpair failed and we were unable to recover it. 00:38:48.723 [2024-12-06 17:54:30.465499] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.723 [2024-12-06 17:54:30.465585] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.723 [2024-12-06 17:54:30.465610] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.723 [2024-12-06 17:54:30.465624] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.723 [2024-12-06 17:54:30.465637] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.723 [2024-12-06 17:54:30.465674] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.723 qpair failed and we were unable to recover it. 00:38:48.723 [2024-12-06 17:54:30.475607] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.723 [2024-12-06 17:54:30.475698] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.723 [2024-12-06 17:54:30.475724] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.723 [2024-12-06 17:54:30.475744] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.723 [2024-12-06 17:54:30.475758] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.723 [2024-12-06 17:54:30.475789] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.723 qpair failed and we were unable to recover it. 00:38:48.723 [2024-12-06 17:54:30.485599] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.723 [2024-12-06 17:54:30.485705] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.723 [2024-12-06 17:54:30.485730] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.723 [2024-12-06 17:54:30.485745] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.723 [2024-12-06 17:54:30.485757] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.723 [2024-12-06 17:54:30.485788] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.723 qpair failed and we were unable to recover it. 00:38:48.723 [2024-12-06 17:54:30.495600] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.723 [2024-12-06 17:54:30.495691] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.723 [2024-12-06 17:54:30.495716] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.723 [2024-12-06 17:54:30.495730] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.723 [2024-12-06 17:54:30.495743] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.723 [2024-12-06 17:54:30.495774] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.723 qpair failed and we were unable to recover it. 00:38:48.723 [2024-12-06 17:54:30.505650] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.723 [2024-12-06 17:54:30.505742] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.723 [2024-12-06 17:54:30.505767] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.723 [2024-12-06 17:54:30.505782] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.723 [2024-12-06 17:54:30.505795] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.723 [2024-12-06 17:54:30.505826] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.723 qpair failed and we were unable to recover it. 00:38:48.723 [2024-12-06 17:54:30.515705] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.723 [2024-12-06 17:54:30.515805] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.723 [2024-12-06 17:54:30.515830] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.723 [2024-12-06 17:54:30.515845] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.723 [2024-12-06 17:54:30.515858] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.723 [2024-12-06 17:54:30.515889] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.723 qpair failed and we were unable to recover it. 00:38:48.723 [2024-12-06 17:54:30.525718] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.724 [2024-12-06 17:54:30.525814] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.724 [2024-12-06 17:54:30.525843] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.724 [2024-12-06 17:54:30.525859] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.724 [2024-12-06 17:54:30.525872] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.724 [2024-12-06 17:54:30.525904] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.724 qpair failed and we were unable to recover it. 00:38:48.724 [2024-12-06 17:54:30.535715] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.724 [2024-12-06 17:54:30.535806] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.724 [2024-12-06 17:54:30.535832] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.724 [2024-12-06 17:54:30.535847] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.724 [2024-12-06 17:54:30.535860] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.724 [2024-12-06 17:54:30.535890] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.724 qpair failed and we were unable to recover it. 00:38:48.724 [2024-12-06 17:54:30.545760] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.724 [2024-12-06 17:54:30.545845] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.724 [2024-12-06 17:54:30.545870] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.724 [2024-12-06 17:54:30.545886] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.724 [2024-12-06 17:54:30.545899] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.724 [2024-12-06 17:54:30.545929] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.724 qpair failed and we were unable to recover it. 00:38:48.724 [2024-12-06 17:54:30.555857] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.724 [2024-12-06 17:54:30.555971] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.724 [2024-12-06 17:54:30.556023] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.724 [2024-12-06 17:54:30.556047] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.724 [2024-12-06 17:54:30.556063] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.724 [2024-12-06 17:54:30.556109] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.724 qpair failed and we were unable to recover it. 00:38:48.983 [2024-12-06 17:54:30.565834] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.983 [2024-12-06 17:54:30.565931] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.983 [2024-12-06 17:54:30.565964] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.983 [2024-12-06 17:54:30.565988] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.983 [2024-12-06 17:54:30.566011] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.983 [2024-12-06 17:54:30.566046] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.983 qpair failed and we were unable to recover it. 00:38:48.983 [2024-12-06 17:54:30.575928] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.983 [2024-12-06 17:54:30.576025] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.983 [2024-12-06 17:54:30.576051] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.983 [2024-12-06 17:54:30.576065] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.983 [2024-12-06 17:54:30.576078] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.983 [2024-12-06 17:54:30.576109] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.983 qpair failed and we were unable to recover it. 00:38:48.983 [2024-12-06 17:54:30.585925] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.983 [2024-12-06 17:54:30.586048] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.983 [2024-12-06 17:54:30.586076] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.983 [2024-12-06 17:54:30.586092] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.983 [2024-12-06 17:54:30.586105] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.983 [2024-12-06 17:54:30.586136] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.983 qpair failed and we were unable to recover it. 00:38:48.983 [2024-12-06 17:54:30.595898] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.983 [2024-12-06 17:54:30.595988] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.983 [2024-12-06 17:54:30.596014] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.983 [2024-12-06 17:54:30.596028] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.983 [2024-12-06 17:54:30.596042] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.983 [2024-12-06 17:54:30.596072] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.983 qpair failed and we were unable to recover it. 00:38:48.983 [2024-12-06 17:54:30.605956] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.983 [2024-12-06 17:54:30.606059] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.983 [2024-12-06 17:54:30.606084] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.983 [2024-12-06 17:54:30.606104] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.983 [2024-12-06 17:54:30.606118] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.983 [2024-12-06 17:54:30.606149] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.983 qpair failed and we were unable to recover it. 00:38:48.983 [2024-12-06 17:54:30.615987] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.983 [2024-12-06 17:54:30.616071] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.983 [2024-12-06 17:54:30.616096] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.983 [2024-12-06 17:54:30.616111] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.983 [2024-12-06 17:54:30.616124] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.983 [2024-12-06 17:54:30.616154] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.983 qpair failed and we were unable to recover it. 00:38:48.983 [2024-12-06 17:54:30.625979] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.983 [2024-12-06 17:54:30.626069] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.983 [2024-12-06 17:54:30.626094] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.983 [2024-12-06 17:54:30.626108] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.983 [2024-12-06 17:54:30.626121] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.983 [2024-12-06 17:54:30.626164] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.983 qpair failed and we were unable to recover it. 00:38:48.984 [2024-12-06 17:54:30.636014] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.984 [2024-12-06 17:54:30.636100] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.984 [2024-12-06 17:54:30.636125] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.984 [2024-12-06 17:54:30.636139] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.984 [2024-12-06 17:54:30.636152] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.984 [2024-12-06 17:54:30.636183] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.984 qpair failed and we were unable to recover it. 00:38:48.984 [2024-12-06 17:54:30.646027] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.984 [2024-12-06 17:54:30.646124] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.984 [2024-12-06 17:54:30.646151] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.984 [2024-12-06 17:54:30.646166] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.984 [2024-12-06 17:54:30.646179] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.984 [2024-12-06 17:54:30.646215] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.984 qpair failed and we were unable to recover it. 00:38:48.984 [2024-12-06 17:54:30.656125] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.984 [2024-12-06 17:54:30.656225] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.984 [2024-12-06 17:54:30.656255] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.984 [2024-12-06 17:54:30.656272] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.984 [2024-12-06 17:54:30.656285] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.984 [2024-12-06 17:54:30.656316] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.984 qpair failed and we were unable to recover it. 00:38:48.984 [2024-12-06 17:54:30.666095] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.984 [2024-12-06 17:54:30.666182] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.984 [2024-12-06 17:54:30.666208] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.984 [2024-12-06 17:54:30.666223] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.984 [2024-12-06 17:54:30.666235] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.984 [2024-12-06 17:54:30.666265] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.984 qpair failed and we were unable to recover it. 00:38:48.984 [2024-12-06 17:54:30.676104] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.984 [2024-12-06 17:54:30.676190] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.984 [2024-12-06 17:54:30.676214] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.984 [2024-12-06 17:54:30.676229] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.984 [2024-12-06 17:54:30.676242] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.984 [2024-12-06 17:54:30.676272] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.984 qpair failed and we were unable to recover it. 00:38:48.984 [2024-12-06 17:54:30.686230] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.984 [2024-12-06 17:54:30.686325] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.984 [2024-12-06 17:54:30.686352] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.984 [2024-12-06 17:54:30.686368] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.984 [2024-12-06 17:54:30.686381] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.984 [2024-12-06 17:54:30.686411] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.984 qpair failed and we were unable to recover it. 00:38:48.984 [2024-12-06 17:54:30.696202] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.984 [2024-12-06 17:54:30.696324] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.984 [2024-12-06 17:54:30.696351] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.984 [2024-12-06 17:54:30.696366] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.984 [2024-12-06 17:54:30.696379] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.984 [2024-12-06 17:54:30.696410] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.984 qpair failed and we were unable to recover it. 00:38:48.984 [2024-12-06 17:54:30.706218] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.984 [2024-12-06 17:54:30.706301] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.984 [2024-12-06 17:54:30.706325] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.984 [2024-12-06 17:54:30.706339] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.984 [2024-12-06 17:54:30.706352] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.984 [2024-12-06 17:54:30.706383] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.984 qpair failed and we were unable to recover it. 00:38:48.984 [2024-12-06 17:54:30.716276] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.984 [2024-12-06 17:54:30.716377] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.984 [2024-12-06 17:54:30.716403] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.984 [2024-12-06 17:54:30.716418] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.984 [2024-12-06 17:54:30.716432] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.984 [2024-12-06 17:54:30.716462] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.984 qpair failed and we were unable to recover it. 00:38:48.984 [2024-12-06 17:54:30.726263] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.984 [2024-12-06 17:54:30.726355] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.984 [2024-12-06 17:54:30.726380] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.984 [2024-12-06 17:54:30.726395] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.984 [2024-12-06 17:54:30.726408] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.984 [2024-12-06 17:54:30.726450] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.984 qpair failed and we were unable to recover it. 00:38:48.984 [2024-12-06 17:54:30.736302] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.984 [2024-12-06 17:54:30.736392] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.984 [2024-12-06 17:54:30.736422] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.985 [2024-12-06 17:54:30.736438] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.985 [2024-12-06 17:54:30.736451] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.985 [2024-12-06 17:54:30.736481] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.985 qpair failed and we were unable to recover it. 00:38:48.985 [2024-12-06 17:54:30.746304] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.985 [2024-12-06 17:54:30.746395] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.985 [2024-12-06 17:54:30.746420] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.985 [2024-12-06 17:54:30.746435] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.985 [2024-12-06 17:54:30.746447] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.985 [2024-12-06 17:54:30.746478] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.985 qpair failed and we were unable to recover it. 00:38:48.985 [2024-12-06 17:54:30.756353] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.985 [2024-12-06 17:54:30.756445] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.985 [2024-12-06 17:54:30.756470] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.985 [2024-12-06 17:54:30.756485] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.985 [2024-12-06 17:54:30.756498] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.985 [2024-12-06 17:54:30.756529] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.985 qpair failed and we were unable to recover it. 00:38:48.985 [2024-12-06 17:54:30.766470] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.985 [2024-12-06 17:54:30.766568] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.985 [2024-12-06 17:54:30.766593] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.985 [2024-12-06 17:54:30.766608] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.985 [2024-12-06 17:54:30.766621] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.985 [2024-12-06 17:54:30.766662] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.985 qpair failed and we were unable to recover it. 00:38:48.985 [2024-12-06 17:54:30.776378] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.985 [2024-12-06 17:54:30.776480] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.985 [2024-12-06 17:54:30.776507] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.985 [2024-12-06 17:54:30.776522] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.985 [2024-12-06 17:54:30.776541] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.985 [2024-12-06 17:54:30.776573] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.985 qpair failed and we were unable to recover it. 00:38:48.985 [2024-12-06 17:54:30.786415] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.985 [2024-12-06 17:54:30.786503] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.985 [2024-12-06 17:54:30.786529] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.985 [2024-12-06 17:54:30.786543] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.985 [2024-12-06 17:54:30.786557] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.985 [2024-12-06 17:54:30.786588] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.985 qpair failed and we were unable to recover it. 00:38:48.985 [2024-12-06 17:54:30.796469] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.985 [2024-12-06 17:54:30.796561] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.985 [2024-12-06 17:54:30.796586] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.985 [2024-12-06 17:54:30.796601] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.985 [2024-12-06 17:54:30.796613] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.985 [2024-12-06 17:54:30.796644] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.985 qpair failed and we were unable to recover it. 00:38:48.985 [2024-12-06 17:54:30.806549] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.985 [2024-12-06 17:54:30.806644] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.985 [2024-12-06 17:54:30.806679] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.985 [2024-12-06 17:54:30.806696] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.985 [2024-12-06 17:54:30.806708] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.985 [2024-12-06 17:54:30.806739] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.985 qpair failed and we were unable to recover it. 00:38:48.985 [2024-12-06 17:54:30.816549] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:48.985 [2024-12-06 17:54:30.816637] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:48.985 [2024-12-06 17:54:30.816675] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:48.985 [2024-12-06 17:54:30.816702] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:48.985 [2024-12-06 17:54:30.816725] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:48.985 [2024-12-06 17:54:30.816779] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:48.985 qpair failed and we were unable to recover it. 00:38:49.244 [2024-12-06 17:54:30.826571] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.244 [2024-12-06 17:54:30.826657] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.244 [2024-12-06 17:54:30.826692] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.244 [2024-12-06 17:54:30.826708] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.244 [2024-12-06 17:54:30.826721] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.244 [2024-12-06 17:54:30.826753] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.244 qpair failed and we were unable to recover it. 00:38:49.244 [2024-12-06 17:54:30.836567] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.244 [2024-12-06 17:54:30.836659] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.244 [2024-12-06 17:54:30.836706] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.244 [2024-12-06 17:54:30.836721] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.244 [2024-12-06 17:54:30.836734] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.244 [2024-12-06 17:54:30.836766] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.244 qpair failed and we were unable to recover it. 00:38:49.244 [2024-12-06 17:54:30.846604] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.244 [2024-12-06 17:54:30.846710] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.244 [2024-12-06 17:54:30.846737] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.244 [2024-12-06 17:54:30.846752] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.244 [2024-12-06 17:54:30.846765] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.244 [2024-12-06 17:54:30.846795] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.244 qpair failed and we were unable to recover it. 00:38:49.244 [2024-12-06 17:54:30.856718] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.244 [2024-12-06 17:54:30.856843] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.244 [2024-12-06 17:54:30.856870] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.244 [2024-12-06 17:54:30.856886] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.244 [2024-12-06 17:54:30.856899] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.244 [2024-12-06 17:54:30.856930] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.244 qpair failed and we were unable to recover it. 00:38:49.244 [2024-12-06 17:54:30.866634] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.244 [2024-12-06 17:54:30.866734] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.244 [2024-12-06 17:54:30.866765] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.244 [2024-12-06 17:54:30.866780] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.244 [2024-12-06 17:54:30.866793] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.244 [2024-12-06 17:54:30.866824] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.244 qpair failed and we were unable to recover it. 00:38:49.244 [2024-12-06 17:54:30.876655] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.244 [2024-12-06 17:54:30.876756] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.244 [2024-12-06 17:54:30.876782] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.244 [2024-12-06 17:54:30.876796] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.244 [2024-12-06 17:54:30.876808] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.244 [2024-12-06 17:54:30.876839] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.244 qpair failed and we were unable to recover it. 00:38:49.244 [2024-12-06 17:54:30.886767] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.244 [2024-12-06 17:54:30.886874] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.244 [2024-12-06 17:54:30.886900] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.244 [2024-12-06 17:54:30.886915] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.244 [2024-12-06 17:54:30.886928] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.244 [2024-12-06 17:54:30.886970] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.244 qpair failed and we were unable to recover it. 00:38:49.244 [2024-12-06 17:54:30.896772] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.245 [2024-12-06 17:54:30.896906] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.245 [2024-12-06 17:54:30.896935] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.245 [2024-12-06 17:54:30.896959] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.245 [2024-12-06 17:54:30.896972] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.245 [2024-12-06 17:54:30.897003] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.245 qpair failed and we were unable to recover it. 00:38:49.245 [2024-12-06 17:54:30.906753] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.245 [2024-12-06 17:54:30.906841] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.245 [2024-12-06 17:54:30.906867] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.245 [2024-12-06 17:54:30.906881] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.245 [2024-12-06 17:54:30.906900] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.245 [2024-12-06 17:54:30.906931] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.245 qpair failed and we were unable to recover it. 00:38:49.245 [2024-12-06 17:54:30.916796] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.245 [2024-12-06 17:54:30.916881] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.245 [2024-12-06 17:54:30.916906] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.245 [2024-12-06 17:54:30.916920] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.245 [2024-12-06 17:54:30.916933] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.245 [2024-12-06 17:54:30.916967] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.245 qpair failed and we were unable to recover it. 00:38:49.245 [2024-12-06 17:54:30.926859] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.245 [2024-12-06 17:54:30.926970] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.245 [2024-12-06 17:54:30.926995] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.245 [2024-12-06 17:54:30.927009] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.245 [2024-12-06 17:54:30.927021] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.245 [2024-12-06 17:54:30.927053] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.245 qpair failed and we were unable to recover it. 00:38:49.245 [2024-12-06 17:54:30.936868] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.245 [2024-12-06 17:54:30.936986] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.245 [2024-12-06 17:54:30.937013] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.245 [2024-12-06 17:54:30.937028] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.245 [2024-12-06 17:54:30.937040] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.245 [2024-12-06 17:54:30.937084] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.245 qpair failed and we were unable to recover it. 00:38:49.245 [2024-12-06 17:54:30.946896] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.245 [2024-12-06 17:54:30.946981] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.245 [2024-12-06 17:54:30.947006] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.245 [2024-12-06 17:54:30.947021] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.245 [2024-12-06 17:54:30.947033] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.245 [2024-12-06 17:54:30.947064] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.245 qpair failed and we were unable to recover it. 00:38:49.245 [2024-12-06 17:54:30.956931] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.245 [2024-12-06 17:54:30.957054] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.245 [2024-12-06 17:54:30.957082] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.245 [2024-12-06 17:54:30.957098] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.245 [2024-12-06 17:54:30.957111] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.245 [2024-12-06 17:54:30.957154] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.245 qpair failed and we were unable to recover it. 00:38:49.245 [2024-12-06 17:54:30.966948] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.245 [2024-12-06 17:54:30.967036] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.245 [2024-12-06 17:54:30.967060] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.245 [2024-12-06 17:54:30.967075] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.245 [2024-12-06 17:54:30.967087] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.245 [2024-12-06 17:54:30.967117] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.245 qpair failed and we were unable to recover it. 00:38:49.245 [2024-12-06 17:54:30.977037] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.245 [2024-12-06 17:54:30.977130] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.245 [2024-12-06 17:54:30.977156] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.245 [2024-12-06 17:54:30.977175] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.245 [2024-12-06 17:54:30.977189] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.245 [2024-12-06 17:54:30.977221] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.245 qpair failed and we were unable to recover it. 00:38:49.245 [2024-12-06 17:54:30.987025] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.245 [2024-12-06 17:54:30.987109] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.245 [2024-12-06 17:54:30.987134] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.245 [2024-12-06 17:54:30.987149] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.245 [2024-12-06 17:54:30.987162] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.245 [2024-12-06 17:54:30.987193] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.245 qpair failed and we were unable to recover it. 00:38:49.245 [2024-12-06 17:54:30.997050] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.245 [2024-12-06 17:54:30.997150] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.246 [2024-12-06 17:54:30.997182] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.246 [2024-12-06 17:54:30.997199] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.246 [2024-12-06 17:54:30.997212] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.246 [2024-12-06 17:54:30.997242] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.246 qpair failed and we were unable to recover it. 00:38:49.246 [2024-12-06 17:54:31.007137] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.246 [2024-12-06 17:54:31.007242] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.246 [2024-12-06 17:54:31.007269] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.246 [2024-12-06 17:54:31.007284] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.246 [2024-12-06 17:54:31.007297] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.246 [2024-12-06 17:54:31.007328] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.246 qpair failed and we were unable to recover it. 00:38:49.246 [2024-12-06 17:54:31.017072] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.246 [2024-12-06 17:54:31.017180] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.246 [2024-12-06 17:54:31.017206] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.246 [2024-12-06 17:54:31.017221] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.246 [2024-12-06 17:54:31.017233] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.246 [2024-12-06 17:54:31.017263] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.246 qpair failed and we were unable to recover it. 00:38:49.246 [2024-12-06 17:54:31.027140] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.246 [2024-12-06 17:54:31.027227] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.246 [2024-12-06 17:54:31.027254] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.246 [2024-12-06 17:54:31.027270] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.246 [2024-12-06 17:54:31.027283] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.246 [2024-12-06 17:54:31.027314] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.246 qpair failed and we were unable to recover it. 00:38:49.246 [2024-12-06 17:54:31.037218] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.246 [2024-12-06 17:54:31.037321] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.246 [2024-12-06 17:54:31.037349] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.246 [2024-12-06 17:54:31.037372] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.246 [2024-12-06 17:54:31.037386] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.246 [2024-12-06 17:54:31.037417] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.246 qpair failed and we were unable to recover it. 00:38:49.246 [2024-12-06 17:54:31.047199] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.246 [2024-12-06 17:54:31.047298] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.246 [2024-12-06 17:54:31.047323] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.246 [2024-12-06 17:54:31.047338] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.246 [2024-12-06 17:54:31.047350] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.246 [2024-12-06 17:54:31.047381] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.246 qpair failed and we were unable to recover it. 00:38:49.246 [2024-12-06 17:54:31.057204] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.246 [2024-12-06 17:54:31.057293] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.246 [2024-12-06 17:54:31.057318] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.246 [2024-12-06 17:54:31.057332] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.246 [2024-12-06 17:54:31.057345] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.246 [2024-12-06 17:54:31.057376] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.246 qpair failed and we were unable to recover it. 00:38:49.246 [2024-12-06 17:54:31.067248] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.246 [2024-12-06 17:54:31.067342] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.246 [2024-12-06 17:54:31.067378] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.246 [2024-12-06 17:54:31.067402] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.246 [2024-12-06 17:54:31.067424] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.246 [2024-12-06 17:54:31.067457] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.246 qpair failed and we were unable to recover it. 00:38:49.246 [2024-12-06 17:54:31.077273] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.246 [2024-12-06 17:54:31.077368] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.246 [2024-12-06 17:54:31.077395] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.246 [2024-12-06 17:54:31.077410] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.246 [2024-12-06 17:54:31.077424] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.246 [2024-12-06 17:54:31.077461] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.246 qpair failed and we were unable to recover it. 00:38:49.505 [2024-12-06 17:54:31.087319] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.505 [2024-12-06 17:54:31.087411] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.505 [2024-12-06 17:54:31.087439] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.505 [2024-12-06 17:54:31.087454] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.505 [2024-12-06 17:54:31.087467] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.505 [2024-12-06 17:54:31.087499] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.505 qpair failed and we were unable to recover it. 00:38:49.505 [2024-12-06 17:54:31.097304] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.505 [2024-12-06 17:54:31.097399] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.505 [2024-12-06 17:54:31.097424] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.505 [2024-12-06 17:54:31.097439] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.505 [2024-12-06 17:54:31.097452] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.505 [2024-12-06 17:54:31.097483] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.505 qpair failed and we were unable to recover it. 00:38:49.505 [2024-12-06 17:54:31.107335] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.505 [2024-12-06 17:54:31.107423] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.505 [2024-12-06 17:54:31.107448] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.505 [2024-12-06 17:54:31.107463] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.505 [2024-12-06 17:54:31.107476] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.505 [2024-12-06 17:54:31.107507] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.505 qpair failed and we were unable to recover it. 00:38:49.505 [2024-12-06 17:54:31.117363] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.505 [2024-12-06 17:54:31.117453] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.505 [2024-12-06 17:54:31.117478] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.505 [2024-12-06 17:54:31.117493] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.505 [2024-12-06 17:54:31.117506] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.505 [2024-12-06 17:54:31.117550] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.505 qpair failed and we were unable to recover it. 00:38:49.505 [2024-12-06 17:54:31.127486] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.505 [2024-12-06 17:54:31.127581] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.505 [2024-12-06 17:54:31.127606] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.505 [2024-12-06 17:54:31.127621] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.505 [2024-12-06 17:54:31.127635] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.505 [2024-12-06 17:54:31.127673] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.505 qpair failed and we were unable to recover it. 00:38:49.505 [2024-12-06 17:54:31.137442] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.505 [2024-12-06 17:54:31.137544] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.505 [2024-12-06 17:54:31.137571] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.505 [2024-12-06 17:54:31.137587] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.505 [2024-12-06 17:54:31.137600] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.505 [2024-12-06 17:54:31.137631] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.505 qpair failed and we were unable to recover it. 00:38:49.505 [2024-12-06 17:54:31.147449] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.505 [2024-12-06 17:54:31.147542] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.505 [2024-12-06 17:54:31.147568] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.505 [2024-12-06 17:54:31.147583] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.505 [2024-12-06 17:54:31.147597] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.505 [2024-12-06 17:54:31.147628] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.505 qpair failed and we were unable to recover it. 00:38:49.505 [2024-12-06 17:54:31.157522] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.505 [2024-12-06 17:54:31.157633] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.505 [2024-12-06 17:54:31.157661] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.505 [2024-12-06 17:54:31.157685] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.505 [2024-12-06 17:54:31.157699] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.505 [2024-12-06 17:54:31.157743] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.505 qpair failed and we were unable to recover it. 00:38:49.505 [2024-12-06 17:54:31.167601] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.505 [2024-12-06 17:54:31.167746] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.505 [2024-12-06 17:54:31.167774] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.505 [2024-12-06 17:54:31.167795] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.505 [2024-12-06 17:54:31.167809] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.505 [2024-12-06 17:54:31.167840] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.505 qpair failed and we were unable to recover it. 00:38:49.505 [2024-12-06 17:54:31.177569] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.505 [2024-12-06 17:54:31.177662] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.505 [2024-12-06 17:54:31.177694] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.505 [2024-12-06 17:54:31.177709] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.506 [2024-12-06 17:54:31.177722] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.506 [2024-12-06 17:54:31.177753] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.506 qpair failed and we were unable to recover it. 00:38:49.506 [2024-12-06 17:54:31.187590] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.506 [2024-12-06 17:54:31.187711] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.506 [2024-12-06 17:54:31.187739] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.506 [2024-12-06 17:54:31.187755] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.506 [2024-12-06 17:54:31.187767] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.506 [2024-12-06 17:54:31.187798] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.506 qpair failed and we were unable to recover it. 00:38:49.506 [2024-12-06 17:54:31.197577] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.506 [2024-12-06 17:54:31.197686] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.506 [2024-12-06 17:54:31.197713] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.506 [2024-12-06 17:54:31.197728] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.506 [2024-12-06 17:54:31.197741] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.506 [2024-12-06 17:54:31.197772] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.506 qpair failed and we were unable to recover it. 00:38:49.506 [2024-12-06 17:54:31.207658] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.506 [2024-12-06 17:54:31.207758] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.506 [2024-12-06 17:54:31.207782] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.506 [2024-12-06 17:54:31.207797] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.506 [2024-12-06 17:54:31.207810] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.506 [2024-12-06 17:54:31.207846] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.506 qpair failed and we were unable to recover it. 00:38:49.506 [2024-12-06 17:54:31.217652] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.506 [2024-12-06 17:54:31.217745] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.506 [2024-12-06 17:54:31.217771] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.506 [2024-12-06 17:54:31.217786] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.506 [2024-12-06 17:54:31.217799] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.506 [2024-12-06 17:54:31.217830] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.506 qpair failed and we were unable to recover it. 00:38:49.506 [2024-12-06 17:54:31.227704] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.506 [2024-12-06 17:54:31.227831] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.506 [2024-12-06 17:54:31.227858] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.506 [2024-12-06 17:54:31.227873] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.506 [2024-12-06 17:54:31.227886] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.506 [2024-12-06 17:54:31.227917] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.506 qpair failed and we were unable to recover it. 00:38:49.506 [2024-12-06 17:54:31.237723] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.506 [2024-12-06 17:54:31.237811] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.506 [2024-12-06 17:54:31.237835] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.506 [2024-12-06 17:54:31.237849] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.506 [2024-12-06 17:54:31.237862] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.506 [2024-12-06 17:54:31.237893] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.506 qpair failed and we were unable to recover it. 00:38:49.506 [2024-12-06 17:54:31.247757] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.506 [2024-12-06 17:54:31.247852] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.506 [2024-12-06 17:54:31.247878] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.506 [2024-12-06 17:54:31.247893] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.506 [2024-12-06 17:54:31.247906] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.506 [2024-12-06 17:54:31.247948] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.506 qpair failed and we were unable to recover it. 00:38:49.506 [2024-12-06 17:54:31.257796] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.506 [2024-12-06 17:54:31.257886] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.506 [2024-12-06 17:54:31.257911] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.506 [2024-12-06 17:54:31.257926] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.506 [2024-12-06 17:54:31.257939] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.506 [2024-12-06 17:54:31.257970] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.506 qpair failed and we were unable to recover it. 00:38:49.506 [2024-12-06 17:54:31.267833] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.506 [2024-12-06 17:54:31.267953] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.506 [2024-12-06 17:54:31.267978] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.506 [2024-12-06 17:54:31.267993] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.506 [2024-12-06 17:54:31.268006] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.506 [2024-12-06 17:54:31.268037] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.506 qpair failed and we were unable to recover it. 00:38:49.506 [2024-12-06 17:54:31.277859] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.506 [2024-12-06 17:54:31.277946] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.506 [2024-12-06 17:54:31.277974] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.506 [2024-12-06 17:54:31.277989] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.506 [2024-12-06 17:54:31.278002] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.506 [2024-12-06 17:54:31.278033] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.506 qpair failed and we were unable to recover it. 00:38:49.507 [2024-12-06 17:54:31.287857] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.507 [2024-12-06 17:54:31.287948] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.507 [2024-12-06 17:54:31.287983] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.507 [2024-12-06 17:54:31.287998] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.507 [2024-12-06 17:54:31.288011] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.507 [2024-12-06 17:54:31.288042] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.507 qpair failed and we were unable to recover it. 00:38:49.507 [2024-12-06 17:54:31.297935] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.507 [2024-12-06 17:54:31.298025] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.507 [2024-12-06 17:54:31.298056] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.507 [2024-12-06 17:54:31.298072] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.507 [2024-12-06 17:54:31.298085] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.507 [2024-12-06 17:54:31.298116] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.507 qpair failed and we were unable to recover it. 00:38:49.507 [2024-12-06 17:54:31.307943] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.507 [2024-12-06 17:54:31.308074] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.507 [2024-12-06 17:54:31.308099] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.507 [2024-12-06 17:54:31.308115] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.507 [2024-12-06 17:54:31.308127] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.507 [2024-12-06 17:54:31.308158] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.507 qpair failed and we were unable to recover it. 00:38:49.507 [2024-12-06 17:54:31.317959] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.507 [2024-12-06 17:54:31.318039] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.507 [2024-12-06 17:54:31.318069] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.507 [2024-12-06 17:54:31.318094] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.507 [2024-12-06 17:54:31.318116] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.507 [2024-12-06 17:54:31.318157] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.507 qpair failed and we were unable to recover it. 00:38:49.507 [2024-12-06 17:54:31.328004] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.507 [2024-12-06 17:54:31.328099] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.507 [2024-12-06 17:54:31.328125] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.507 [2024-12-06 17:54:31.328141] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.507 [2024-12-06 17:54:31.328154] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.507 [2024-12-06 17:54:31.328185] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.507 qpair failed and we were unable to recover it. 00:38:49.507 [2024-12-06 17:54:31.338001] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.507 [2024-12-06 17:54:31.338090] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.507 [2024-12-06 17:54:31.338116] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.507 [2024-12-06 17:54:31.338132] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.507 [2024-12-06 17:54:31.338152] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.507 [2024-12-06 17:54:31.338184] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.507 qpair failed and we were unable to recover it. 00:38:49.766 [2024-12-06 17:54:31.348005] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.766 [2024-12-06 17:54:31.348095] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.766 [2024-12-06 17:54:31.348121] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.766 [2024-12-06 17:54:31.348136] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.766 [2024-12-06 17:54:31.348149] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.766 [2024-12-06 17:54:31.348180] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.766 qpair failed and we were unable to recover it. 00:38:49.766 [2024-12-06 17:54:31.358054] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.766 [2024-12-06 17:54:31.358179] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.766 [2024-12-06 17:54:31.358204] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.766 [2024-12-06 17:54:31.358219] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.766 [2024-12-06 17:54:31.358232] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.766 [2024-12-06 17:54:31.358263] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.766 qpair failed and we were unable to recover it. 00:38:49.766 [2024-12-06 17:54:31.368142] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.766 [2024-12-06 17:54:31.368236] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.766 [2024-12-06 17:54:31.368260] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.766 [2024-12-06 17:54:31.368276] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.766 [2024-12-06 17:54:31.368289] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.766 [2024-12-06 17:54:31.368319] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.766 qpair failed and we were unable to recover it. 00:38:49.766 [2024-12-06 17:54:31.378094] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.766 [2024-12-06 17:54:31.378178] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.766 [2024-12-06 17:54:31.378203] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.766 [2024-12-06 17:54:31.378218] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.766 [2024-12-06 17:54:31.378231] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.766 [2024-12-06 17:54:31.378262] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.766 qpair failed and we were unable to recover it. 00:38:49.766 [2024-12-06 17:54:31.388125] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.766 [2024-12-06 17:54:31.388216] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.766 [2024-12-06 17:54:31.388241] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.766 [2024-12-06 17:54:31.388255] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.766 [2024-12-06 17:54:31.388267] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.766 [2024-12-06 17:54:31.388299] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.766 qpair failed and we were unable to recover it. 00:38:49.766 [2024-12-06 17:54:31.398179] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.766 [2024-12-06 17:54:31.398265] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.766 [2024-12-06 17:54:31.398289] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.766 [2024-12-06 17:54:31.398303] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.767 [2024-12-06 17:54:31.398316] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.767 [2024-12-06 17:54:31.398347] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.767 qpair failed and we were unable to recover it. 00:38:49.767 [2024-12-06 17:54:31.408237] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.767 [2024-12-06 17:54:31.408331] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.767 [2024-12-06 17:54:31.408355] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.767 [2024-12-06 17:54:31.408370] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.767 [2024-12-06 17:54:31.408383] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.767 [2024-12-06 17:54:31.408413] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.767 qpair failed and we were unable to recover it. 00:38:49.767 [2024-12-06 17:54:31.418217] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.767 [2024-12-06 17:54:31.418306] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.767 [2024-12-06 17:54:31.418331] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.767 [2024-12-06 17:54:31.418345] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.767 [2024-12-06 17:54:31.418358] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.767 [2024-12-06 17:54:31.418388] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.767 qpair failed and we were unable to recover it. 00:38:49.767 [2024-12-06 17:54:31.428295] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.767 [2024-12-06 17:54:31.428383] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.767 [2024-12-06 17:54:31.428421] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.767 [2024-12-06 17:54:31.428437] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.767 [2024-12-06 17:54:31.428449] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.767 [2024-12-06 17:54:31.428479] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.767 qpair failed and we were unable to recover it. 00:38:49.767 [2024-12-06 17:54:31.438273] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.767 [2024-12-06 17:54:31.438398] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.767 [2024-12-06 17:54:31.438425] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.767 [2024-12-06 17:54:31.438441] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.767 [2024-12-06 17:54:31.438453] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.767 [2024-12-06 17:54:31.438484] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.767 qpair failed and we were unable to recover it. 00:38:49.767 [2024-12-06 17:54:31.448312] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.767 [2024-12-06 17:54:31.448401] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.767 [2024-12-06 17:54:31.448426] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.767 [2024-12-06 17:54:31.448440] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.767 [2024-12-06 17:54:31.448456] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.767 [2024-12-06 17:54:31.448486] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.767 qpair failed and we were unable to recover it. 00:38:49.767 [2024-12-06 17:54:31.458334] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.767 [2024-12-06 17:54:31.458420] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.767 [2024-12-06 17:54:31.458445] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.767 [2024-12-06 17:54:31.458459] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.767 [2024-12-06 17:54:31.458472] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.767 [2024-12-06 17:54:31.458503] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.767 qpair failed and we were unable to recover it. 00:38:49.767 [2024-12-06 17:54:31.468419] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.767 [2024-12-06 17:54:31.468518] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.767 [2024-12-06 17:54:31.468543] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.767 [2024-12-06 17:54:31.468558] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.767 [2024-12-06 17:54:31.468577] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.767 [2024-12-06 17:54:31.468609] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.767 qpair failed and we were unable to recover it. 00:38:49.767 [2024-12-06 17:54:31.478379] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.767 [2024-12-06 17:54:31.478462] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.767 [2024-12-06 17:54:31.478488] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.767 [2024-12-06 17:54:31.478503] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.767 [2024-12-06 17:54:31.478516] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.767 [2024-12-06 17:54:31.478547] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.767 qpair failed and we were unable to recover it. 00:38:49.767 [2024-12-06 17:54:31.488417] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.767 [2024-12-06 17:54:31.488508] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.767 [2024-12-06 17:54:31.488534] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.767 [2024-12-06 17:54:31.488549] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.767 [2024-12-06 17:54:31.488562] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.767 [2024-12-06 17:54:31.488592] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.767 qpair failed and we were unable to recover it. 00:38:49.767 [2024-12-06 17:54:31.498522] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.767 [2024-12-06 17:54:31.498611] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.767 [2024-12-06 17:54:31.498637] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.767 [2024-12-06 17:54:31.498653] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.767 [2024-12-06 17:54:31.498672] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.767 [2024-12-06 17:54:31.498711] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.767 qpair failed and we were unable to recover it. 00:38:49.767 [2024-12-06 17:54:31.508467] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.767 [2024-12-06 17:54:31.508546] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.767 [2024-12-06 17:54:31.508572] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.767 [2024-12-06 17:54:31.508587] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.767 [2024-12-06 17:54:31.508600] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.767 [2024-12-06 17:54:31.508630] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.767 qpair failed and we were unable to recover it. 00:38:49.767 [2024-12-06 17:54:31.518556] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.767 [2024-12-06 17:54:31.518655] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.767 [2024-12-06 17:54:31.518688] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.767 [2024-12-06 17:54:31.518704] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.767 [2024-12-06 17:54:31.518717] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.767 [2024-12-06 17:54:31.518748] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.767 qpair failed and we were unable to recover it. 00:38:49.767 [2024-12-06 17:54:31.528647] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.767 [2024-12-06 17:54:31.528774] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.767 [2024-12-06 17:54:31.528799] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.767 [2024-12-06 17:54:31.528813] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.767 [2024-12-06 17:54:31.528826] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.768 [2024-12-06 17:54:31.528858] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.768 qpair failed and we were unable to recover it. 00:38:49.768 [2024-12-06 17:54:31.538559] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.768 [2024-12-06 17:54:31.538645] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.768 [2024-12-06 17:54:31.538679] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.768 [2024-12-06 17:54:31.538696] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.768 [2024-12-06 17:54:31.538709] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.768 [2024-12-06 17:54:31.538741] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.768 qpair failed and we were unable to recover it. 00:38:49.768 [2024-12-06 17:54:31.548612] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.768 [2024-12-06 17:54:31.548701] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.768 [2024-12-06 17:54:31.548727] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.768 [2024-12-06 17:54:31.548742] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.768 [2024-12-06 17:54:31.548756] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.768 [2024-12-06 17:54:31.548787] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.768 qpair failed and we were unable to recover it. 00:38:49.768 [2024-12-06 17:54:31.558694] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.768 [2024-12-06 17:54:31.558796] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.768 [2024-12-06 17:54:31.558836] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.768 [2024-12-06 17:54:31.558852] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.768 [2024-12-06 17:54:31.558865] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.768 [2024-12-06 17:54:31.558896] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.768 qpair failed and we were unable to recover it. 00:38:49.768 [2024-12-06 17:54:31.568680] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.768 [2024-12-06 17:54:31.568770] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.768 [2024-12-06 17:54:31.568796] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.768 [2024-12-06 17:54:31.568812] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.768 [2024-12-06 17:54:31.568825] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.768 [2024-12-06 17:54:31.568869] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.768 qpair failed and we were unable to recover it. 00:38:49.768 [2024-12-06 17:54:31.578698] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.768 [2024-12-06 17:54:31.578791] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.768 [2024-12-06 17:54:31.578817] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.768 [2024-12-06 17:54:31.578832] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.768 [2024-12-06 17:54:31.578846] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.768 [2024-12-06 17:54:31.578878] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.768 qpair failed and we were unable to recover it. 00:38:49.768 [2024-12-06 17:54:31.588725] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.768 [2024-12-06 17:54:31.588830] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.768 [2024-12-06 17:54:31.588859] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.768 [2024-12-06 17:54:31.588876] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.768 [2024-12-06 17:54:31.588890] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.768 [2024-12-06 17:54:31.588922] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.768 qpair failed and we were unable to recover it. 00:38:49.768 [2024-12-06 17:54:31.598775] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:49.768 [2024-12-06 17:54:31.598905] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:49.768 [2024-12-06 17:54:31.598930] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:49.768 [2024-12-06 17:54:31.598952] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:49.768 [2024-12-06 17:54:31.598965] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:49.768 [2024-12-06 17:54:31.598997] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:49.768 qpair failed and we were unable to recover it. 00:38:50.026 [2024-12-06 17:54:31.608891] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.026 [2024-12-06 17:54:31.608986] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.026 [2024-12-06 17:54:31.609014] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.026 [2024-12-06 17:54:31.609030] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.026 [2024-12-06 17:54:31.609043] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:50.026 [2024-12-06 17:54:31.609074] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:50.026 qpair failed and we were unable to recover it. 00:38:50.026 [2024-12-06 17:54:31.618811] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.026 [2024-12-06 17:54:31.618899] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.026 [2024-12-06 17:54:31.618925] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.026 [2024-12-06 17:54:31.618939] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.026 [2024-12-06 17:54:31.618952] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:50.026 [2024-12-06 17:54:31.618983] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:50.026 qpair failed and we were unable to recover it. 00:38:50.026 [2024-12-06 17:54:31.628897] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.026 [2024-12-06 17:54:31.628986] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.026 [2024-12-06 17:54:31.629011] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.026 [2024-12-06 17:54:31.629026] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.026 [2024-12-06 17:54:31.629039] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:50.026 [2024-12-06 17:54:31.629071] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:50.026 qpair failed and we were unable to recover it. 00:38:50.026 [2024-12-06 17:54:31.638886] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.027 [2024-12-06 17:54:31.638970] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.027 [2024-12-06 17:54:31.638996] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.027 [2024-12-06 17:54:31.639011] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.027 [2024-12-06 17:54:31.639025] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:50.027 [2024-12-06 17:54:31.639074] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:50.027 qpair failed and we were unable to recover it. 00:38:50.027 [2024-12-06 17:54:31.648910] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.027 [2024-12-06 17:54:31.648995] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.027 [2024-12-06 17:54:31.649021] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.027 [2024-12-06 17:54:31.649036] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.027 [2024-12-06 17:54:31.649049] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:50.027 [2024-12-06 17:54:31.649079] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:50.027 qpair failed and we were unable to recover it. 00:38:50.027 [2024-12-06 17:54:31.659078] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.027 [2024-12-06 17:54:31.659167] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.027 [2024-12-06 17:54:31.659193] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.027 [2024-12-06 17:54:31.659208] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.027 [2024-12-06 17:54:31.659220] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:50.027 [2024-12-06 17:54:31.659251] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:50.027 qpair failed and we were unable to recover it. 00:38:50.027 [2024-12-06 17:54:31.669017] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.027 [2024-12-06 17:54:31.669141] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.027 [2024-12-06 17:54:31.669170] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.027 [2024-12-06 17:54:31.669188] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.027 [2024-12-06 17:54:31.669202] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:50.027 [2024-12-06 17:54:31.669233] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:50.027 qpair failed and we were unable to recover it. 00:38:50.027 [2024-12-06 17:54:31.678990] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.027 [2024-12-06 17:54:31.679076] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.027 [2024-12-06 17:54:31.679102] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.027 [2024-12-06 17:54:31.679117] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.027 [2024-12-06 17:54:31.679131] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:50.027 [2024-12-06 17:54:31.679162] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:50.027 qpair failed and we were unable to recover it. 00:38:50.027 [2024-12-06 17:54:31.689125] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.027 [2024-12-06 17:54:31.689240] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.027 [2024-12-06 17:54:31.689265] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.027 [2024-12-06 17:54:31.689280] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.027 [2024-12-06 17:54:31.689292] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:50.027 [2024-12-06 17:54:31.689324] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:50.027 qpair failed and we were unable to recover it. 00:38:50.027 [2024-12-06 17:54:31.699132] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.027 [2024-12-06 17:54:31.699268] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.027 [2024-12-06 17:54:31.699298] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.027 [2024-12-06 17:54:31.699315] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.027 [2024-12-06 17:54:31.699328] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:50.027 [2024-12-06 17:54:31.699375] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:50.027 qpair failed and we were unable to recover it. 00:38:50.027 [2024-12-06 17:54:31.709091] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.027 [2024-12-06 17:54:31.709177] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.027 [2024-12-06 17:54:31.709203] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.027 [2024-12-06 17:54:31.709218] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.027 [2024-12-06 17:54:31.709231] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:50.027 [2024-12-06 17:54:31.709262] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:50.027 qpair failed and we were unable to recover it. 00:38:50.027 [2024-12-06 17:54:31.719129] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.027 [2024-12-06 17:54:31.719250] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.027 [2024-12-06 17:54:31.719275] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.027 [2024-12-06 17:54:31.719290] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.027 [2024-12-06 17:54:31.719304] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:50.027 [2024-12-06 17:54:31.719334] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:50.027 qpair failed and we were unable to recover it. 00:38:50.027 [2024-12-06 17:54:31.729226] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.027 [2024-12-06 17:54:31.729317] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.027 [2024-12-06 17:54:31.729343] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.027 [2024-12-06 17:54:31.729364] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.027 [2024-12-06 17:54:31.729378] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:50.027 [2024-12-06 17:54:31.729421] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:50.027 qpair failed and we were unable to recover it. 00:38:50.027 [2024-12-06 17:54:31.739161] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.027 [2024-12-06 17:54:31.739249] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.027 [2024-12-06 17:54:31.739275] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.027 [2024-12-06 17:54:31.739291] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.027 [2024-12-06 17:54:31.739303] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:50.027 [2024-12-06 17:54:31.739334] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:50.027 qpair failed and we were unable to recover it. 00:38:50.027 [2024-12-06 17:54:31.749210] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.027 [2024-12-06 17:54:31.749292] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.027 [2024-12-06 17:54:31.749318] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.027 [2024-12-06 17:54:31.749333] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.027 [2024-12-06 17:54:31.749346] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:50.027 [2024-12-06 17:54:31.749376] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:50.027 qpair failed and we were unable to recover it. 00:38:50.027 [2024-12-06 17:54:31.759238] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.027 [2024-12-06 17:54:31.759320] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.027 [2024-12-06 17:54:31.759346] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.027 [2024-12-06 17:54:31.759362] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.027 [2024-12-06 17:54:31.759375] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:50.027 [2024-12-06 17:54:31.759406] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:50.027 qpair failed and we were unable to recover it. 00:38:50.027 [2024-12-06 17:54:31.769251] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.028 [2024-12-06 17:54:31.769340] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.028 [2024-12-06 17:54:31.769366] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.028 [2024-12-06 17:54:31.769381] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.028 [2024-12-06 17:54:31.769394] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:50.028 [2024-12-06 17:54:31.769433] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:50.028 qpair failed and we were unable to recover it. 00:38:50.028 [2024-12-06 17:54:31.779291] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.028 [2024-12-06 17:54:31.779423] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.028 [2024-12-06 17:54:31.779448] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.028 [2024-12-06 17:54:31.779464] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.028 [2024-12-06 17:54:31.779478] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:50.028 [2024-12-06 17:54:31.779508] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:50.028 qpair failed and we were unable to recover it. 00:38:50.028 [2024-12-06 17:54:31.789354] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.028 [2024-12-06 17:54:31.789433] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.028 [2024-12-06 17:54:31.789459] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.028 [2024-12-06 17:54:31.789474] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.028 [2024-12-06 17:54:31.789487] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:50.028 [2024-12-06 17:54:31.789518] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:50.028 qpair failed and we were unable to recover it. 00:38:50.028 [2024-12-06 17:54:31.799356] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.028 [2024-12-06 17:54:31.799444] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.028 [2024-12-06 17:54:31.799469] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.028 [2024-12-06 17:54:31.799484] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.028 [2024-12-06 17:54:31.799497] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:50.028 [2024-12-06 17:54:31.799527] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:50.028 qpair failed and we were unable to recover it. 00:38:50.028 [2024-12-06 17:54:31.809376] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.028 [2024-12-06 17:54:31.809500] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.028 [2024-12-06 17:54:31.809526] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.028 [2024-12-06 17:54:31.809540] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.028 [2024-12-06 17:54:31.809554] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:50.028 [2024-12-06 17:54:31.809585] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:50.028 qpair failed and we were unable to recover it. 00:38:50.028 [2024-12-06 17:54:31.819419] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.028 [2024-12-06 17:54:31.819512] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.028 [2024-12-06 17:54:31.819544] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.028 [2024-12-06 17:54:31.819560] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.028 [2024-12-06 17:54:31.819572] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.028 [2024-12-06 17:54:31.819605] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.028 qpair failed and we were unable to recover it. 00:38:50.028 [2024-12-06 17:54:31.829422] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.028 [2024-12-06 17:54:31.829508] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.028 [2024-12-06 17:54:31.829534] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.028 [2024-12-06 17:54:31.829549] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.028 [2024-12-06 17:54:31.829562] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.028 [2024-12-06 17:54:31.829593] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.028 qpair failed and we were unable to recover it. 00:38:50.028 [2024-12-06 17:54:31.839534] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.028 [2024-12-06 17:54:31.839615] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.028 [2024-12-06 17:54:31.839641] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.028 [2024-12-06 17:54:31.839656] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.028 [2024-12-06 17:54:31.839678] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.028 [2024-12-06 17:54:31.839709] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.028 qpair failed and we were unable to recover it. 00:38:50.028 [2024-12-06 17:54:31.849507] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.028 [2024-12-06 17:54:31.849601] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.028 [2024-12-06 17:54:31.849627] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.028 [2024-12-06 17:54:31.849642] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.028 [2024-12-06 17:54:31.849654] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.028 [2024-12-06 17:54:31.849693] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.028 qpair failed and we were unable to recover it. 00:38:50.028 [2024-12-06 17:54:31.859532] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.028 [2024-12-06 17:54:31.859645] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.028 [2024-12-06 17:54:31.859684] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.028 [2024-12-06 17:54:31.859701] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.028 [2024-12-06 17:54:31.859715] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.028 [2024-12-06 17:54:31.859745] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.028 qpair failed and we were unable to recover it. 00:38:50.286 [2024-12-06 17:54:31.869541] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.286 [2024-12-06 17:54:31.869626] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.286 [2024-12-06 17:54:31.869651] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.286 [2024-12-06 17:54:31.869673] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.286 [2024-12-06 17:54:31.869688] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.286 [2024-12-06 17:54:31.869719] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.286 qpair failed and we were unable to recover it. 00:38:50.286 [2024-12-06 17:54:31.879539] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.286 [2024-12-06 17:54:31.879623] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.286 [2024-12-06 17:54:31.879647] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.286 [2024-12-06 17:54:31.879662] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.286 [2024-12-06 17:54:31.879688] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.286 [2024-12-06 17:54:31.879720] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.286 qpair failed and we were unable to recover it. 00:38:50.286 [2024-12-06 17:54:31.889611] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.286 [2024-12-06 17:54:31.889714] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.287 [2024-12-06 17:54:31.889740] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.287 [2024-12-06 17:54:31.889754] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.287 [2024-12-06 17:54:31.889766] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.287 [2024-12-06 17:54:31.889796] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.287 qpair failed and we were unable to recover it. 00:38:50.287 [2024-12-06 17:54:31.899624] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.287 [2024-12-06 17:54:31.899720] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.287 [2024-12-06 17:54:31.899746] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.287 [2024-12-06 17:54:31.899761] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.287 [2024-12-06 17:54:31.899780] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.287 [2024-12-06 17:54:31.899811] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.287 qpair failed and we were unable to recover it. 00:38:50.287 [2024-12-06 17:54:31.909682] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.287 [2024-12-06 17:54:31.909769] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.287 [2024-12-06 17:54:31.909795] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.287 [2024-12-06 17:54:31.909809] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.287 [2024-12-06 17:54:31.909823] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.287 [2024-12-06 17:54:31.909854] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.287 qpair failed and we were unable to recover it. 00:38:50.287 [2024-12-06 17:54:31.919674] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.287 [2024-12-06 17:54:31.919757] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.287 [2024-12-06 17:54:31.919782] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.287 [2024-12-06 17:54:31.919797] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.287 [2024-12-06 17:54:31.919811] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.287 [2024-12-06 17:54:31.919841] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.287 qpair failed and we were unable to recover it. 00:38:50.287 [2024-12-06 17:54:31.929752] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.287 [2024-12-06 17:54:31.929859] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.287 [2024-12-06 17:54:31.929885] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.287 [2024-12-06 17:54:31.929901] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.287 [2024-12-06 17:54:31.929914] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.287 [2024-12-06 17:54:31.929945] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.287 qpair failed and we were unable to recover it. 00:38:50.287 [2024-12-06 17:54:31.939748] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.287 [2024-12-06 17:54:31.939840] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.287 [2024-12-06 17:54:31.939865] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.287 [2024-12-06 17:54:31.939880] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.287 [2024-12-06 17:54:31.939893] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.287 [2024-12-06 17:54:31.939924] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.287 qpair failed and we were unable to recover it. 00:38:50.287 [2024-12-06 17:54:31.949776] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.287 [2024-12-06 17:54:31.949868] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.287 [2024-12-06 17:54:31.949894] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.287 [2024-12-06 17:54:31.949908] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.287 [2024-12-06 17:54:31.949921] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.287 [2024-12-06 17:54:31.949951] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.287 qpair failed and we were unable to recover it. 00:38:50.287 [2024-12-06 17:54:31.959833] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.287 [2024-12-06 17:54:31.959920] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.287 [2024-12-06 17:54:31.959945] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.287 [2024-12-06 17:54:31.959960] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.287 [2024-12-06 17:54:31.959973] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.287 [2024-12-06 17:54:31.960003] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.287 qpair failed and we were unable to recover it. 00:38:50.287 [2024-12-06 17:54:31.969887] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.287 [2024-12-06 17:54:31.969973] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.287 [2024-12-06 17:54:31.969999] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.287 [2024-12-06 17:54:31.970014] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.287 [2024-12-06 17:54:31.970027] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.287 [2024-12-06 17:54:31.970057] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.287 qpair failed and we were unable to recover it. 00:38:50.287 [2024-12-06 17:54:31.979897] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.287 [2024-12-06 17:54:31.979995] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.287 [2024-12-06 17:54:31.980021] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.287 [2024-12-06 17:54:31.980036] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.287 [2024-12-06 17:54:31.980049] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.287 [2024-12-06 17:54:31.980092] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.287 qpair failed and we were unable to recover it. 00:38:50.287 [2024-12-06 17:54:31.989901] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.287 [2024-12-06 17:54:31.989983] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.287 [2024-12-06 17:54:31.990013] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.287 [2024-12-06 17:54:31.990029] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.287 [2024-12-06 17:54:31.990043] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.287 [2024-12-06 17:54:31.990073] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.287 qpair failed and we were unable to recover it. 00:38:50.287 [2024-12-06 17:54:31.999912] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.287 [2024-12-06 17:54:32.000007] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.287 [2024-12-06 17:54:32.000035] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.287 [2024-12-06 17:54:32.000052] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.287 [2024-12-06 17:54:32.000066] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.287 [2024-12-06 17:54:32.000099] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.287 qpair failed and we were unable to recover it. 00:38:50.287 [2024-12-06 17:54:32.009980] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.287 [2024-12-06 17:54:32.010078] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.287 [2024-12-06 17:54:32.010103] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.287 [2024-12-06 17:54:32.010118] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.287 [2024-12-06 17:54:32.010131] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.287 [2024-12-06 17:54:32.010175] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.287 qpair failed and we were unable to recover it. 00:38:50.287 [2024-12-06 17:54:32.019987] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.287 [2024-12-06 17:54:32.020071] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.287 [2024-12-06 17:54:32.020097] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.288 [2024-12-06 17:54:32.020112] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.288 [2024-12-06 17:54:32.020125] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.288 [2024-12-06 17:54:32.020155] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.288 qpair failed and we were unable to recover it. 00:38:50.288 [2024-12-06 17:54:32.029987] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.288 [2024-12-06 17:54:32.030082] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.288 [2024-12-06 17:54:32.030107] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.288 [2024-12-06 17:54:32.030122] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.288 [2024-12-06 17:54:32.030141] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.288 [2024-12-06 17:54:32.030175] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.288 qpair failed and we were unable to recover it. 00:38:50.288 [2024-12-06 17:54:32.040036] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.288 [2024-12-06 17:54:32.040118] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.288 [2024-12-06 17:54:32.040144] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.288 [2024-12-06 17:54:32.040159] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.288 [2024-12-06 17:54:32.040171] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.288 [2024-12-06 17:54:32.040202] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.288 qpair failed and we were unable to recover it. 00:38:50.288 [2024-12-06 17:54:32.050084] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.288 [2024-12-06 17:54:32.050175] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.288 [2024-12-06 17:54:32.050201] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.288 [2024-12-06 17:54:32.050215] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.288 [2024-12-06 17:54:32.050228] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.288 [2024-12-06 17:54:32.050259] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.288 qpair failed and we were unable to recover it. 00:38:50.288 [2024-12-06 17:54:32.060083] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.288 [2024-12-06 17:54:32.060168] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.288 [2024-12-06 17:54:32.060192] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.288 [2024-12-06 17:54:32.060207] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.288 [2024-12-06 17:54:32.060220] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.288 [2024-12-06 17:54:32.060250] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.288 qpair failed and we were unable to recover it. 00:38:50.288 [2024-12-06 17:54:32.070097] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.288 [2024-12-06 17:54:32.070185] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.288 [2024-12-06 17:54:32.070210] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.288 [2024-12-06 17:54:32.070225] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.288 [2024-12-06 17:54:32.070238] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.288 [2024-12-06 17:54:32.070268] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.288 qpair failed and we were unable to recover it. 00:38:50.288 [2024-12-06 17:54:32.080188] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.288 [2024-12-06 17:54:32.080275] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.288 [2024-12-06 17:54:32.080303] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.288 [2024-12-06 17:54:32.080318] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.288 [2024-12-06 17:54:32.080332] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.288 [2024-12-06 17:54:32.080375] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.288 qpair failed and we were unable to recover it. 00:38:50.288 [2024-12-06 17:54:32.090184] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.288 [2024-12-06 17:54:32.090275] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.288 [2024-12-06 17:54:32.090301] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.288 [2024-12-06 17:54:32.090316] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.288 [2024-12-06 17:54:32.090329] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.288 [2024-12-06 17:54:32.090361] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.288 qpair failed and we were unable to recover it. 00:38:50.288 [2024-12-06 17:54:32.100228] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.288 [2024-12-06 17:54:32.100319] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.288 [2024-12-06 17:54:32.100345] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.288 [2024-12-06 17:54:32.100361] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.288 [2024-12-06 17:54:32.100374] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.288 [2024-12-06 17:54:32.100404] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.288 qpair failed and we were unable to recover it. 00:38:50.288 [2024-12-06 17:54:32.110250] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.288 [2024-12-06 17:54:32.110328] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.288 [2024-12-06 17:54:32.110356] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.288 [2024-12-06 17:54:32.110372] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.288 [2024-12-06 17:54:32.110385] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.288 [2024-12-06 17:54:32.110428] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.288 qpair failed and we were unable to recover it. 00:38:50.288 [2024-12-06 17:54:32.120399] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.288 [2024-12-06 17:54:32.120529] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.288 [2024-12-06 17:54:32.120565] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.288 [2024-12-06 17:54:32.120582] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.288 [2024-12-06 17:54:32.120595] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.288 [2024-12-06 17:54:32.120625] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.288 qpair failed and we were unable to recover it. 00:38:50.547 [2024-12-06 17:54:32.130281] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.547 [2024-12-06 17:54:32.130372] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.547 [2024-12-06 17:54:32.130397] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.547 [2024-12-06 17:54:32.130413] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.547 [2024-12-06 17:54:32.130426] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.547 [2024-12-06 17:54:32.130456] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.547 qpair failed and we were unable to recover it. 00:38:50.547 [2024-12-06 17:54:32.140299] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.547 [2024-12-06 17:54:32.140396] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.547 [2024-12-06 17:54:32.140421] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.547 [2024-12-06 17:54:32.140435] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.547 [2024-12-06 17:54:32.140449] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.547 [2024-12-06 17:54:32.140479] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.547 qpair failed and we were unable to recover it. 00:38:50.547 [2024-12-06 17:54:32.150366] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.547 [2024-12-06 17:54:32.150446] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.547 [2024-12-06 17:54:32.150471] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.547 [2024-12-06 17:54:32.150487] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.547 [2024-12-06 17:54:32.150500] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.547 [2024-12-06 17:54:32.150529] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.547 qpair failed and we were unable to recover it. 00:38:50.547 [2024-12-06 17:54:32.160380] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.547 [2024-12-06 17:54:32.160469] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.547 [2024-12-06 17:54:32.160495] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.547 [2024-12-06 17:54:32.160515] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.547 [2024-12-06 17:54:32.160529] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.547 [2024-12-06 17:54:32.160559] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.547 qpair failed and we were unable to recover it. 00:38:50.547 [2024-12-06 17:54:32.170402] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.547 [2024-12-06 17:54:32.170491] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.547 [2024-12-06 17:54:32.170516] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.547 [2024-12-06 17:54:32.170531] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.547 [2024-12-06 17:54:32.170544] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.547 [2024-12-06 17:54:32.170575] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.547 qpair failed and we were unable to recover it. 00:38:50.547 [2024-12-06 17:54:32.180443] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.547 [2024-12-06 17:54:32.180529] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.547 [2024-12-06 17:54:32.180554] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.547 [2024-12-06 17:54:32.180569] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.547 [2024-12-06 17:54:32.180582] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.547 [2024-12-06 17:54:32.180612] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.547 qpair failed and we were unable to recover it. 00:38:50.547 [2024-12-06 17:54:32.190433] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.548 [2024-12-06 17:54:32.190521] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.548 [2024-12-06 17:54:32.190547] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.548 [2024-12-06 17:54:32.190562] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.548 [2024-12-06 17:54:32.190576] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.548 [2024-12-06 17:54:32.190606] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.548 qpair failed and we were unable to recover it. 00:38:50.548 [2024-12-06 17:54:32.200474] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.548 [2024-12-06 17:54:32.200560] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.548 [2024-12-06 17:54:32.200586] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.548 [2024-12-06 17:54:32.200601] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.548 [2024-12-06 17:54:32.200614] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.548 [2024-12-06 17:54:32.200650] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.548 qpair failed and we were unable to recover it. 00:38:50.548 [2024-12-06 17:54:32.210500] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.548 [2024-12-06 17:54:32.210590] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.548 [2024-12-06 17:54:32.210615] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.548 [2024-12-06 17:54:32.210630] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.548 [2024-12-06 17:54:32.210644] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.548 [2024-12-06 17:54:32.210682] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.548 qpair failed and we were unable to recover it. 00:38:50.548 [2024-12-06 17:54:32.220530] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.548 [2024-12-06 17:54:32.220613] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.548 [2024-12-06 17:54:32.220638] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.548 [2024-12-06 17:54:32.220653] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.548 [2024-12-06 17:54:32.220672] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.548 [2024-12-06 17:54:32.220704] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.548 qpair failed and we were unable to recover it. 00:38:50.548 [2024-12-06 17:54:32.230536] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.548 [2024-12-06 17:54:32.230617] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.548 [2024-12-06 17:54:32.230642] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.548 [2024-12-06 17:54:32.230658] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.548 [2024-12-06 17:54:32.230680] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.548 [2024-12-06 17:54:32.230711] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.548 qpair failed and we were unable to recover it. 00:38:50.548 [2024-12-06 17:54:32.240595] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.548 [2024-12-06 17:54:32.240684] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.548 [2024-12-06 17:54:32.240710] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.548 [2024-12-06 17:54:32.240726] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.548 [2024-12-06 17:54:32.240738] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.548 [2024-12-06 17:54:32.240768] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.548 qpair failed and we were unable to recover it. 00:38:50.548 [2024-12-06 17:54:32.250619] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.548 [2024-12-06 17:54:32.250720] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.548 [2024-12-06 17:54:32.250746] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.548 [2024-12-06 17:54:32.250761] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.548 [2024-12-06 17:54:32.250774] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.548 [2024-12-06 17:54:32.250805] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.548 qpair failed and we were unable to recover it. 00:38:50.548 [2024-12-06 17:54:32.260640] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.548 [2024-12-06 17:54:32.260733] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.548 [2024-12-06 17:54:32.260758] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.548 [2024-12-06 17:54:32.260773] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.548 [2024-12-06 17:54:32.260786] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.548 [2024-12-06 17:54:32.260817] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.548 qpair failed and we were unable to recover it. 00:38:50.548 [2024-12-06 17:54:32.270658] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.548 [2024-12-06 17:54:32.270755] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.548 [2024-12-06 17:54:32.270779] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.548 [2024-12-06 17:54:32.270793] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.548 [2024-12-06 17:54:32.270806] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.548 [2024-12-06 17:54:32.270837] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.548 qpair failed and we were unable to recover it. 00:38:50.548 [2024-12-06 17:54:32.280690] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.548 [2024-12-06 17:54:32.280786] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.548 [2024-12-06 17:54:32.280810] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.548 [2024-12-06 17:54:32.280825] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.548 [2024-12-06 17:54:32.280839] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.548 [2024-12-06 17:54:32.280869] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.548 qpair failed and we were unable to recover it. 00:38:50.548 [2024-12-06 17:54:32.290774] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.549 [2024-12-06 17:54:32.290867] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.549 [2024-12-06 17:54:32.290892] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.549 [2024-12-06 17:54:32.290912] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.549 [2024-12-06 17:54:32.290926] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.549 [2024-12-06 17:54:32.290957] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.549 qpair failed and we were unable to recover it. 00:38:50.549 [2024-12-06 17:54:32.300763] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.549 [2024-12-06 17:54:32.300849] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.549 [2024-12-06 17:54:32.300874] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.549 [2024-12-06 17:54:32.300888] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.549 [2024-12-06 17:54:32.300902] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.549 [2024-12-06 17:54:32.300932] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.549 qpair failed and we were unable to recover it. 00:38:50.549 [2024-12-06 17:54:32.310790] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.549 [2024-12-06 17:54:32.310875] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.549 [2024-12-06 17:54:32.310899] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.549 [2024-12-06 17:54:32.310913] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.549 [2024-12-06 17:54:32.310926] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.549 [2024-12-06 17:54:32.310957] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.549 qpair failed and we were unable to recover it. 00:38:50.549 [2024-12-06 17:54:32.320850] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.549 [2024-12-06 17:54:32.320938] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.549 [2024-12-06 17:54:32.320964] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.549 [2024-12-06 17:54:32.320978] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.549 [2024-12-06 17:54:32.320992] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.549 [2024-12-06 17:54:32.321036] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.549 qpair failed and we were unable to recover it. 00:38:50.549 [2024-12-06 17:54:32.330884] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.549 [2024-12-06 17:54:32.330977] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.549 [2024-12-06 17:54:32.331004] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.549 [2024-12-06 17:54:32.331020] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.549 [2024-12-06 17:54:32.331033] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.549 [2024-12-06 17:54:32.331071] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.549 qpair failed and we were unable to recover it. 00:38:50.549 [2024-12-06 17:54:32.340887] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.549 [2024-12-06 17:54:32.341011] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.549 [2024-12-06 17:54:32.341041] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.549 [2024-12-06 17:54:32.341058] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.549 [2024-12-06 17:54:32.341071] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.549 [2024-12-06 17:54:32.341102] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.549 qpair failed and we were unable to recover it. 00:38:50.549 [2024-12-06 17:54:32.350930] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.549 [2024-12-06 17:54:32.351011] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.549 [2024-12-06 17:54:32.351036] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.549 [2024-12-06 17:54:32.351051] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.549 [2024-12-06 17:54:32.351064] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.549 [2024-12-06 17:54:32.351106] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.549 qpair failed and we were unable to recover it. 00:38:50.549 [2024-12-06 17:54:32.360995] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.549 [2024-12-06 17:54:32.361083] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.549 [2024-12-06 17:54:32.361108] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.549 [2024-12-06 17:54:32.361122] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.549 [2024-12-06 17:54:32.361135] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.549 [2024-12-06 17:54:32.361166] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.549 qpair failed and we were unable to recover it. 00:38:50.549 [2024-12-06 17:54:32.370983] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.549 [2024-12-06 17:54:32.371073] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.549 [2024-12-06 17:54:32.371098] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.549 [2024-12-06 17:54:32.371112] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.549 [2024-12-06 17:54:32.371125] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.549 [2024-12-06 17:54:32.371155] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.549 qpair failed and we were unable to recover it. 00:38:50.549 [2024-12-06 17:54:32.381026] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.549 [2024-12-06 17:54:32.381115] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.549 [2024-12-06 17:54:32.381144] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.550 [2024-12-06 17:54:32.381161] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.550 [2024-12-06 17:54:32.381174] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.550 [2024-12-06 17:54:32.381205] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.550 qpair failed and we were unable to recover it. 00:38:50.808 [2024-12-06 17:54:32.391128] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.808 [2024-12-06 17:54:32.391216] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.808 [2024-12-06 17:54:32.391242] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.808 [2024-12-06 17:54:32.391257] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.808 [2024-12-06 17:54:32.391270] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.808 [2024-12-06 17:54:32.391300] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.808 qpair failed and we were unable to recover it. 00:38:50.808 [2024-12-06 17:54:32.401071] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.808 [2024-12-06 17:54:32.401153] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.808 [2024-12-06 17:54:32.401178] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.808 [2024-12-06 17:54:32.401193] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.808 [2024-12-06 17:54:32.401206] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.808 [2024-12-06 17:54:32.401248] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.808 qpair failed and we were unable to recover it. 00:38:50.808 [2024-12-06 17:54:32.411068] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.808 [2024-12-06 17:54:32.411162] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.808 [2024-12-06 17:54:32.411187] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.808 [2024-12-06 17:54:32.411201] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.808 [2024-12-06 17:54:32.411214] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.808 [2024-12-06 17:54:32.411244] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.808 qpair failed and we were unable to recover it. 00:38:50.809 [2024-12-06 17:54:32.421119] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.809 [2024-12-06 17:54:32.421210] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.809 [2024-12-06 17:54:32.421240] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.809 [2024-12-06 17:54:32.421256] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.809 [2024-12-06 17:54:32.421269] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.809 [2024-12-06 17:54:32.421299] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.809 qpair failed and we were unable to recover it. 00:38:50.809 [2024-12-06 17:54:32.431102] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.809 [2024-12-06 17:54:32.431186] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.809 [2024-12-06 17:54:32.431211] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.809 [2024-12-06 17:54:32.431226] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.809 [2024-12-06 17:54:32.431239] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.809 [2024-12-06 17:54:32.431269] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.809 qpair failed and we were unable to recover it. 00:38:50.809 [2024-12-06 17:54:32.441194] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.809 [2024-12-06 17:54:32.441273] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.809 [2024-12-06 17:54:32.441298] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.809 [2024-12-06 17:54:32.441312] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.809 [2024-12-06 17:54:32.441325] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.809 [2024-12-06 17:54:32.441355] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.809 qpair failed and we were unable to recover it. 00:38:50.809 [2024-12-06 17:54:32.451194] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.809 [2024-12-06 17:54:32.451285] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.809 [2024-12-06 17:54:32.451309] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.809 [2024-12-06 17:54:32.451324] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.809 [2024-12-06 17:54:32.451338] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.809 [2024-12-06 17:54:32.451368] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.809 qpair failed and we were unable to recover it. 00:38:50.809 [2024-12-06 17:54:32.461197] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.809 [2024-12-06 17:54:32.461283] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.809 [2024-12-06 17:54:32.461308] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.809 [2024-12-06 17:54:32.461323] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.809 [2024-12-06 17:54:32.461341] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.809 [2024-12-06 17:54:32.461372] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.809 qpair failed and we were unable to recover it. 00:38:50.809 [2024-12-06 17:54:32.471206] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.809 [2024-12-06 17:54:32.471290] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.809 [2024-12-06 17:54:32.471314] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.809 [2024-12-06 17:54:32.471329] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.809 [2024-12-06 17:54:32.471342] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.809 [2024-12-06 17:54:32.471372] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.809 qpair failed and we were unable to recover it. 00:38:50.809 [2024-12-06 17:54:32.481347] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.809 [2024-12-06 17:54:32.481429] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.809 [2024-12-06 17:54:32.481453] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.809 [2024-12-06 17:54:32.481467] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.809 [2024-12-06 17:54:32.481480] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.809 [2024-12-06 17:54:32.481510] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.809 qpair failed and we were unable to recover it. 00:38:50.809 [2024-12-06 17:54:32.491286] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.809 [2024-12-06 17:54:32.491379] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.809 [2024-12-06 17:54:32.491404] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.809 [2024-12-06 17:54:32.491419] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.809 [2024-12-06 17:54:32.491432] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.809 [2024-12-06 17:54:32.491462] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.809 qpair failed and we were unable to recover it. 00:38:50.809 [2024-12-06 17:54:32.501337] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.809 [2024-12-06 17:54:32.501430] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.809 [2024-12-06 17:54:32.501458] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.809 [2024-12-06 17:54:32.501475] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.809 [2024-12-06 17:54:32.501488] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.809 [2024-12-06 17:54:32.501531] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.809 qpair failed and we were unable to recover it. 00:38:50.809 [2024-12-06 17:54:32.511334] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.809 [2024-12-06 17:54:32.511465] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.809 [2024-12-06 17:54:32.511493] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.809 [2024-12-06 17:54:32.511508] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.809 [2024-12-06 17:54:32.511521] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.809 [2024-12-06 17:54:32.511551] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.810 qpair failed and we were unable to recover it. 00:38:50.810 [2024-12-06 17:54:32.521348] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.810 [2024-12-06 17:54:32.521440] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.810 [2024-12-06 17:54:32.521464] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.810 [2024-12-06 17:54:32.521479] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.810 [2024-12-06 17:54:32.521492] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.810 [2024-12-06 17:54:32.521522] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.810 qpair failed and we were unable to recover it. 00:38:50.810 [2024-12-06 17:54:32.531391] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.810 [2024-12-06 17:54:32.531507] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.810 [2024-12-06 17:54:32.531533] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.810 [2024-12-06 17:54:32.531548] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.810 [2024-12-06 17:54:32.531561] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.810 [2024-12-06 17:54:32.531591] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.810 qpair failed and we were unable to recover it. 00:38:50.810 [2024-12-06 17:54:32.541411] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.810 [2024-12-06 17:54:32.541495] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.810 [2024-12-06 17:54:32.541521] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.810 [2024-12-06 17:54:32.541536] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.810 [2024-12-06 17:54:32.541549] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.810 [2024-12-06 17:54:32.541579] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.810 qpair failed and we were unable to recover it. 00:38:50.810 [2024-12-06 17:54:32.551472] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.810 [2024-12-06 17:54:32.551585] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.810 [2024-12-06 17:54:32.551617] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.810 [2024-12-06 17:54:32.551633] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.810 [2024-12-06 17:54:32.551646] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.810 [2024-12-06 17:54:32.551688] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.810 qpair failed and we were unable to recover it. 00:38:50.810 [2024-12-06 17:54:32.561544] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.810 [2024-12-06 17:54:32.561654] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.810 [2024-12-06 17:54:32.561692] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.810 [2024-12-06 17:54:32.561708] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.810 [2024-12-06 17:54:32.561722] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.810 [2024-12-06 17:54:32.561753] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.810 qpair failed and we were unable to recover it. 00:38:50.810 [2024-12-06 17:54:32.571525] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.810 [2024-12-06 17:54:32.571624] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.810 [2024-12-06 17:54:32.571649] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.810 [2024-12-06 17:54:32.571672] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.810 [2024-12-06 17:54:32.571688] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.810 [2024-12-06 17:54:32.571718] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.810 qpair failed and we were unable to recover it. 00:38:50.810 [2024-12-06 17:54:32.581571] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.810 [2024-12-06 17:54:32.581686] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.810 [2024-12-06 17:54:32.581715] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.810 [2024-12-06 17:54:32.581730] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.810 [2024-12-06 17:54:32.581744] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.810 [2024-12-06 17:54:32.581775] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.810 qpair failed and we were unable to recover it. 00:38:50.810 [2024-12-06 17:54:32.591574] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.810 [2024-12-06 17:54:32.591678] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.810 [2024-12-06 17:54:32.591706] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.810 [2024-12-06 17:54:32.591721] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.810 [2024-12-06 17:54:32.591739] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.810 [2024-12-06 17:54:32.591771] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.810 qpair failed and we were unable to recover it. 00:38:50.810 [2024-12-06 17:54:32.601616] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.810 [2024-12-06 17:54:32.601723] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.810 [2024-12-06 17:54:32.601749] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.810 [2024-12-06 17:54:32.601764] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.810 [2024-12-06 17:54:32.601777] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.810 [2024-12-06 17:54:32.601808] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.810 qpair failed and we were unable to recover it. 00:38:50.810 [2024-12-06 17:54:32.611646] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.810 [2024-12-06 17:54:32.611752] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.810 [2024-12-06 17:54:32.611778] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.810 [2024-12-06 17:54:32.611792] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.810 [2024-12-06 17:54:32.611805] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.811 [2024-12-06 17:54:32.611836] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.811 qpair failed and we were unable to recover it. 00:38:50.811 [2024-12-06 17:54:32.621639] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.811 [2024-12-06 17:54:32.621768] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.811 [2024-12-06 17:54:32.621793] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.811 [2024-12-06 17:54:32.621808] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.811 [2024-12-06 17:54:32.621822] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.811 [2024-12-06 17:54:32.621852] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.811 qpair failed and we were unable to recover it. 00:38:50.811 [2024-12-06 17:54:32.631661] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.811 [2024-12-06 17:54:32.631762] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.811 [2024-12-06 17:54:32.631788] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.811 [2024-12-06 17:54:32.631802] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.811 [2024-12-06 17:54:32.631815] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.811 [2024-12-06 17:54:32.631846] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.811 qpair failed and we were unable to recover it. 00:38:50.811 [2024-12-06 17:54:32.641716] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:50.811 [2024-12-06 17:54:32.641845] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:50.811 [2024-12-06 17:54:32.641871] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:50.811 [2024-12-06 17:54:32.641887] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:50.811 [2024-12-06 17:54:32.641900] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:50.811 [2024-12-06 17:54:32.641931] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:50.811 qpair failed and we were unable to recover it. 00:38:51.070 [2024-12-06 17:54:32.651787] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.070 [2024-12-06 17:54:32.651902] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.070 [2024-12-06 17:54:32.651931] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.070 [2024-12-06 17:54:32.651950] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.070 [2024-12-06 17:54:32.651963] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.070 [2024-12-06 17:54:32.652009] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.070 qpair failed and we were unable to recover it. 00:38:51.070 [2024-12-06 17:54:32.661833] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.070 [2024-12-06 17:54:32.661924] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.070 [2024-12-06 17:54:32.661949] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.070 [2024-12-06 17:54:32.661964] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.070 [2024-12-06 17:54:32.661977] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.070 [2024-12-06 17:54:32.662007] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.070 qpair failed and we were unable to recover it. 00:38:51.070 [2024-12-06 17:54:32.671793] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.070 [2024-12-06 17:54:32.671887] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.070 [2024-12-06 17:54:32.671913] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.070 [2024-12-06 17:54:32.671928] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.070 [2024-12-06 17:54:32.671941] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.070 [2024-12-06 17:54:32.671972] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.070 qpair failed and we were unable to recover it. 00:38:51.070 [2024-12-06 17:54:32.681852] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.070 [2024-12-06 17:54:32.681940] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.070 [2024-12-06 17:54:32.681970] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.070 [2024-12-06 17:54:32.681986] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.070 [2024-12-06 17:54:32.681999] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.070 [2024-12-06 17:54:32.682029] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.070 qpair failed and we were unable to recover it. 00:38:51.070 [2024-12-06 17:54:32.691868] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.070 [2024-12-06 17:54:32.691962] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.070 [2024-12-06 17:54:32.691986] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.070 [2024-12-06 17:54:32.692001] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.070 [2024-12-06 17:54:32.692014] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.070 [2024-12-06 17:54:32.692044] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.070 qpair failed and we were unable to recover it. 00:38:51.070 [2024-12-06 17:54:32.701902] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.070 [2024-12-06 17:54:32.701991] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.070 [2024-12-06 17:54:32.702015] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.070 [2024-12-06 17:54:32.702030] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.070 [2024-12-06 17:54:32.702043] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.070 [2024-12-06 17:54:32.702072] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.070 qpair failed and we were unable to recover it. 00:38:51.070 [2024-12-06 17:54:32.711966] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.070 [2024-12-06 17:54:32.712051] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.070 [2024-12-06 17:54:32.712075] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.070 [2024-12-06 17:54:32.712090] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.070 [2024-12-06 17:54:32.712103] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.070 [2024-12-06 17:54:32.712133] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.070 qpair failed and we were unable to recover it. 00:38:51.070 [2024-12-06 17:54:32.721982] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.070 [2024-12-06 17:54:32.722079] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.070 [2024-12-06 17:54:32.722106] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.070 [2024-12-06 17:54:32.722127] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.070 [2024-12-06 17:54:32.722140] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.070 [2024-12-06 17:54:32.722186] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.070 qpair failed and we were unable to recover it. 00:38:51.070 [2024-12-06 17:54:32.732008] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.070 [2024-12-06 17:54:32.732100] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.070 [2024-12-06 17:54:32.732124] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.070 [2024-12-06 17:54:32.732138] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.070 [2024-12-06 17:54:32.732151] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.070 [2024-12-06 17:54:32.732182] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.070 qpair failed and we were unable to recover it. 00:38:51.070 [2024-12-06 17:54:32.742007] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.070 [2024-12-06 17:54:32.742098] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.070 [2024-12-06 17:54:32.742128] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.070 [2024-12-06 17:54:32.742144] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.070 [2024-12-06 17:54:32.742158] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.070 [2024-12-06 17:54:32.742187] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.070 qpair failed and we were unable to recover it. 00:38:51.070 [2024-12-06 17:54:32.752011] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.070 [2024-12-06 17:54:32.752097] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.070 [2024-12-06 17:54:32.752122] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.070 [2024-12-06 17:54:32.752136] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.071 [2024-12-06 17:54:32.752150] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.071 [2024-12-06 17:54:32.752180] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.071 qpair failed and we were unable to recover it. 00:38:51.071 [2024-12-06 17:54:32.762114] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.071 [2024-12-06 17:54:32.762223] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.071 [2024-12-06 17:54:32.762250] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.071 [2024-12-06 17:54:32.762266] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.071 [2024-12-06 17:54:32.762279] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.071 [2024-12-06 17:54:32.762323] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.071 qpair failed and we were unable to recover it. 00:38:51.071 [2024-12-06 17:54:32.772096] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.071 [2024-12-06 17:54:32.772188] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.071 [2024-12-06 17:54:32.772212] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.071 [2024-12-06 17:54:32.772228] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.071 [2024-12-06 17:54:32.772241] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.071 [2024-12-06 17:54:32.772271] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.071 qpair failed and we were unable to recover it. 00:38:51.071 [2024-12-06 17:54:32.782113] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.071 [2024-12-06 17:54:32.782201] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.071 [2024-12-06 17:54:32.782226] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.071 [2024-12-06 17:54:32.782240] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.071 [2024-12-06 17:54:32.782253] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.071 [2024-12-06 17:54:32.782283] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.071 qpair failed and we were unable to recover it. 00:38:51.071 [2024-12-06 17:54:32.792171] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.071 [2024-12-06 17:54:32.792274] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.071 [2024-12-06 17:54:32.792301] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.071 [2024-12-06 17:54:32.792317] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.071 [2024-12-06 17:54:32.792329] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.071 [2024-12-06 17:54:32.792359] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.071 qpair failed and we were unable to recover it. 00:38:51.071 [2024-12-06 17:54:32.802188] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.071 [2024-12-06 17:54:32.802303] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.071 [2024-12-06 17:54:32.802329] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.071 [2024-12-06 17:54:32.802344] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.071 [2024-12-06 17:54:32.802357] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.071 [2024-12-06 17:54:32.802387] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.071 qpair failed and we were unable to recover it. 00:38:51.071 [2024-12-06 17:54:32.812229] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.071 [2024-12-06 17:54:32.812352] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.071 [2024-12-06 17:54:32.812379] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.071 [2024-12-06 17:54:32.812394] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.071 [2024-12-06 17:54:32.812407] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.071 [2024-12-06 17:54:32.812437] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.071 qpair failed and we were unable to recover it. 00:38:51.071 [2024-12-06 17:54:32.822281] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.071 [2024-12-06 17:54:32.822373] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.071 [2024-12-06 17:54:32.822398] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.071 [2024-12-06 17:54:32.822413] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.071 [2024-12-06 17:54:32.822425] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.071 [2024-12-06 17:54:32.822471] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.071 qpair failed and we were unable to recover it. 00:38:51.071 [2024-12-06 17:54:32.832280] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.071 [2024-12-06 17:54:32.832387] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.071 [2024-12-06 17:54:32.832416] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.071 [2024-12-06 17:54:32.832432] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.071 [2024-12-06 17:54:32.832445] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.071 [2024-12-06 17:54:32.832476] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.071 qpair failed and we were unable to recover it. 00:38:51.071 [2024-12-06 17:54:32.842299] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.071 [2024-12-06 17:54:32.842394] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.071 [2024-12-06 17:54:32.842419] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.071 [2024-12-06 17:54:32.842435] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.071 [2024-12-06 17:54:32.842448] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.071 [2024-12-06 17:54:32.842479] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.071 qpair failed and we were unable to recover it. 00:38:51.071 [2024-12-06 17:54:32.852436] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.071 [2024-12-06 17:54:32.852566] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.071 [2024-12-06 17:54:32.852592] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.071 [2024-12-06 17:54:32.852615] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.071 [2024-12-06 17:54:32.852630] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.072 [2024-12-06 17:54:32.852661] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.072 qpair failed and we were unable to recover it. 00:38:51.072 [2024-12-06 17:54:32.862318] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.072 [2024-12-06 17:54:32.862412] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.072 [2024-12-06 17:54:32.862437] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.072 [2024-12-06 17:54:32.862452] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.072 [2024-12-06 17:54:32.862465] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.072 [2024-12-06 17:54:32.862495] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.072 qpair failed and we were unable to recover it. 00:38:51.072 [2024-12-06 17:54:32.872362] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.072 [2024-12-06 17:54:32.872446] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.072 [2024-12-06 17:54:32.872471] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.072 [2024-12-06 17:54:32.872486] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.072 [2024-12-06 17:54:32.872499] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.072 [2024-12-06 17:54:32.872529] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.072 qpair failed and we were unable to recover it. 00:38:51.072 [2024-12-06 17:54:32.882409] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.072 [2024-12-06 17:54:32.882504] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.072 [2024-12-06 17:54:32.882530] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.072 [2024-12-06 17:54:32.882545] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.072 [2024-12-06 17:54:32.882558] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.072 [2024-12-06 17:54:32.882589] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.072 qpair failed and we were unable to recover it. 00:38:51.072 [2024-12-06 17:54:32.892441] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.072 [2024-12-06 17:54:32.892537] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.072 [2024-12-06 17:54:32.892563] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.072 [2024-12-06 17:54:32.892579] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.072 [2024-12-06 17:54:32.892593] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.072 [2024-12-06 17:54:32.892630] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.072 qpair failed and we were unable to recover it. 00:38:51.072 [2024-12-06 17:54:32.902435] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.072 [2024-12-06 17:54:32.902562] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.072 [2024-12-06 17:54:32.902589] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.072 [2024-12-06 17:54:32.902605] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.072 [2024-12-06 17:54:32.902618] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.072 [2024-12-06 17:54:32.902649] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.072 qpair failed and we were unable to recover it. 00:38:51.332 [2024-12-06 17:54:32.912506] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.332 [2024-12-06 17:54:32.912603] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.332 [2024-12-06 17:54:32.912630] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.332 [2024-12-06 17:54:32.912645] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.332 [2024-12-06 17:54:32.912658] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.332 [2024-12-06 17:54:32.912699] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.332 qpair failed and we were unable to recover it. 00:38:51.332 [2024-12-06 17:54:32.922509] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.332 [2024-12-06 17:54:32.922604] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.332 [2024-12-06 17:54:32.922629] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.332 [2024-12-06 17:54:32.922645] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.332 [2024-12-06 17:54:32.922658] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.332 [2024-12-06 17:54:32.922698] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.332 qpair failed and we were unable to recover it. 00:38:51.332 [2024-12-06 17:54:32.932528] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.332 [2024-12-06 17:54:32.932626] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.332 [2024-12-06 17:54:32.932651] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.332 [2024-12-06 17:54:32.932674] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.332 [2024-12-06 17:54:32.932690] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe58000b90 00:38:51.332 [2024-12-06 17:54:32.932722] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:38:51.332 qpair failed and we were unable to recover it. 00:38:51.332 [2024-12-06 17:54:32.942553] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.332 [2024-12-06 17:54:32.942646] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.332 [2024-12-06 17:54:32.942690] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.332 [2024-12-06 17:54:32.942709] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.332 [2024-12-06 17:54:32.942722] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.332 [2024-12-06 17:54:32.942755] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.332 qpair failed and we were unable to recover it. 00:38:51.332 [2024-12-06 17:54:32.952601] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.332 [2024-12-06 17:54:32.952701] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.332 [2024-12-06 17:54:32.952730] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.332 [2024-12-06 17:54:32.952745] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.332 [2024-12-06 17:54:32.952757] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.332 [2024-12-06 17:54:32.952788] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.332 qpair failed and we were unable to recover it. 00:38:51.332 [2024-12-06 17:54:32.962635] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.332 [2024-12-06 17:54:32.962739] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.332 [2024-12-06 17:54:32.962768] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.332 [2024-12-06 17:54:32.962784] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.332 [2024-12-06 17:54:32.962797] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.332 [2024-12-06 17:54:32.962828] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.332 qpair failed and we were unable to recover it. 00:38:51.332 [2024-12-06 17:54:32.972656] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.332 [2024-12-06 17:54:32.972760] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.332 [2024-12-06 17:54:32.972787] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.332 [2024-12-06 17:54:32.972802] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.332 [2024-12-06 17:54:32.972814] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.332 [2024-12-06 17:54:32.972845] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.332 qpair failed and we were unable to recover it. 00:38:51.332 [2024-12-06 17:54:32.982744] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.332 [2024-12-06 17:54:32.982835] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.332 [2024-12-06 17:54:32.982867] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.332 [2024-12-06 17:54:32.982883] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.332 [2024-12-06 17:54:32.982896] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.332 [2024-12-06 17:54:32.982927] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.332 qpair failed and we were unable to recover it. 00:38:51.332 [2024-12-06 17:54:32.992679] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.332 [2024-12-06 17:54:32.992806] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.332 [2024-12-06 17:54:32.992833] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.332 [2024-12-06 17:54:32.992847] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.332 [2024-12-06 17:54:32.992860] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.332 [2024-12-06 17:54:32.992892] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.332 qpair failed and we were unable to recover it. 00:38:51.332 [2024-12-06 17:54:33.002742] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.332 [2024-12-06 17:54:33.002829] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.332 [2024-12-06 17:54:33.002854] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.332 [2024-12-06 17:54:33.002869] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.332 [2024-12-06 17:54:33.002882] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.332 [2024-12-06 17:54:33.002913] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.332 qpair failed and we were unable to recover it. 00:38:51.332 [2024-12-06 17:54:33.012761] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.332 [2024-12-06 17:54:33.012852] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.332 [2024-12-06 17:54:33.012876] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.332 [2024-12-06 17:54:33.012891] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.332 [2024-12-06 17:54:33.012903] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.332 [2024-12-06 17:54:33.012934] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.332 qpair failed and we were unable to recover it. 00:38:51.332 [2024-12-06 17:54:33.022826] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.332 [2024-12-06 17:54:33.022943] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.332 [2024-12-06 17:54:33.022969] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.332 [2024-12-06 17:54:33.022984] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.332 [2024-12-06 17:54:33.023003] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.332 [2024-12-06 17:54:33.023035] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.332 qpair failed and we were unable to recover it. 00:38:51.333 [2024-12-06 17:54:33.032858] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.333 [2024-12-06 17:54:33.032942] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.333 [2024-12-06 17:54:33.032968] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.333 [2024-12-06 17:54:33.032983] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.333 [2024-12-06 17:54:33.032995] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.333 [2024-12-06 17:54:33.033026] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.333 qpair failed and we were unable to recover it. 00:38:51.333 [2024-12-06 17:54:33.042868] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.333 [2024-12-06 17:54:33.042954] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.333 [2024-12-06 17:54:33.042979] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.333 [2024-12-06 17:54:33.042993] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.333 [2024-12-06 17:54:33.043006] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.333 [2024-12-06 17:54:33.043037] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.333 qpair failed and we were unable to recover it. 00:38:51.333 [2024-12-06 17:54:33.052895] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.333 [2024-12-06 17:54:33.052996] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.333 [2024-12-06 17:54:33.053026] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.333 [2024-12-06 17:54:33.053043] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.333 [2024-12-06 17:54:33.053057] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.333 [2024-12-06 17:54:33.053089] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.333 qpair failed and we were unable to recover it. 00:38:51.333 [2024-12-06 17:54:33.062941] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.333 [2024-12-06 17:54:33.063045] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.333 [2024-12-06 17:54:33.063071] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.333 [2024-12-06 17:54:33.063086] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.333 [2024-12-06 17:54:33.063100] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.333 [2024-12-06 17:54:33.063131] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.333 qpair failed and we were unable to recover it. 00:38:51.333 [2024-12-06 17:54:33.073060] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.333 [2024-12-06 17:54:33.073197] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.333 [2024-12-06 17:54:33.073232] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.333 [2024-12-06 17:54:33.073252] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.333 [2024-12-06 17:54:33.073265] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.333 [2024-12-06 17:54:33.073297] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.333 qpair failed and we were unable to recover it. 00:38:51.333 [2024-12-06 17:54:33.082950] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.333 [2024-12-06 17:54:33.083049] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.333 [2024-12-06 17:54:33.083075] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.333 [2024-12-06 17:54:33.083089] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.333 [2024-12-06 17:54:33.083102] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.333 [2024-12-06 17:54:33.083135] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.333 qpair failed and we were unable to recover it. 00:38:51.333 [2024-12-06 17:54:33.093091] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.333 [2024-12-06 17:54:33.093213] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.333 [2024-12-06 17:54:33.093241] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.333 [2024-12-06 17:54:33.093257] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.333 [2024-12-06 17:54:33.093269] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.333 [2024-12-06 17:54:33.093300] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.333 qpair failed and we were unable to recover it. 00:38:51.333 [2024-12-06 17:54:33.103024] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.333 [2024-12-06 17:54:33.103111] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.333 [2024-12-06 17:54:33.103137] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.333 [2024-12-06 17:54:33.103152] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.333 [2024-12-06 17:54:33.103164] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.333 [2024-12-06 17:54:33.103195] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.333 qpair failed and we were unable to recover it. 00:38:51.333 [2024-12-06 17:54:33.113051] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.333 [2024-12-06 17:54:33.113141] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.333 [2024-12-06 17:54:33.113173] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.333 [2024-12-06 17:54:33.113189] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.333 [2024-12-06 17:54:33.113201] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.333 [2024-12-06 17:54:33.113233] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.333 qpair failed and we were unable to recover it. 00:38:51.333 [2024-12-06 17:54:33.123084] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.333 [2024-12-06 17:54:33.123191] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.333 [2024-12-06 17:54:33.123218] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.333 [2024-12-06 17:54:33.123234] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.333 [2024-12-06 17:54:33.123247] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.333 [2024-12-06 17:54:33.123278] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.333 qpair failed and we were unable to recover it. 00:38:51.333 [2024-12-06 17:54:33.133097] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.333 [2024-12-06 17:54:33.133192] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.333 [2024-12-06 17:54:33.133218] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.333 [2024-12-06 17:54:33.133233] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.333 [2024-12-06 17:54:33.133246] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.333 [2024-12-06 17:54:33.133277] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.333 qpair failed and we were unable to recover it. 00:38:51.333 [2024-12-06 17:54:33.143116] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.333 [2024-12-06 17:54:33.143207] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.333 [2024-12-06 17:54:33.143232] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.333 [2024-12-06 17:54:33.143247] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.333 [2024-12-06 17:54:33.143260] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.333 [2024-12-06 17:54:33.143291] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.333 qpair failed and we were unable to recover it. 00:38:51.333 [2024-12-06 17:54:33.153153] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.333 [2024-12-06 17:54:33.153271] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.333 [2024-12-06 17:54:33.153297] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.333 [2024-12-06 17:54:33.153312] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.333 [2024-12-06 17:54:33.153331] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.333 [2024-12-06 17:54:33.153364] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.333 qpair failed and we were unable to recover it. 00:38:51.333 [2024-12-06 17:54:33.163188] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.333 [2024-12-06 17:54:33.163318] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.333 [2024-12-06 17:54:33.163345] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.333 [2024-12-06 17:54:33.163360] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.333 [2024-12-06 17:54:33.163373] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.333 [2024-12-06 17:54:33.163404] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.333 qpair failed and we were unable to recover it. 00:38:51.594 [2024-12-06 17:54:33.173310] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.594 [2024-12-06 17:54:33.173405] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.594 [2024-12-06 17:54:33.173430] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.594 [2024-12-06 17:54:33.173445] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.594 [2024-12-06 17:54:33.173457] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.594 [2024-12-06 17:54:33.173487] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.594 qpair failed and we were unable to recover it. 00:38:51.594 [2024-12-06 17:54:33.183287] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.594 [2024-12-06 17:54:33.183396] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.594 [2024-12-06 17:54:33.183421] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.594 [2024-12-06 17:54:33.183436] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.594 [2024-12-06 17:54:33.183449] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.594 [2024-12-06 17:54:33.183479] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.594 qpair failed and we were unable to recover it. 00:38:51.594 [2024-12-06 17:54:33.193291] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.594 [2024-12-06 17:54:33.193421] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.594 [2024-12-06 17:54:33.193448] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.594 [2024-12-06 17:54:33.193463] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.594 [2024-12-06 17:54:33.193476] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.594 [2024-12-06 17:54:33.193506] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.594 qpair failed and we were unable to recover it. 00:38:51.594 [2024-12-06 17:54:33.203322] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.594 [2024-12-06 17:54:33.203410] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.594 [2024-12-06 17:54:33.203436] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.594 [2024-12-06 17:54:33.203451] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.594 [2024-12-06 17:54:33.203464] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.594 [2024-12-06 17:54:33.203494] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.594 qpair failed and we were unable to recover it. 00:38:51.594 [2024-12-06 17:54:33.213372] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.594 [2024-12-06 17:54:33.213501] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.594 [2024-12-06 17:54:33.213528] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.594 [2024-12-06 17:54:33.213543] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.594 [2024-12-06 17:54:33.213555] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.594 [2024-12-06 17:54:33.213586] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.594 qpair failed and we were unable to recover it. 00:38:51.594 [2024-12-06 17:54:33.223401] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.594 [2024-12-06 17:54:33.223490] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.594 [2024-12-06 17:54:33.223516] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.594 [2024-12-06 17:54:33.223531] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.594 [2024-12-06 17:54:33.223543] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.594 [2024-12-06 17:54:33.223574] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.594 qpair failed and we were unable to recover it. 00:38:51.594 [2024-12-06 17:54:33.233427] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.594 [2024-12-06 17:54:33.233508] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.594 [2024-12-06 17:54:33.233534] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.595 [2024-12-06 17:54:33.233549] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.595 [2024-12-06 17:54:33.233561] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.595 [2024-12-06 17:54:33.233604] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.595 qpair failed and we were unable to recover it. 00:38:51.595 [2024-12-06 17:54:33.243403] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.595 [2024-12-06 17:54:33.243527] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.595 [2024-12-06 17:54:33.243555] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.595 [2024-12-06 17:54:33.243570] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.595 [2024-12-06 17:54:33.243582] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.595 [2024-12-06 17:54:33.243613] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.595 qpair failed and we were unable to recover it. 00:38:51.595 [2024-12-06 17:54:33.253469] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.595 [2024-12-06 17:54:33.253564] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.595 [2024-12-06 17:54:33.253589] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.595 [2024-12-06 17:54:33.253603] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.595 [2024-12-06 17:54:33.253616] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.595 [2024-12-06 17:54:33.253647] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.595 qpair failed and we were unable to recover it. 00:38:51.595 [2024-12-06 17:54:33.263479] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.595 [2024-12-06 17:54:33.263572] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.595 [2024-12-06 17:54:33.263596] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.595 [2024-12-06 17:54:33.263611] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.595 [2024-12-06 17:54:33.263624] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.595 [2024-12-06 17:54:33.263654] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.595 qpair failed and we were unable to recover it. 00:38:51.595 [2024-12-06 17:54:33.273554] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.595 [2024-12-06 17:54:33.273639] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.595 [2024-12-06 17:54:33.273671] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.595 [2024-12-06 17:54:33.273689] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.595 [2024-12-06 17:54:33.273702] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.595 [2024-12-06 17:54:33.273733] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.595 qpair failed and we were unable to recover it. 00:38:51.595 [2024-12-06 17:54:33.283614] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.595 [2024-12-06 17:54:33.283708] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.595 [2024-12-06 17:54:33.283735] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.595 [2024-12-06 17:54:33.283756] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.595 [2024-12-06 17:54:33.283769] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.595 [2024-12-06 17:54:33.283800] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.595 qpair failed and we were unable to recover it. 00:38:51.595 [2024-12-06 17:54:33.293637] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.595 [2024-12-06 17:54:33.293742] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.595 [2024-12-06 17:54:33.293769] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.595 [2024-12-06 17:54:33.293785] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.595 [2024-12-06 17:54:33.293798] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.595 [2024-12-06 17:54:33.293828] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.595 qpair failed and we were unable to recover it. 00:38:51.595 [2024-12-06 17:54:33.303622] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.595 [2024-12-06 17:54:33.303732] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.595 [2024-12-06 17:54:33.303759] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.595 [2024-12-06 17:54:33.303775] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.595 [2024-12-06 17:54:33.303787] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.595 [2024-12-06 17:54:33.303818] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.595 qpair failed and we were unable to recover it. 00:38:51.595 [2024-12-06 17:54:33.313614] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.595 [2024-12-06 17:54:33.313709] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.595 [2024-12-06 17:54:33.313734] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.595 [2024-12-06 17:54:33.313749] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.595 [2024-12-06 17:54:33.313763] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.595 [2024-12-06 17:54:33.313795] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.595 qpair failed and we were unable to recover it. 00:38:51.595 [2024-12-06 17:54:33.323632] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.595 [2024-12-06 17:54:33.323780] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.595 [2024-12-06 17:54:33.323813] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.595 [2024-12-06 17:54:33.323830] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.595 [2024-12-06 17:54:33.323844] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.595 [2024-12-06 17:54:33.323881] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.595 qpair failed and we were unable to recover it. 00:38:51.595 [2024-12-06 17:54:33.333703] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.595 [2024-12-06 17:54:33.333806] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.595 [2024-12-06 17:54:33.333832] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.595 [2024-12-06 17:54:33.333847] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.595 [2024-12-06 17:54:33.333860] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.595 [2024-12-06 17:54:33.333893] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.595 qpair failed and we were unable to recover it. 00:38:51.595 [2024-12-06 17:54:33.343790] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.595 [2024-12-06 17:54:33.343877] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.595 [2024-12-06 17:54:33.343904] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.595 [2024-12-06 17:54:33.343919] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.595 [2024-12-06 17:54:33.343932] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.595 [2024-12-06 17:54:33.343963] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.595 qpair failed and we were unable to recover it. 00:38:51.595 [2024-12-06 17:54:33.353748] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.595 [2024-12-06 17:54:33.353832] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.595 [2024-12-06 17:54:33.353857] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.595 [2024-12-06 17:54:33.353872] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.595 [2024-12-06 17:54:33.353885] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.595 [2024-12-06 17:54:33.353916] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.595 qpair failed and we were unable to recover it. 00:38:51.595 [2024-12-06 17:54:33.363750] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.595 [2024-12-06 17:54:33.363859] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.595 [2024-12-06 17:54:33.363885] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.595 [2024-12-06 17:54:33.363900] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.595 [2024-12-06 17:54:33.363913] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.595 [2024-12-06 17:54:33.363944] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.595 qpair failed and we were unable to recover it. 00:38:51.595 [2024-12-06 17:54:33.373790] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.595 [2024-12-06 17:54:33.373894] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.595 [2024-12-06 17:54:33.373920] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.595 [2024-12-06 17:54:33.373935] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.595 [2024-12-06 17:54:33.373948] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.595 [2024-12-06 17:54:33.373979] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.595 qpair failed and we were unable to recover it. 00:38:51.595 [2024-12-06 17:54:33.383866] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.595 [2024-12-06 17:54:33.383979] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.595 [2024-12-06 17:54:33.384005] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.595 [2024-12-06 17:54:33.384020] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.595 [2024-12-06 17:54:33.384033] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.595 [2024-12-06 17:54:33.384079] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.595 qpair failed and we were unable to recover it. 00:38:51.595 [2024-12-06 17:54:33.393881] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.595 [2024-12-06 17:54:33.393971] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.595 [2024-12-06 17:54:33.393997] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.595 [2024-12-06 17:54:33.394011] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.595 [2024-12-06 17:54:33.394024] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.595 [2024-12-06 17:54:33.394055] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.595 qpair failed and we were unable to recover it. 00:38:51.596 [2024-12-06 17:54:33.403914] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.596 [2024-12-06 17:54:33.404029] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.596 [2024-12-06 17:54:33.404054] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.596 [2024-12-06 17:54:33.404070] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.596 [2024-12-06 17:54:33.404083] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.596 [2024-12-06 17:54:33.404127] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.596 qpair failed and we were unable to recover it. 00:38:51.596 [2024-12-06 17:54:33.413940] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.596 [2024-12-06 17:54:33.414030] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.596 [2024-12-06 17:54:33.414056] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.596 [2024-12-06 17:54:33.414076] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.596 [2024-12-06 17:54:33.414090] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.596 [2024-12-06 17:54:33.414120] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.596 qpair failed and we were unable to recover it. 00:38:51.596 [2024-12-06 17:54:33.423934] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.596 [2024-12-06 17:54:33.424024] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.596 [2024-12-06 17:54:33.424049] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.596 [2024-12-06 17:54:33.424064] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.596 [2024-12-06 17:54:33.424077] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.596 [2024-12-06 17:54:33.424108] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.596 qpair failed and we were unable to recover it. 00:38:51.853 [2024-12-06 17:54:33.433956] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.853 [2024-12-06 17:54:33.434043] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.853 [2024-12-06 17:54:33.434068] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.853 [2024-12-06 17:54:33.434083] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.853 [2024-12-06 17:54:33.434096] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.853 [2024-12-06 17:54:33.434127] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.853 qpair failed and we were unable to recover it. 00:38:51.853 [2024-12-06 17:54:33.444015] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.853 [2024-12-06 17:54:33.444108] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.853 [2024-12-06 17:54:33.444134] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.853 [2024-12-06 17:54:33.444149] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.853 [2024-12-06 17:54:33.444162] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.853 [2024-12-06 17:54:33.444205] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.853 qpair failed and we were unable to recover it. 00:38:51.853 [2024-12-06 17:54:33.454045] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.853 [2024-12-06 17:54:33.454139] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.853 [2024-12-06 17:54:33.454168] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.853 [2024-12-06 17:54:33.454185] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.853 [2024-12-06 17:54:33.454198] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.853 [2024-12-06 17:54:33.454236] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.853 qpair failed and we were unable to recover it. 00:38:51.853 [2024-12-06 17:54:33.464057] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.853 [2024-12-06 17:54:33.464140] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.853 [2024-12-06 17:54:33.464166] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.853 [2024-12-06 17:54:33.464181] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.853 [2024-12-06 17:54:33.464194] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.853 [2024-12-06 17:54:33.464224] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.853 qpair failed and we were unable to recover it. 00:38:51.853 [2024-12-06 17:54:33.474068] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.853 [2024-12-06 17:54:33.474154] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.853 [2024-12-06 17:54:33.474180] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.853 [2024-12-06 17:54:33.474195] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.853 [2024-12-06 17:54:33.474208] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.853 [2024-12-06 17:54:33.474237] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.853 qpair failed and we were unable to recover it. 00:38:51.853 [2024-12-06 17:54:33.484238] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.853 [2024-12-06 17:54:33.484325] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.853 [2024-12-06 17:54:33.484350] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.853 [2024-12-06 17:54:33.484365] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.853 [2024-12-06 17:54:33.484378] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.853 [2024-12-06 17:54:33.484409] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.853 qpair failed and we were unable to recover it. 00:38:51.853 [2024-12-06 17:54:33.494152] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.853 [2024-12-06 17:54:33.494242] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.853 [2024-12-06 17:54:33.494267] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.853 [2024-12-06 17:54:33.494282] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.853 [2024-12-06 17:54:33.494295] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.853 [2024-12-06 17:54:33.494325] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.853 qpair failed and we were unable to recover it. 00:38:51.853 [2024-12-06 17:54:33.504150] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.853 [2024-12-06 17:54:33.504239] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.853 [2024-12-06 17:54:33.504264] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.853 [2024-12-06 17:54:33.504279] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.853 [2024-12-06 17:54:33.504292] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.853 [2024-12-06 17:54:33.504336] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.853 qpair failed and we were unable to recover it. 00:38:51.853 [2024-12-06 17:54:33.514158] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.853 [2024-12-06 17:54:33.514240] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.853 [2024-12-06 17:54:33.514265] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.853 [2024-12-06 17:54:33.514280] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.853 [2024-12-06 17:54:33.514294] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.853 [2024-12-06 17:54:33.514325] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.853 qpair failed and we were unable to recover it. 00:38:51.853 [2024-12-06 17:54:33.524239] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.853 [2024-12-06 17:54:33.524328] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.853 [2024-12-06 17:54:33.524354] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.853 [2024-12-06 17:54:33.524370] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.853 [2024-12-06 17:54:33.524383] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.853 [2024-12-06 17:54:33.524414] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.853 qpair failed and we were unable to recover it. 00:38:51.853 [2024-12-06 17:54:33.534239] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.854 [2024-12-06 17:54:33.534331] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.854 [2024-12-06 17:54:33.534356] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.854 [2024-12-06 17:54:33.534371] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.854 [2024-12-06 17:54:33.534384] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.854 [2024-12-06 17:54:33.534415] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.854 qpair failed and we were unable to recover it. 00:38:51.854 [2024-12-06 17:54:33.544290] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.854 [2024-12-06 17:54:33.544375] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.854 [2024-12-06 17:54:33.544410] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.854 [2024-12-06 17:54:33.544429] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.854 [2024-12-06 17:54:33.544441] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.854 [2024-12-06 17:54:33.544473] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.854 qpair failed and we were unable to recover it. 00:38:51.854 [2024-12-06 17:54:33.554320] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.854 [2024-12-06 17:54:33.554405] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.854 [2024-12-06 17:54:33.554435] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.854 [2024-12-06 17:54:33.554451] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.854 [2024-12-06 17:54:33.554464] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.854 [2024-12-06 17:54:33.554496] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.854 qpair failed and we were unable to recover it. 00:38:51.854 [2024-12-06 17:54:33.564341] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.854 [2024-12-06 17:54:33.564476] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.854 [2024-12-06 17:54:33.564501] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.854 [2024-12-06 17:54:33.564518] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.854 [2024-12-06 17:54:33.564531] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.854 [2024-12-06 17:54:33.564576] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.854 qpair failed and we were unable to recover it. 00:38:51.854 [2024-12-06 17:54:33.574403] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.854 [2024-12-06 17:54:33.574546] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.854 [2024-12-06 17:54:33.574579] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.854 [2024-12-06 17:54:33.574618] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.854 [2024-12-06 17:54:33.574642] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.854 [2024-12-06 17:54:33.574708] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.854 qpair failed and we were unable to recover it. 00:38:51.854 [2024-12-06 17:54:33.584397] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.854 [2024-12-06 17:54:33.584493] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.854 [2024-12-06 17:54:33.584520] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.854 [2024-12-06 17:54:33.584535] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.854 [2024-12-06 17:54:33.584554] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.854 [2024-12-06 17:54:33.584587] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.854 qpair failed and we were unable to recover it. 00:38:51.854 [2024-12-06 17:54:33.594401] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.854 [2024-12-06 17:54:33.594518] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.854 [2024-12-06 17:54:33.594544] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.854 [2024-12-06 17:54:33.594559] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.854 [2024-12-06 17:54:33.594573] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.854 [2024-12-06 17:54:33.594604] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.854 qpair failed and we were unable to recover it. 00:38:51.854 [2024-12-06 17:54:33.604424] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.854 [2024-12-06 17:54:33.604554] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.854 [2024-12-06 17:54:33.604583] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.854 [2024-12-06 17:54:33.604599] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.854 [2024-12-06 17:54:33.604612] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.854 [2024-12-06 17:54:33.604643] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.854 qpair failed and we were unable to recover it. 00:38:51.854 [2024-12-06 17:54:33.614533] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.854 [2024-12-06 17:54:33.614629] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.854 [2024-12-06 17:54:33.614655] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.854 [2024-12-06 17:54:33.614680] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.854 [2024-12-06 17:54:33.614695] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.854 [2024-12-06 17:54:33.614726] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.854 qpair failed and we were unable to recover it. 00:38:51.854 [2024-12-06 17:54:33.624503] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.854 [2024-12-06 17:54:33.624627] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.854 [2024-12-06 17:54:33.624653] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.854 [2024-12-06 17:54:33.624676] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.854 [2024-12-06 17:54:33.624691] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.854 [2024-12-06 17:54:33.624725] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.854 qpair failed and we were unable to recover it. 00:38:51.854 [2024-12-06 17:54:33.634526] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.854 [2024-12-06 17:54:33.634610] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.854 [2024-12-06 17:54:33.634635] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.854 [2024-12-06 17:54:33.634650] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.854 [2024-12-06 17:54:33.634673] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.854 [2024-12-06 17:54:33.634715] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.854 qpair failed and we were unable to recover it. 00:38:51.854 [2024-12-06 17:54:33.644545] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.854 [2024-12-06 17:54:33.644629] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.854 [2024-12-06 17:54:33.644655] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.854 [2024-12-06 17:54:33.644679] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.854 [2024-12-06 17:54:33.644693] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.854 [2024-12-06 17:54:33.644724] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.854 qpair failed and we were unable to recover it. 00:38:51.854 [2024-12-06 17:54:33.654589] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.854 [2024-12-06 17:54:33.654714] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.854 [2024-12-06 17:54:33.654740] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.854 [2024-12-06 17:54:33.654755] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.854 [2024-12-06 17:54:33.654767] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.854 [2024-12-06 17:54:33.654798] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.854 qpair failed and we were unable to recover it. 00:38:51.854 [2024-12-06 17:54:33.664631] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.854 [2024-12-06 17:54:33.664749] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.854 [2024-12-06 17:54:33.664775] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.854 [2024-12-06 17:54:33.664790] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.854 [2024-12-06 17:54:33.664803] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.854 [2024-12-06 17:54:33.664847] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.854 qpair failed and we were unable to recover it. 00:38:51.854 [2024-12-06 17:54:33.674652] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.854 [2024-12-06 17:54:33.674768] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.854 [2024-12-06 17:54:33.674799] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.854 [2024-12-06 17:54:33.674814] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.854 [2024-12-06 17:54:33.674828] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.854 [2024-12-06 17:54:33.674859] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.854 qpair failed and we were unable to recover it. 00:38:51.854 [2024-12-06 17:54:33.684757] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:51.854 [2024-12-06 17:54:33.684840] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:51.854 [2024-12-06 17:54:33.684866] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:51.854 [2024-12-06 17:54:33.684881] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:51.854 [2024-12-06 17:54:33.684894] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:51.854 [2024-12-06 17:54:33.684925] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:51.854 qpair failed and we were unable to recover it. 00:38:52.112 [2024-12-06 17:54:33.694761] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.112 [2024-12-06 17:54:33.694851] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.112 [2024-12-06 17:54:33.694876] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.112 [2024-12-06 17:54:33.694891] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.112 [2024-12-06 17:54:33.694903] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.112 [2024-12-06 17:54:33.694934] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.112 qpair failed and we were unable to recover it. 00:38:52.112 [2024-12-06 17:54:33.704749] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.112 [2024-12-06 17:54:33.704839] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.112 [2024-12-06 17:54:33.704870] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.112 [2024-12-06 17:54:33.704887] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.112 [2024-12-06 17:54:33.704902] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.112 [2024-12-06 17:54:33.704939] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.112 qpair failed and we were unable to recover it. 00:38:52.112 [2024-12-06 17:54:33.714743] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.112 [2024-12-06 17:54:33.714830] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.112 [2024-12-06 17:54:33.714855] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.112 [2024-12-06 17:54:33.714870] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.112 [2024-12-06 17:54:33.714889] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.112 [2024-12-06 17:54:33.714922] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.112 qpair failed and we were unable to recover it. 00:38:52.112 [2024-12-06 17:54:33.724768] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.112 [2024-12-06 17:54:33.724854] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.112 [2024-12-06 17:54:33.724880] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.112 [2024-12-06 17:54:33.724894] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.112 [2024-12-06 17:54:33.724908] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.112 [2024-12-06 17:54:33.724939] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.112 qpair failed and we were unable to recover it. 00:38:52.112 [2024-12-06 17:54:33.734822] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.112 [2024-12-06 17:54:33.734924] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.112 [2024-12-06 17:54:33.734952] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.112 [2024-12-06 17:54:33.734968] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.112 [2024-12-06 17:54:33.734990] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.112 [2024-12-06 17:54:33.735023] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.112 qpair failed and we were unable to recover it. 00:38:52.112 [2024-12-06 17:54:33.744862] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.112 [2024-12-06 17:54:33.744946] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.112 [2024-12-06 17:54:33.744971] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.112 [2024-12-06 17:54:33.744996] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.112 [2024-12-06 17:54:33.745009] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.112 [2024-12-06 17:54:33.745040] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.112 qpair failed and we were unable to recover it. 00:38:52.112 [2024-12-06 17:54:33.754901] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.112 [2024-12-06 17:54:33.754993] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.112 [2024-12-06 17:54:33.755019] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.112 [2024-12-06 17:54:33.755033] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.112 [2024-12-06 17:54:33.755047] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.112 [2024-12-06 17:54:33.755077] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.112 qpair failed and we were unable to recover it. 00:38:52.112 [2024-12-06 17:54:33.764980] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.112 [2024-12-06 17:54:33.765064] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.112 [2024-12-06 17:54:33.765089] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.112 [2024-12-06 17:54:33.765104] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.112 [2024-12-06 17:54:33.765116] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.112 [2024-12-06 17:54:33.765148] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.112 qpair failed and we were unable to recover it. 00:38:52.112 [2024-12-06 17:54:33.774969] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.112 [2024-12-06 17:54:33.775062] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.112 [2024-12-06 17:54:33.775087] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.112 [2024-12-06 17:54:33.775102] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.112 [2024-12-06 17:54:33.775116] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.112 [2024-12-06 17:54:33.775147] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.112 qpair failed and we were unable to recover it. 00:38:52.112 [2024-12-06 17:54:33.784954] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.112 [2024-12-06 17:54:33.785038] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.112 [2024-12-06 17:54:33.785064] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.112 [2024-12-06 17:54:33.785079] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.112 [2024-12-06 17:54:33.785092] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.112 [2024-12-06 17:54:33.785123] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.112 qpair failed and we were unable to recover it. 00:38:52.112 [2024-12-06 17:54:33.795024] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.112 [2024-12-06 17:54:33.795116] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.112 [2024-12-06 17:54:33.795142] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.112 [2024-12-06 17:54:33.795157] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.112 [2024-12-06 17:54:33.795170] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.112 [2024-12-06 17:54:33.795201] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.112 qpair failed and we were unable to recover it. 00:38:52.112 [2024-12-06 17:54:33.805019] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.112 [2024-12-06 17:54:33.805107] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.112 [2024-12-06 17:54:33.805132] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.112 [2024-12-06 17:54:33.805148] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.112 [2024-12-06 17:54:33.805161] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.113 [2024-12-06 17:54:33.805192] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.113 qpair failed and we were unable to recover it. 00:38:52.113 [2024-12-06 17:54:33.815056] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.113 [2024-12-06 17:54:33.815146] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.113 [2024-12-06 17:54:33.815170] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.113 [2024-12-06 17:54:33.815186] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.113 [2024-12-06 17:54:33.815199] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.113 [2024-12-06 17:54:33.815230] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.113 qpair failed and we were unable to recover it. 00:38:52.113 [2024-12-06 17:54:33.825109] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.113 [2024-12-06 17:54:33.825230] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.113 [2024-12-06 17:54:33.825260] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.113 [2024-12-06 17:54:33.825285] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.113 [2024-12-06 17:54:33.825308] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.113 [2024-12-06 17:54:33.825350] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.113 qpair failed and we were unable to recover it. 00:38:52.113 [2024-12-06 17:54:33.835103] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.113 [2024-12-06 17:54:33.835187] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.113 [2024-12-06 17:54:33.835214] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.113 [2024-12-06 17:54:33.835230] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.113 [2024-12-06 17:54:33.835243] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.113 [2024-12-06 17:54:33.835274] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.113 qpair failed and we were unable to recover it. 00:38:52.113 [2024-12-06 17:54:33.845141] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.113 [2024-12-06 17:54:33.845227] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.113 [2024-12-06 17:54:33.845253] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.113 [2024-12-06 17:54:33.845274] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.113 [2024-12-06 17:54:33.845289] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.113 [2024-12-06 17:54:33.845319] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.113 qpair failed and we were unable to recover it. 00:38:52.113 [2024-12-06 17:54:33.855168] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.113 [2024-12-06 17:54:33.855259] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.113 [2024-12-06 17:54:33.855285] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.113 [2024-12-06 17:54:33.855300] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.113 [2024-12-06 17:54:33.855313] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.113 [2024-12-06 17:54:33.855344] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.113 qpair failed and we were unable to recover it. 00:38:52.113 [2024-12-06 17:54:33.865223] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.113 [2024-12-06 17:54:33.865334] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.113 [2024-12-06 17:54:33.865359] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.113 [2024-12-06 17:54:33.865374] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.113 [2024-12-06 17:54:33.865388] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.113 [2024-12-06 17:54:33.865420] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.113 qpair failed and we were unable to recover it. 00:38:52.113 [2024-12-06 17:54:33.875214] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.113 [2024-12-06 17:54:33.875308] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.113 [2024-12-06 17:54:33.875333] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.113 [2024-12-06 17:54:33.875348] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.113 [2024-12-06 17:54:33.875361] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.113 [2024-12-06 17:54:33.875391] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.113 qpair failed and we were unable to recover it. 00:38:52.113 [2024-12-06 17:54:33.885232] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.113 [2024-12-06 17:54:33.885334] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.113 [2024-12-06 17:54:33.885360] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.113 [2024-12-06 17:54:33.885374] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.113 [2024-12-06 17:54:33.885387] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.113 [2024-12-06 17:54:33.885425] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.113 qpair failed and we were unable to recover it. 00:38:52.113 [2024-12-06 17:54:33.895300] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.113 [2024-12-06 17:54:33.895389] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.113 [2024-12-06 17:54:33.895415] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.113 [2024-12-06 17:54:33.895429] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.113 [2024-12-06 17:54:33.895442] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.113 [2024-12-06 17:54:33.895474] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.113 qpair failed and we were unable to recover it. 00:38:52.113 [2024-12-06 17:54:33.905324] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.113 [2024-12-06 17:54:33.905407] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.113 [2024-12-06 17:54:33.905434] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.113 [2024-12-06 17:54:33.905449] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.113 [2024-12-06 17:54:33.905462] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.113 [2024-12-06 17:54:33.905493] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.113 qpair failed and we were unable to recover it. 00:38:52.113 [2024-12-06 17:54:33.915356] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.113 [2024-12-06 17:54:33.915482] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.113 [2024-12-06 17:54:33.915508] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.113 [2024-12-06 17:54:33.915522] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.113 [2024-12-06 17:54:33.915536] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.113 [2024-12-06 17:54:33.915566] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.113 qpair failed and we were unable to recover it. 00:38:52.113 [2024-12-06 17:54:33.925380] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.113 [2024-12-06 17:54:33.925462] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.113 [2024-12-06 17:54:33.925489] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.113 [2024-12-06 17:54:33.925504] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.113 [2024-12-06 17:54:33.925517] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.113 [2024-12-06 17:54:33.925548] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.113 qpair failed and we were unable to recover it. 00:38:52.113 [2024-12-06 17:54:33.935438] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.113 [2024-12-06 17:54:33.935544] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.113 [2024-12-06 17:54:33.935569] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.113 [2024-12-06 17:54:33.935583] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.113 [2024-12-06 17:54:33.935596] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.114 [2024-12-06 17:54:33.935627] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.114 qpair failed and we were unable to recover it. 00:38:52.114 [2024-12-06 17:54:33.945434] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.114 [2024-12-06 17:54:33.945544] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.114 [2024-12-06 17:54:33.945570] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.114 [2024-12-06 17:54:33.945585] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.114 [2024-12-06 17:54:33.945598] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.114 [2024-12-06 17:54:33.945630] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.114 qpair failed and we were unable to recover it. 00:38:52.395 [2024-12-06 17:54:33.955505] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.395 [2024-12-06 17:54:33.955590] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.395 [2024-12-06 17:54:33.955616] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.395 [2024-12-06 17:54:33.955631] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.395 [2024-12-06 17:54:33.955644] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.395 [2024-12-06 17:54:33.955684] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.395 qpair failed and we were unable to recover it. 00:38:52.395 [2024-12-06 17:54:33.965522] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.395 [2024-12-06 17:54:33.965614] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.395 [2024-12-06 17:54:33.965640] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.395 [2024-12-06 17:54:33.965655] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.395 [2024-12-06 17:54:33.965676] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.395 [2024-12-06 17:54:33.965709] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.395 qpair failed and we were unable to recover it. 00:38:52.395 [2024-12-06 17:54:33.975531] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.395 [2024-12-06 17:54:33.975622] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.395 [2024-12-06 17:54:33.975653] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.395 [2024-12-06 17:54:33.975676] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.395 [2024-12-06 17:54:33.975691] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.395 [2024-12-06 17:54:33.975722] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.395 qpair failed and we were unable to recover it. 00:38:52.395 [2024-12-06 17:54:33.985551] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.395 [2024-12-06 17:54:33.985634] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.395 [2024-12-06 17:54:33.985660] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.395 [2024-12-06 17:54:33.985684] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.395 [2024-12-06 17:54:33.985698] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.395 [2024-12-06 17:54:33.985729] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.395 qpair failed and we were unable to recover it. 00:38:52.395 [2024-12-06 17:54:33.995566] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.395 [2024-12-06 17:54:33.995687] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.395 [2024-12-06 17:54:33.995713] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.395 [2024-12-06 17:54:33.995728] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.395 [2024-12-06 17:54:33.995741] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.395 [2024-12-06 17:54:33.995773] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.395 qpair failed and we were unable to recover it. 00:38:52.395 [2024-12-06 17:54:34.005649] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.395 [2024-12-06 17:54:34.005753] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.395 [2024-12-06 17:54:34.005778] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.395 [2024-12-06 17:54:34.005793] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.395 [2024-12-06 17:54:34.005806] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.395 [2024-12-06 17:54:34.005850] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.395 qpair failed and we were unable to recover it. 00:38:52.395 [2024-12-06 17:54:34.015655] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.395 [2024-12-06 17:54:34.015766] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.395 [2024-12-06 17:54:34.015791] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.395 [2024-12-06 17:54:34.015806] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.395 [2024-12-06 17:54:34.015819] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.395 [2024-12-06 17:54:34.015856] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.395 qpair failed and we were unable to recover it. 00:38:52.395 [2024-12-06 17:54:34.025649] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.395 [2024-12-06 17:54:34.025744] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.395 [2024-12-06 17:54:34.025770] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.395 [2024-12-06 17:54:34.025785] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.395 [2024-12-06 17:54:34.025798] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.395 [2024-12-06 17:54:34.025829] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.395 qpair failed and we were unable to recover it. 00:38:52.395 [2024-12-06 17:54:34.035684] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.395 [2024-12-06 17:54:34.035769] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.395 [2024-12-06 17:54:34.035794] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.395 [2024-12-06 17:54:34.035809] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.395 [2024-12-06 17:54:34.035822] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.395 [2024-12-06 17:54:34.035853] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.395 qpair failed and we were unable to recover it. 00:38:52.395 [2024-12-06 17:54:34.045695] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.395 [2024-12-06 17:54:34.045781] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.396 [2024-12-06 17:54:34.045806] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.396 [2024-12-06 17:54:34.045821] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.396 [2024-12-06 17:54:34.045834] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.396 [2024-12-06 17:54:34.045865] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.396 qpair failed and we were unable to recover it. 00:38:52.396 [2024-12-06 17:54:34.055749] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.396 [2024-12-06 17:54:34.055836] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.396 [2024-12-06 17:54:34.055862] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.396 [2024-12-06 17:54:34.055877] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.396 [2024-12-06 17:54:34.055890] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.396 [2024-12-06 17:54:34.055933] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.396 qpair failed and we were unable to recover it. 00:38:52.396 [2024-12-06 17:54:34.065759] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.396 [2024-12-06 17:54:34.065847] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.396 [2024-12-06 17:54:34.065873] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.396 [2024-12-06 17:54:34.065888] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.396 [2024-12-06 17:54:34.065901] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.396 [2024-12-06 17:54:34.065932] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.396 qpair failed and we were unable to recover it. 00:38:52.396 [2024-12-06 17:54:34.075784] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.396 [2024-12-06 17:54:34.075868] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.396 [2024-12-06 17:54:34.075894] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.396 [2024-12-06 17:54:34.075910] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.396 [2024-12-06 17:54:34.075923] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.396 [2024-12-06 17:54:34.075962] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.396 qpair failed and we were unable to recover it. 00:38:52.396 [2024-12-06 17:54:34.085830] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.396 [2024-12-06 17:54:34.085917] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.396 [2024-12-06 17:54:34.085944] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.396 [2024-12-06 17:54:34.085959] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.396 [2024-12-06 17:54:34.085972] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.396 [2024-12-06 17:54:34.086003] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.396 qpair failed and we were unable to recover it. 00:38:52.396 [2024-12-06 17:54:34.095862] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.396 [2024-12-06 17:54:34.095949] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.396 [2024-12-06 17:54:34.095975] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.396 [2024-12-06 17:54:34.095991] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.396 [2024-12-06 17:54:34.096004] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.396 [2024-12-06 17:54:34.096047] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.396 qpair failed and we were unable to recover it. 00:38:52.396 [2024-12-06 17:54:34.105873] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.396 [2024-12-06 17:54:34.105959] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.396 [2024-12-06 17:54:34.105990] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.396 [2024-12-06 17:54:34.106006] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.396 [2024-12-06 17:54:34.106018] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.396 [2024-12-06 17:54:34.106049] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.396 qpair failed and we were unable to recover it. 00:38:52.396 [2024-12-06 17:54:34.115984] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.396 [2024-12-06 17:54:34.116063] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.396 [2024-12-06 17:54:34.116089] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.396 [2024-12-06 17:54:34.116104] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.396 [2024-12-06 17:54:34.116116] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.396 [2024-12-06 17:54:34.116147] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.396 qpair failed and we were unable to recover it. 00:38:52.396 [2024-12-06 17:54:34.125946] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.396 [2024-12-06 17:54:34.126028] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.396 [2024-12-06 17:54:34.126054] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.396 [2024-12-06 17:54:34.126069] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.396 [2024-12-06 17:54:34.126082] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.396 [2024-12-06 17:54:34.126112] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.396 qpair failed and we were unable to recover it. 00:38:52.396 [2024-12-06 17:54:34.136009] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.396 [2024-12-06 17:54:34.136104] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.396 [2024-12-06 17:54:34.136130] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.396 [2024-12-06 17:54:34.136145] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.396 [2024-12-06 17:54:34.136158] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.396 [2024-12-06 17:54:34.136189] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.396 qpair failed and we were unable to recover it. 00:38:52.396 [2024-12-06 17:54:34.146013] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.396 [2024-12-06 17:54:34.146095] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.396 [2024-12-06 17:54:34.146123] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.396 [2024-12-06 17:54:34.146140] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.396 [2024-12-06 17:54:34.146159] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.397 [2024-12-06 17:54:34.146190] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.397 qpair failed and we were unable to recover it. 00:38:52.397 [2024-12-06 17:54:34.156077] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.397 [2024-12-06 17:54:34.156200] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.397 [2024-12-06 17:54:34.156229] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.397 [2024-12-06 17:54:34.156246] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.397 [2024-12-06 17:54:34.156259] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.397 [2024-12-06 17:54:34.156291] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.397 qpair failed and we were unable to recover it. 00:38:52.397 [2024-12-06 17:54:34.166070] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.397 [2024-12-06 17:54:34.166152] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.397 [2024-12-06 17:54:34.166178] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.397 [2024-12-06 17:54:34.166193] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.397 [2024-12-06 17:54:34.166206] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.397 [2024-12-06 17:54:34.166237] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.397 qpair failed and we were unable to recover it. 00:38:52.397 [2024-12-06 17:54:34.176145] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.397 [2024-12-06 17:54:34.176243] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.397 [2024-12-06 17:54:34.176269] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.397 [2024-12-06 17:54:34.176284] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.397 [2024-12-06 17:54:34.176297] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.397 [2024-12-06 17:54:34.176329] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.397 qpair failed and we were unable to recover it. 00:38:52.397 [2024-12-06 17:54:34.186096] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.397 [2024-12-06 17:54:34.186223] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.397 [2024-12-06 17:54:34.186249] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.397 [2024-12-06 17:54:34.186263] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.397 [2024-12-06 17:54:34.186277] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.397 [2024-12-06 17:54:34.186308] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.397 qpair failed and we were unable to recover it. 00:38:52.397 [2024-12-06 17:54:34.196164] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.397 [2024-12-06 17:54:34.196254] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.397 [2024-12-06 17:54:34.196279] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.397 [2024-12-06 17:54:34.196294] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.397 [2024-12-06 17:54:34.196307] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.397 [2024-12-06 17:54:34.196338] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.397 qpair failed and we were unable to recover it. 00:38:52.397 [2024-12-06 17:54:34.206235] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.397 [2024-12-06 17:54:34.206336] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.397 [2024-12-06 17:54:34.206360] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.397 [2024-12-06 17:54:34.206375] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.397 [2024-12-06 17:54:34.206388] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.397 [2024-12-06 17:54:34.206420] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.397 qpair failed and we were unable to recover it. 00:38:52.397 [2024-12-06 17:54:34.216225] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.397 [2024-12-06 17:54:34.216312] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.397 [2024-12-06 17:54:34.216336] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.397 [2024-12-06 17:54:34.216352] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.397 [2024-12-06 17:54:34.216366] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.397 [2024-12-06 17:54:34.216396] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.397 qpair failed and we were unable to recover it. 00:38:52.397 [2024-12-06 17:54:34.226231] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.397 [2024-12-06 17:54:34.226362] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.397 [2024-12-06 17:54:34.226389] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.397 [2024-12-06 17:54:34.226404] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.397 [2024-12-06 17:54:34.226417] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.397 [2024-12-06 17:54:34.226448] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.397 qpair failed and we were unable to recover it. 00:38:52.655 [2024-12-06 17:54:34.236258] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.655 [2024-12-06 17:54:34.236342] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.655 [2024-12-06 17:54:34.236372] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.655 [2024-12-06 17:54:34.236388] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.655 [2024-12-06 17:54:34.236401] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.655 [2024-12-06 17:54:34.236432] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.655 qpair failed and we were unable to recover it. 00:38:52.655 [2024-12-06 17:54:34.246294] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.655 [2024-12-06 17:54:34.246380] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.655 [2024-12-06 17:54:34.246406] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.655 [2024-12-06 17:54:34.246421] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.655 [2024-12-06 17:54:34.246434] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.655 [2024-12-06 17:54:34.246464] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.655 qpair failed and we were unable to recover it. 00:38:52.655 [2024-12-06 17:54:34.256319] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.655 [2024-12-06 17:54:34.256409] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.655 [2024-12-06 17:54:34.256435] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.655 [2024-12-06 17:54:34.256450] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.655 [2024-12-06 17:54:34.256463] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.655 [2024-12-06 17:54:34.256494] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.655 qpair failed and we were unable to recover it. 00:38:52.655 [2024-12-06 17:54:34.266392] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.655 [2024-12-06 17:54:34.266485] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.655 [2024-12-06 17:54:34.266510] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.655 [2024-12-06 17:54:34.266525] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.655 [2024-12-06 17:54:34.266538] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.655 [2024-12-06 17:54:34.266569] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.656 qpair failed and we were unable to recover it. 00:38:52.656 [2024-12-06 17:54:34.276412] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.656 [2024-12-06 17:54:34.276509] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.656 [2024-12-06 17:54:34.276534] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.656 [2024-12-06 17:54:34.276554] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.656 [2024-12-06 17:54:34.276569] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.656 [2024-12-06 17:54:34.276600] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.656 qpair failed and we were unable to recover it. 00:38:52.656 [2024-12-06 17:54:34.286407] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.656 [2024-12-06 17:54:34.286485] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.656 [2024-12-06 17:54:34.286510] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.656 [2024-12-06 17:54:34.286525] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.656 [2024-12-06 17:54:34.286539] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.656 [2024-12-06 17:54:34.286569] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.656 qpair failed and we were unable to recover it. 00:38:52.656 [2024-12-06 17:54:34.296478] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.656 [2024-12-06 17:54:34.296587] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.656 [2024-12-06 17:54:34.296613] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.656 [2024-12-06 17:54:34.296629] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.656 [2024-12-06 17:54:34.296642] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.656 [2024-12-06 17:54:34.296681] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.656 qpair failed and we were unable to recover it. 00:38:52.656 [2024-12-06 17:54:34.306472] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.656 [2024-12-06 17:54:34.306560] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.656 [2024-12-06 17:54:34.306585] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.656 [2024-12-06 17:54:34.306600] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.656 [2024-12-06 17:54:34.306612] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.656 [2024-12-06 17:54:34.306642] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.656 qpair failed and we were unable to recover it. 00:38:52.656 [2024-12-06 17:54:34.316520] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.656 [2024-12-06 17:54:34.316605] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.656 [2024-12-06 17:54:34.316631] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.656 [2024-12-06 17:54:34.316646] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.656 [2024-12-06 17:54:34.316659] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.656 [2024-12-06 17:54:34.316699] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.656 qpair failed and we were unable to recover it. 00:38:52.656 [2024-12-06 17:54:34.326517] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.656 [2024-12-06 17:54:34.326606] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.656 [2024-12-06 17:54:34.326638] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.656 [2024-12-06 17:54:34.326662] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.656 [2024-12-06 17:54:34.326696] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.656 [2024-12-06 17:54:34.326750] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.656 qpair failed and we were unable to recover it. 00:38:52.656 [2024-12-06 17:54:34.336579] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.656 [2024-12-06 17:54:34.336681] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.656 [2024-12-06 17:54:34.336707] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.656 [2024-12-06 17:54:34.336722] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.656 [2024-12-06 17:54:34.336735] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.656 [2024-12-06 17:54:34.336768] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.656 qpair failed and we were unable to recover it. 00:38:52.656 [2024-12-06 17:54:34.346570] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.656 [2024-12-06 17:54:34.346655] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.656 [2024-12-06 17:54:34.346689] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.656 [2024-12-06 17:54:34.346705] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.656 [2024-12-06 17:54:34.346718] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.656 [2024-12-06 17:54:34.346749] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.656 qpair failed and we were unable to recover it. 00:38:52.656 [2024-12-06 17:54:34.356621] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.656 [2024-12-06 17:54:34.356748] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.656 [2024-12-06 17:54:34.356775] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.656 [2024-12-06 17:54:34.356790] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.656 [2024-12-06 17:54:34.356804] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.656 [2024-12-06 17:54:34.356835] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.656 qpair failed and we were unable to recover it. 00:38:52.656 [2024-12-06 17:54:34.366619] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.656 [2024-12-06 17:54:34.366715] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.656 [2024-12-06 17:54:34.366749] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.656 [2024-12-06 17:54:34.366765] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.656 [2024-12-06 17:54:34.366778] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.656 [2024-12-06 17:54:34.366809] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.656 qpair failed and we were unable to recover it. 00:38:52.656 [2024-12-06 17:54:34.376736] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.656 [2024-12-06 17:54:34.376879] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.656 [2024-12-06 17:54:34.376909] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.656 [2024-12-06 17:54:34.376926] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.656 [2024-12-06 17:54:34.376939] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.656 [2024-12-06 17:54:34.376986] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.656 qpair failed and we were unable to recover it. 00:38:52.656 [2024-12-06 17:54:34.386715] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.656 [2024-12-06 17:54:34.386800] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.656 [2024-12-06 17:54:34.386825] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.656 [2024-12-06 17:54:34.386840] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.656 [2024-12-06 17:54:34.386853] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.656 [2024-12-06 17:54:34.386884] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.656 qpair failed and we were unable to recover it. 00:38:52.656 [2024-12-06 17:54:34.396716] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.656 [2024-12-06 17:54:34.396805] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.656 [2024-12-06 17:54:34.396830] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.656 [2024-12-06 17:54:34.396844] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.656 [2024-12-06 17:54:34.396857] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.656 [2024-12-06 17:54:34.396888] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.656 qpair failed and we were unable to recover it. 00:38:52.656 [2024-12-06 17:54:34.406744] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.656 [2024-12-06 17:54:34.406830] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.656 [2024-12-06 17:54:34.406855] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.656 [2024-12-06 17:54:34.406879] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.656 [2024-12-06 17:54:34.406893] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.656 [2024-12-06 17:54:34.406923] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.656 qpair failed and we were unable to recover it. 00:38:52.656 [2024-12-06 17:54:34.416828] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.656 [2024-12-06 17:54:34.416919] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.656 [2024-12-06 17:54:34.416944] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.656 [2024-12-06 17:54:34.416958] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.656 [2024-12-06 17:54:34.416971] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.656 [2024-12-06 17:54:34.417002] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.656 qpair failed and we were unable to recover it. 00:38:52.656 [2024-12-06 17:54:34.426947] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.656 [2024-12-06 17:54:34.427034] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.656 [2024-12-06 17:54:34.427059] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.656 [2024-12-06 17:54:34.427074] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.656 [2024-12-06 17:54:34.427087] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.656 [2024-12-06 17:54:34.427117] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.656 qpair failed and we were unable to recover it. 00:38:52.657 [2024-12-06 17:54:34.436842] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.657 [2024-12-06 17:54:34.436926] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.657 [2024-12-06 17:54:34.436951] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.657 [2024-12-06 17:54:34.436965] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.657 [2024-12-06 17:54:34.436979] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.657 [2024-12-06 17:54:34.437009] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.657 qpair failed and we were unable to recover it. 00:38:52.657 [2024-12-06 17:54:34.446876] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.657 [2024-12-06 17:54:34.446960] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.657 [2024-12-06 17:54:34.446984] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.657 [2024-12-06 17:54:34.446999] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.657 [2024-12-06 17:54:34.447012] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.657 [2024-12-06 17:54:34.447048] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.657 qpair failed and we were unable to recover it. 00:38:52.657 [2024-12-06 17:54:34.456931] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.657 [2024-12-06 17:54:34.457024] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.657 [2024-12-06 17:54:34.457048] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.657 [2024-12-06 17:54:34.457063] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.657 [2024-12-06 17:54:34.457076] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.657 [2024-12-06 17:54:34.457107] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.657 qpair failed and we were unable to recover it. 00:38:52.657 [2024-12-06 17:54:34.466940] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.657 [2024-12-06 17:54:34.467025] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.657 [2024-12-06 17:54:34.467051] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.657 [2024-12-06 17:54:34.467065] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.657 [2024-12-06 17:54:34.467079] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.657 [2024-12-06 17:54:34.467122] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.657 qpair failed and we were unable to recover it. 00:38:52.657 [2024-12-06 17:54:34.476944] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.657 [2024-12-06 17:54:34.477026] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.657 [2024-12-06 17:54:34.477051] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.657 [2024-12-06 17:54:34.477065] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.657 [2024-12-06 17:54:34.477079] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.657 [2024-12-06 17:54:34.477110] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.657 qpair failed and we were unable to recover it. 00:38:52.657 [2024-12-06 17:54:34.486968] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.657 [2024-12-06 17:54:34.487048] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.657 [2024-12-06 17:54:34.487073] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.657 [2024-12-06 17:54:34.487087] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.657 [2024-12-06 17:54:34.487100] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.657 [2024-12-06 17:54:34.487130] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.657 qpair failed and we were unable to recover it. 00:38:52.915 [2024-12-06 17:54:34.497044] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.915 [2024-12-06 17:54:34.497163] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.915 [2024-12-06 17:54:34.497189] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.915 [2024-12-06 17:54:34.497205] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.915 [2024-12-06 17:54:34.497217] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.915 [2024-12-06 17:54:34.497248] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.915 qpair failed and we were unable to recover it. 00:38:52.915 [2024-12-06 17:54:34.507141] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.915 [2024-12-06 17:54:34.507226] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.915 [2024-12-06 17:54:34.507252] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.915 [2024-12-06 17:54:34.507266] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.915 [2024-12-06 17:54:34.507279] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe4c000b90 00:38:52.915 [2024-12-06 17:54:34.507309] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:38:52.915 qpair failed and we were unable to recover it. 00:38:52.915 [2024-12-06 17:54:34.507426] nvme_ctrlr.c:4518:nvme_ctrlr_keep_alive: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] Submitting Keep Alive failed 00:38:52.915 A controller has encountered a failure and is being reset. 00:38:52.915 [2024-12-06 17:54:34.517082] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.915 [2024-12-06 17:54:34.517163] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.915 [2024-12-06 17:54:34.517195] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.915 [2024-12-06 17:54:34.517212] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.915 [2024-12-06 17:54:34.517225] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe50000b90 00:38:52.915 [2024-12-06 17:54:34.517257] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:38:52.915 qpair failed and we were unable to recover it. 00:38:52.915 [2024-12-06 17:54:34.527135] ctrlr.c: 764:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:38:52.915 [2024-12-06 17:54:34.527219] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:38:52.915 [2024-12-06 17:54:34.527246] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:38:52.915 [2024-12-06 17:54:34.527261] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:38:52.915 [2024-12-06 17:54:34.527275] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fbe50000b90 00:38:52.915 [2024-12-06 17:54:34.527306] nvme_qpair.c: 812:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:38:52.915 qpair failed and we were unable to recover it. 00:38:52.915 Controller properly reset. 00:38:52.915 Initializing NVMe Controllers 00:38:52.915 Attaching to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:38:52.915 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:38:52.915 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) with lcore 0 00:38:52.915 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) with lcore 1 00:38:52.915 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) with lcore 2 00:38:52.915 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) with lcore 3 00:38:52.915 Initialization complete. Launching workers. 00:38:52.915 Starting thread on core 1 00:38:52.915 Starting thread on core 2 00:38:52.915 Starting thread on core 3 00:38:52.915 Starting thread on core 0 00:38:52.915 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@51 -- # sync 00:38:52.915 00:38:52.915 real 0m10.675s 00:38:52.915 user 0m18.955s 00:38:52.915 sys 0m5.348s 00:38:52.915 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@1130 -- # xtrace_disable 00:38:52.915 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:38:52.915 ************************************ 00:38:52.915 END TEST nvmf_target_disconnect_tc2 00:38:52.915 ************************************ 00:38:52.915 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@72 -- # '[' -n '' ']' 00:38:52.915 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@76 -- # trap - SIGINT SIGTERM EXIT 00:38:52.915 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@77 -- # nvmftestfini 00:38:52.915 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@516 -- # nvmfcleanup 00:38:52.915 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@121 -- # sync 00:38:52.915 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:38:52.915 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@124 -- # set +e 00:38:52.915 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@125 -- # for i in {1..20} 00:38:52.915 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:38:52.915 rmmod nvme_tcp 00:38:52.915 rmmod nvme_fabrics 00:38:52.915 rmmod nvme_keyring 00:38:52.915 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:38:52.915 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@128 -- # set -e 00:38:52.915 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@129 -- # return 0 00:38:52.915 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@517 -- # '[' -n 415275 ']' 00:38:52.915 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@518 -- # killprocess 415275 00:38:52.915 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@954 -- # '[' -z 415275 ']' 00:38:52.915 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@958 -- # kill -0 415275 00:38:52.915 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@959 -- # uname 00:38:52.915 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:38:52.915 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 415275 00:38:52.915 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@960 -- # process_name=reactor_4 00:38:52.915 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@964 -- # '[' reactor_4 = sudo ']' 00:38:52.915 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@972 -- # echo 'killing process with pid 415275' 00:38:52.915 killing process with pid 415275 00:38:52.915 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@973 -- # kill 415275 00:38:52.915 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@978 -- # wait 415275 00:38:53.173 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:38:53.173 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:38:53.173 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:38:53.173 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@297 -- # iptr 00:38:53.173 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@791 -- # iptables-save 00:38:53.173 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:38:53.173 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@791 -- # iptables-restore 00:38:53.173 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:38:53.173 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@302 -- # remove_spdk_ns 00:38:53.173 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:38:53.174 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:38:53.174 17:54:34 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:38:55.716 17:54:36 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:38:55.716 00:38:55.716 real 0m15.791s 00:38:55.716 user 0m45.146s 00:38:55.716 sys 0m7.545s 00:38:55.716 17:54:36 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1130 -- # xtrace_disable 00:38:55.716 17:54:36 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@10 -- # set +x 00:38:55.716 ************************************ 00:38:55.716 END TEST nvmf_target_disconnect 00:38:55.716 ************************************ 00:38:55.716 17:54:36 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@51 -- # trap - SIGINT SIGTERM EXIT 00:38:55.716 00:38:55.716 real 6m41.563s 00:38:55.716 user 17m8.924s 00:38:55.716 sys 1m26.763s 00:38:55.716 17:54:36 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1130 -- # xtrace_disable 00:38:55.716 17:54:36 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:38:55.716 ************************************ 00:38:55.716 END TEST nvmf_host 00:38:55.716 ************************************ 00:38:55.716 17:54:36 nvmf_tcp -- nvmf/nvmf.sh@19 -- # [[ tcp = \t\c\p ]] 00:38:55.716 17:54:36 nvmf_tcp -- nvmf/nvmf.sh@19 -- # [[ 0 -eq 0 ]] 00:38:55.716 17:54:36 nvmf_tcp -- nvmf/nvmf.sh@20 -- # run_test nvmf_target_core_interrupt_mode /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_target_core.sh --transport=tcp --interrupt-mode 00:38:55.716 17:54:36 nvmf_tcp -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:38:55.716 17:54:36 nvmf_tcp -- common/autotest_common.sh@1111 -- # xtrace_disable 00:38:55.716 17:54:36 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:38:55.716 ************************************ 00:38:55.716 START TEST nvmf_target_core_interrupt_mode 00:38:55.716 ************************************ 00:38:55.716 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_target_core.sh --transport=tcp --interrupt-mode 00:38:55.716 * Looking for test storage... 00:38:55.716 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf 00:38:55.716 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:38:55.716 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1711 -- # lcov --version 00:38:55.716 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:38:55.716 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:38:55.716 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:38:55.716 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@333 -- # local ver1 ver1_l 00:38:55.716 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@334 -- # local ver2 ver2_l 00:38:55.716 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@336 -- # IFS=.-: 00:38:55.716 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@336 -- # read -ra ver1 00:38:55.716 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@337 -- # IFS=.-: 00:38:55.716 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@337 -- # read -ra ver2 00:38:55.716 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@338 -- # local 'op=<' 00:38:55.716 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@340 -- # ver1_l=2 00:38:55.716 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@341 -- # ver2_l=1 00:38:55.716 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:38:55.716 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@344 -- # case "$op" in 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@345 -- # : 1 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@364 -- # (( v = 0 )) 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@365 -- # decimal 1 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@353 -- # local d=1 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@355 -- # echo 1 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@365 -- # ver1[v]=1 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@366 -- # decimal 2 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@353 -- # local d=2 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@355 -- # echo 2 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@366 -- # ver2[v]=2 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@368 -- # return 0 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:38:55.717 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:55.717 --rc genhtml_branch_coverage=1 00:38:55.717 --rc genhtml_function_coverage=1 00:38:55.717 --rc genhtml_legend=1 00:38:55.717 --rc geninfo_all_blocks=1 00:38:55.717 --rc geninfo_unexecuted_blocks=1 00:38:55.717 00:38:55.717 ' 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:38:55.717 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:55.717 --rc genhtml_branch_coverage=1 00:38:55.717 --rc genhtml_function_coverage=1 00:38:55.717 --rc genhtml_legend=1 00:38:55.717 --rc geninfo_all_blocks=1 00:38:55.717 --rc geninfo_unexecuted_blocks=1 00:38:55.717 00:38:55.717 ' 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:38:55.717 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:55.717 --rc genhtml_branch_coverage=1 00:38:55.717 --rc genhtml_function_coverage=1 00:38:55.717 --rc genhtml_legend=1 00:38:55.717 --rc geninfo_all_blocks=1 00:38:55.717 --rc geninfo_unexecuted_blocks=1 00:38:55.717 00:38:55.717 ' 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:38:55.717 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:55.717 --rc genhtml_branch_coverage=1 00:38:55.717 --rc genhtml_function_coverage=1 00:38:55.717 --rc genhtml_legend=1 00:38:55.717 --rc geninfo_all_blocks=1 00:38:55.717 --rc geninfo_unexecuted_blocks=1 00:38:55.717 00:38:55.717 ' 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@10 -- # uname -s 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@10 -- # '[' '!' Linux = Linux ']' 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@14 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@7 -- # uname -s 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@15 -- # shopt -s extglob 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- paths/export.sh@5 -- # export PATH 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@51 -- # : 0 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@55 -- # have_pci_nics=0 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@16 -- # trap 'exit 1' SIGINT SIGTERM EXIT 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@18 -- # TEST_ARGS=("$@") 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@20 -- # [[ 0 -eq 0 ]] 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@21 -- # run_test nvmf_abort /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort.sh --transport=tcp --interrupt-mode 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1111 -- # xtrace_disable 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:38:55.717 ************************************ 00:38:55.717 START TEST nvmf_abort 00:38:55.717 ************************************ 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort.sh --transport=tcp --interrupt-mode 00:38:55.717 * Looking for test storage... 00:38:55.717 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1711 -- # lcov --version 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:38:55.717 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@333 -- # local ver1 ver1_l 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@334 -- # local ver2 ver2_l 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@336 -- # IFS=.-: 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@336 -- # read -ra ver1 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@337 -- # IFS=.-: 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@337 -- # read -ra ver2 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@338 -- # local 'op=<' 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@340 -- # ver1_l=2 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@341 -- # ver2_l=1 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@344 -- # case "$op" in 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@345 -- # : 1 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@364 -- # (( v = 0 )) 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@365 -- # decimal 1 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@353 -- # local d=1 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@355 -- # echo 1 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@365 -- # ver1[v]=1 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@366 -- # decimal 2 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@353 -- # local d=2 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@355 -- # echo 2 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@366 -- # ver2[v]=2 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@368 -- # return 0 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:38:55.718 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:55.718 --rc genhtml_branch_coverage=1 00:38:55.718 --rc genhtml_function_coverage=1 00:38:55.718 --rc genhtml_legend=1 00:38:55.718 --rc geninfo_all_blocks=1 00:38:55.718 --rc geninfo_unexecuted_blocks=1 00:38:55.718 00:38:55.718 ' 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:38:55.718 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:55.718 --rc genhtml_branch_coverage=1 00:38:55.718 --rc genhtml_function_coverage=1 00:38:55.718 --rc genhtml_legend=1 00:38:55.718 --rc geninfo_all_blocks=1 00:38:55.718 --rc geninfo_unexecuted_blocks=1 00:38:55.718 00:38:55.718 ' 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:38:55.718 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:55.718 --rc genhtml_branch_coverage=1 00:38:55.718 --rc genhtml_function_coverage=1 00:38:55.718 --rc genhtml_legend=1 00:38:55.718 --rc geninfo_all_blocks=1 00:38:55.718 --rc geninfo_unexecuted_blocks=1 00:38:55.718 00:38:55.718 ' 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:38:55.718 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:55.718 --rc genhtml_branch_coverage=1 00:38:55.718 --rc genhtml_function_coverage=1 00:38:55.718 --rc genhtml_legend=1 00:38:55.718 --rc geninfo_all_blocks=1 00:38:55.718 --rc geninfo_unexecuted_blocks=1 00:38:55.718 00:38:55.718 ' 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@7 -- # uname -s 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@15 -- # shopt -s extglob 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- paths/export.sh@5 -- # export PATH 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@51 -- # : 0 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:38:55.718 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:38:55.719 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:38:55.719 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:38:55.719 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:38:55.719 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:38:55.719 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:38:55.719 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@55 -- # have_pci_nics=0 00:38:55.719 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@11 -- # MALLOC_BDEV_SIZE=64 00:38:55.719 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@12 -- # MALLOC_BLOCK_SIZE=4096 00:38:55.719 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@14 -- # nvmftestinit 00:38:55.719 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:38:55.719 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:38:55.719 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@476 -- # prepare_net_devs 00:38:55.719 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@438 -- # local -g is_hw=no 00:38:55.719 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@440 -- # remove_spdk_ns 00:38:55.719 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:38:55.719 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:38:55.719 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:38:55.719 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:38:55.719 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:38:55.719 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@309 -- # xtrace_disable 00:38:55.719 17:54:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:38:58.246 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:38:58.246 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@315 -- # pci_devs=() 00:38:58.246 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@315 -- # local -a pci_devs 00:38:58.246 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@316 -- # pci_net_devs=() 00:38:58.246 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:38:58.246 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@317 -- # pci_drivers=() 00:38:58.246 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@317 -- # local -A pci_drivers 00:38:58.246 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@319 -- # net_devs=() 00:38:58.246 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@319 -- # local -ga net_devs 00:38:58.246 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@320 -- # e810=() 00:38:58.246 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@320 -- # local -ga e810 00:38:58.246 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@321 -- # x722=() 00:38:58.246 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@321 -- # local -ga x722 00:38:58.246 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@322 -- # mlx=() 00:38:58.246 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@322 -- # local -ga mlx 00:38:58.246 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:38:58.246 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:38:58.246 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:38:58.246 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:38:58.246 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:38:58.246 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:38:58.246 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:38:58.247 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:38:58.247 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@418 -- # [[ up == up ]] 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:38:58.247 Found net devices under 0000:0a:00.0: cvl_0_0 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@418 -- # [[ up == up ]] 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:38:58.247 Found net devices under 0000:0a:00.1: cvl_0_1 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@442 -- # is_hw=yes 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:38:58.247 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:38:58.247 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.142 ms 00:38:58.247 00:38:58.247 --- 10.0.0.2 ping statistics --- 00:38:58.247 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:38:58.247 rtt min/avg/max/mdev = 0.142/0.142/0.142/0.000 ms 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:38:58.247 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:38:58.247 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.067 ms 00:38:58.247 00:38:58.247 --- 10.0.0.1 ping statistics --- 00:38:58.247 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:38:58.247 rtt min/avg/max/mdev = 0.067/0.067/0.067/0.000 ms 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:38:58.247 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@450 -- # return 0 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@15 -- # nvmfappstart -m 0xE 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@726 -- # xtrace_disable 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@509 -- # nvmfpid=418081 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0xE 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@510 -- # waitforlisten 418081 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@835 -- # '[' -z 418081 ']' 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@840 -- # local max_retries=100 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:38:58.248 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@844 -- # xtrace_disable 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:38:58.248 [2024-12-06 17:54:39.676997] thread.c:3005:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:38:58.248 [2024-12-06 17:54:39.678023] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:38:58.248 [2024-12-06 17:54:39.678093] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:38:58.248 [2024-12-06 17:54:39.752566] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:38:58.248 [2024-12-06 17:54:39.796283] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:38:58.248 [2024-12-06 17:54:39.796344] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:38:58.248 [2024-12-06 17:54:39.796367] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:38:58.248 [2024-12-06 17:54:39.796379] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:38:58.248 [2024-12-06 17:54:39.796388] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:38:58.248 [2024-12-06 17:54:39.797932] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:38:58.248 [2024-12-06 17:54:39.797990] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:38:58.248 [2024-12-06 17:54:39.797994] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:38:58.248 [2024-12-06 17:54:39.881239] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:38:58.248 [2024-12-06 17:54:39.881423] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:38:58.248 [2024-12-06 17:54:39.881451] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:38:58.248 [2024-12-06 17:54:39.881709] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@868 -- # return 0 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@732 -- # xtrace_disable 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -a 256 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:38:58.248 [2024-12-06 17:54:39.934710] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@20 -- # rpc_cmd bdev_malloc_create 64 4096 -b Malloc0 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:38:58.248 Malloc0 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@21 -- # rpc_cmd bdev_delay_create -b Malloc0 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:38:58.248 Delay0 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@25 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 Delay0 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:58.248 17:54:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:38:58.248 17:54:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:58.248 17:54:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@26 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:38:58.248 17:54:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:58.248 17:54:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:38:58.248 [2024-12-06 17:54:40.010879] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:38:58.248 17:54:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:58.248 17:54:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:38:58.248 17:54:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:38:58.248 17:54:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:38:58.248 17:54:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:38:58.248 17:54:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -c 0x1 -t 1 -l warning -q 128 00:38:58.505 [2024-12-06 17:54:40.126612] nvme_fabric.c: 295:nvme_fabric_discover_probe: *WARNING*: Skipping unsupported current discovery service or discovery service referral 00:39:00.402 Initializing NVMe Controllers 00:39:00.402 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode0 00:39:00.402 controller IO queue size 128 less than required 00:39:00.402 Consider using lower queue depth or small IO size because IO requests may be queued at the NVMe driver. 00:39:00.402 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 0 00:39:00.402 Initialization complete. Launching workers. 00:39:00.402 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 I/O completed: 123, failed: 29462 00:39:00.402 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) abort submitted 29519, failed to submit 66 00:39:00.402 success 29462, unsuccessful 57, failed 0 00:39:00.402 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@34 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:39:00.402 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:39:00.402 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:39:00.402 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:39:00.402 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@36 -- # trap - SIGINT SIGTERM EXIT 00:39:00.402 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@38 -- # nvmftestfini 00:39:00.402 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@516 -- # nvmfcleanup 00:39:00.402 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@121 -- # sync 00:39:00.402 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:39:00.402 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@124 -- # set +e 00:39:00.402 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@125 -- # for i in {1..20} 00:39:00.402 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:39:00.402 rmmod nvme_tcp 00:39:00.402 rmmod nvme_fabrics 00:39:00.402 rmmod nvme_keyring 00:39:00.402 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:39:00.402 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@128 -- # set -e 00:39:00.402 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@129 -- # return 0 00:39:00.402 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@517 -- # '[' -n 418081 ']' 00:39:00.402 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@518 -- # killprocess 418081 00:39:00.402 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@954 -- # '[' -z 418081 ']' 00:39:00.402 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@958 -- # kill -0 418081 00:39:00.402 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@959 -- # uname 00:39:00.402 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:39:00.402 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 418081 00:39:00.661 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:39:00.661 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:39:00.661 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@972 -- # echo 'killing process with pid 418081' 00:39:00.661 killing process with pid 418081 00:39:00.661 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@973 -- # kill 418081 00:39:00.661 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@978 -- # wait 418081 00:39:00.661 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:39:00.661 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:39:00.661 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:39:00.661 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@297 -- # iptr 00:39:00.661 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@791 -- # iptables-save 00:39:00.661 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:39:00.661 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@791 -- # iptables-restore 00:39:00.920 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:39:00.920 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@302 -- # remove_spdk_ns 00:39:00.920 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:39:00.920 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:39:00.920 17:54:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:39:02.820 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:39:02.820 00:39:02.820 real 0m7.351s 00:39:02.820 user 0m9.200s 00:39:02.820 sys 0m2.933s 00:39:02.820 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1130 -- # xtrace_disable 00:39:02.820 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:39:02.820 ************************************ 00:39:02.820 END TEST nvmf_abort 00:39:02.820 ************************************ 00:39:02.820 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@22 -- # run_test nvmf_ns_hotplug_stress /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh --transport=tcp --interrupt-mode 00:39:02.820 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:39:02.820 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1111 -- # xtrace_disable 00:39:02.820 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:39:02.820 ************************************ 00:39:02.820 START TEST nvmf_ns_hotplug_stress 00:39:02.820 ************************************ 00:39:02.820 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh --transport=tcp --interrupt-mode 00:39:02.820 * Looking for test storage... 00:39:02.820 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:39:02.820 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:39:02.820 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1711 -- # lcov --version 00:39:02.820 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@333 -- # local ver1 ver1_l 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@334 -- # local ver2 ver2_l 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@336 -- # IFS=.-: 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@336 -- # read -ra ver1 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@337 -- # IFS=.-: 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@337 -- # read -ra ver2 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@338 -- # local 'op=<' 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@340 -- # ver1_l=2 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@341 -- # ver2_l=1 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@344 -- # case "$op" in 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@345 -- # : 1 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@364 -- # (( v = 0 )) 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@365 -- # decimal 1 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@353 -- # local d=1 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@355 -- # echo 1 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@365 -- # ver1[v]=1 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@366 -- # decimal 2 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@353 -- # local d=2 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@355 -- # echo 2 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@366 -- # ver2[v]=2 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@368 -- # return 0 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:39:03.080 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:39:03.080 --rc genhtml_branch_coverage=1 00:39:03.080 --rc genhtml_function_coverage=1 00:39:03.080 --rc genhtml_legend=1 00:39:03.080 --rc geninfo_all_blocks=1 00:39:03.080 --rc geninfo_unexecuted_blocks=1 00:39:03.080 00:39:03.080 ' 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:39:03.080 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:39:03.080 --rc genhtml_branch_coverage=1 00:39:03.080 --rc genhtml_function_coverage=1 00:39:03.080 --rc genhtml_legend=1 00:39:03.080 --rc geninfo_all_blocks=1 00:39:03.080 --rc geninfo_unexecuted_blocks=1 00:39:03.080 00:39:03.080 ' 00:39:03.080 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:39:03.080 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:39:03.080 --rc genhtml_branch_coverage=1 00:39:03.080 --rc genhtml_function_coverage=1 00:39:03.081 --rc genhtml_legend=1 00:39:03.081 --rc geninfo_all_blocks=1 00:39:03.081 --rc geninfo_unexecuted_blocks=1 00:39:03.081 00:39:03.081 ' 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:39:03.081 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:39:03.081 --rc genhtml_branch_coverage=1 00:39:03.081 --rc genhtml_function_coverage=1 00:39:03.081 --rc genhtml_legend=1 00:39:03.081 --rc geninfo_all_blocks=1 00:39:03.081 --rc geninfo_unexecuted_blocks=1 00:39:03.081 00:39:03.081 ' 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@7 -- # uname -s 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@15 -- # shopt -s extglob 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- paths/export.sh@5 -- # export PATH 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@51 -- # : 0 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@55 -- # have_pci_nics=0 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@11 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@22 -- # nvmftestinit 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@476 -- # prepare_net_devs 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@438 -- # local -g is_hw=no 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@440 -- # remove_spdk_ns 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@309 -- # xtrace_disable 00:39:03.081 17:54:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:39:04.982 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:39:04.982 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@315 -- # pci_devs=() 00:39:04.982 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@315 -- # local -a pci_devs 00:39:04.982 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@316 -- # pci_net_devs=() 00:39:04.982 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:39:04.982 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@317 -- # pci_drivers=() 00:39:04.982 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@317 -- # local -A pci_drivers 00:39:04.982 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@319 -- # net_devs=() 00:39:04.982 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@319 -- # local -ga net_devs 00:39:04.982 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@320 -- # e810=() 00:39:04.982 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@320 -- # local -ga e810 00:39:04.982 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@321 -- # x722=() 00:39:04.982 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@321 -- # local -ga x722 00:39:04.982 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@322 -- # mlx=() 00:39:04.982 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@322 -- # local -ga mlx 00:39:04.983 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:39:04.983 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:39:04.983 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:39:04.983 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:39:05.242 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:39:05.242 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@418 -- # [[ up == up ]] 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:39:05.242 Found net devices under 0000:0a:00.0: cvl_0_0 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@418 -- # [[ up == up ]] 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:39:05.242 Found net devices under 0000:0a:00.1: cvl_0_1 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@442 -- # is_hw=yes 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:39:05.242 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:39:05.242 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.122 ms 00:39:05.242 00:39:05.242 --- 10.0.0.2 ping statistics --- 00:39:05.242 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:39:05.242 rtt min/avg/max/mdev = 0.122/0.122/0.122/0.000 ms 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:39:05.242 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:39:05.242 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.143 ms 00:39:05.242 00:39:05.242 --- 10.0.0.1 ping statistics --- 00:39:05.242 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:39:05.242 rtt min/avg/max/mdev = 0.143/0.143/0.143/0.000 ms 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@450 -- # return 0 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:39:05.242 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:39:05.243 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:39:05.243 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:39:05.243 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:39:05.243 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@23 -- # nvmfappstart -m 0xE 00:39:05.243 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:39:05.243 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@726 -- # xtrace_disable 00:39:05.243 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:39:05.243 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@509 -- # nvmfpid=420301 00:39:05.243 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0xE 00:39:05.243 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@510 -- # waitforlisten 420301 00:39:05.243 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@835 -- # '[' -z 420301 ']' 00:39:05.243 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:39:05.243 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@840 -- # local max_retries=100 00:39:05.243 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:39:05.243 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:39:05.243 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@844 -- # xtrace_disable 00:39:05.243 17:54:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:39:05.243 [2024-12-06 17:54:47.030187] thread.c:3005:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:39:05.243 [2024-12-06 17:54:47.031251] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:39:05.243 [2024-12-06 17:54:47.031313] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:39:05.501 [2024-12-06 17:54:47.103891] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:39:05.501 [2024-12-06 17:54:47.146732] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:39:05.501 [2024-12-06 17:54:47.146793] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:39:05.501 [2024-12-06 17:54:47.146820] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:39:05.501 [2024-12-06 17:54:47.146831] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:39:05.501 [2024-12-06 17:54:47.146840] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:39:05.501 [2024-12-06 17:54:47.148439] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:39:05.501 [2024-12-06 17:54:47.148545] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:39:05.501 [2024-12-06 17:54:47.148548] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:39:05.501 [2024-12-06 17:54:47.230361] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:39:05.501 [2024-12-06 17:54:47.230465] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:39:05.501 [2024-12-06 17:54:47.230467] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:39:05.501 [2024-12-06 17:54:47.230729] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:39:05.501 17:54:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:39:05.501 17:54:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@868 -- # return 0 00:39:05.501 17:54:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:39:05.501 17:54:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@732 -- # xtrace_disable 00:39:05.501 17:54:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:39:05.501 17:54:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:39:05.501 17:54:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@25 -- # null_size=1000 00:39:05.501 17:54:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:39:05.759 [2024-12-06 17:54:47.545241] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:39:05.759 17:54:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:39:06.324 17:54:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:39:06.581 [2024-12-06 17:54:48.169590] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:39:06.581 17:54:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:39:06.839 17:54:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 512 -b Malloc0 00:39:07.096 Malloc0 00:39:07.096 17:54:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_delay_create -b Malloc0 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:39:07.353 Delay0 00:39:07.353 17:54:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:07.610 17:54:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create NULL1 1000 512 00:39:07.867 NULL1 00:39:08.125 17:54:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 NULL1 00:39:08.383 17:54:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@42 -- # PERF_PID=420715 00:39:08.383 17:54:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0x1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 30 -q 128 -w randread -o 512 -Q 1000 00:39:08.383 17:54:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:08.383 17:54:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:09.316 Read completed with error (sct=0, sc=11) 00:39:09.316 17:54:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:09.316 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:39:09.574 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:39:09.574 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:39:09.574 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:39:09.574 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:39:09.574 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:39:09.574 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:39:09.574 17:54:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1001 00:39:09.574 17:54:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1001 00:39:09.831 true 00:39:10.090 17:54:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:10.090 17:54:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:10.655 17:54:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:10.913 17:54:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1002 00:39:10.913 17:54:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1002 00:39:11.170 true 00:39:11.170 17:54:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:11.170 17:54:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:11.428 17:54:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:11.759 17:54:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1003 00:39:11.759 17:54:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1003 00:39:12.016 true 00:39:12.016 17:54:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:12.016 17:54:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:12.274 17:54:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:12.531 17:54:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1004 00:39:12.531 17:54:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1004 00:39:12.788 true 00:39:12.788 17:54:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:12.789 17:54:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:13.720 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:39:13.720 17:54:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:13.720 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:39:13.976 17:54:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1005 00:39:13.976 17:54:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1005 00:39:14.233 true 00:39:14.233 17:54:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:14.233 17:54:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:14.798 17:54:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:14.798 17:54:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1006 00:39:14.798 17:54:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1006 00:39:15.055 true 00:39:15.055 17:54:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:15.056 17:54:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:15.621 17:54:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:15.621 17:54:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1007 00:39:15.622 17:54:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1007 00:39:15.879 true 00:39:15.879 17:54:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:15.879 17:54:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:17.253 17:54:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:17.253 17:54:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1008 00:39:17.253 17:54:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1008 00:39:17.511 true 00:39:17.511 17:54:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:17.512 17:54:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:17.769 17:54:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:18.028 17:54:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1009 00:39:18.028 17:54:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1009 00:39:18.289 true 00:39:18.289 17:55:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:18.289 17:55:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:18.580 17:55:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:18.862 17:55:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1010 00:39:18.862 17:55:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1010 00:39:19.121 true 00:39:19.121 17:55:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:19.121 17:55:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:20.054 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:39:20.054 17:55:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:20.054 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:39:20.311 17:55:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1011 00:39:20.311 17:55:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1011 00:39:20.570 true 00:39:20.570 17:55:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:20.570 17:55:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:20.827 17:55:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:21.085 17:55:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1012 00:39:21.085 17:55:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1012 00:39:21.343 true 00:39:21.343 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:21.343 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:21.601 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:21.860 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1013 00:39:21.860 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1013 00:39:22.118 true 00:39:22.118 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:22.118 17:55:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:23.499 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:39:23.499 17:55:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:23.499 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1014 00:39:23.500 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1014 00:39:23.761 true 00:39:23.761 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:23.761 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:24.018 17:55:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:24.275 17:55:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1015 00:39:24.275 17:55:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1015 00:39:24.531 true 00:39:24.531 17:55:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:24.531 17:55:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:24.788 17:55:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:25.046 17:55:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1016 00:39:25.046 17:55:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1016 00:39:25.303 true 00:39:25.303 17:55:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:25.303 17:55:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:26.233 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:39:26.233 17:55:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:26.796 17:55:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1017 00:39:26.796 17:55:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1017 00:39:26.796 true 00:39:26.796 17:55:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:26.796 17:55:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:27.052 17:55:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:27.616 17:55:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1018 00:39:27.616 17:55:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1018 00:39:27.616 true 00:39:27.616 17:55:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:27.617 17:55:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:27.875 17:55:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:28.441 17:55:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1019 00:39:28.441 17:55:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1019 00:39:28.441 true 00:39:28.441 17:55:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:28.441 17:55:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:29.391 17:55:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:29.649 17:55:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1020 00:39:29.649 17:55:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1020 00:39:29.907 true 00:39:29.907 17:55:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:29.907 17:55:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:30.165 17:55:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:30.423 17:55:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1021 00:39:30.423 17:55:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1021 00:39:30.681 true 00:39:30.681 17:55:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:30.681 17:55:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:30.938 17:55:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:31.196 17:55:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1022 00:39:31.196 17:55:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1022 00:39:31.454 true 00:39:31.454 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:31.454 17:55:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:32.386 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:39:32.386 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:32.644 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1023 00:39:32.644 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1023 00:39:32.902 true 00:39:32.902 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:32.902 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:33.159 17:55:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:33.417 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1024 00:39:33.417 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1024 00:39:33.675 true 00:39:33.675 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:33.675 17:55:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:34.609 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:39:34.609 17:55:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:34.867 17:55:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1025 00:39:34.867 17:55:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1025 00:39:35.123 true 00:39:35.123 17:55:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:35.123 17:55:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:35.380 17:55:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:35.637 17:55:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1026 00:39:35.637 17:55:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1026 00:39:35.893 true 00:39:35.893 17:55:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:35.893 17:55:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:36.150 17:55:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:36.406 17:55:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1027 00:39:36.406 17:55:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1027 00:39:36.663 true 00:39:36.663 17:55:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:36.663 17:55:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:38.035 Message suppressed 999 times: Read completed with error (sct=0, sc=11) 00:39:38.035 17:55:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:38.035 17:55:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1028 00:39:38.035 17:55:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1028 00:39:38.293 true 00:39:38.293 17:55:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:38.293 17:55:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:38.550 Initializing NVMe Controllers 00:39:38.550 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:39:38.550 Controller IO queue size 128, less than required. 00:39:38.550 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:39:38.550 Controller IO queue size 128, less than required. 00:39:38.550 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:39:38.550 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:39:38.550 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:39:38.550 Initialization complete. Launching workers. 00:39:38.550 ======================================================== 00:39:38.550 Latency(us) 00:39:38.550 Device Information : IOPS MiB/s Average min max 00:39:38.550 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 609.94 0.30 86215.96 2320.68 1078195.27 00:39:38.550 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 8556.00 4.18 14960.13 1836.81 536854.40 00:39:38.550 ======================================================== 00:39:38.550 Total : 9165.94 4.48 19701.80 1836.81 1078195.27 00:39:38.550 00:39:38.550 17:55:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:38.807 17:55:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1029 00:39:38.807 17:55:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1029 00:39:39.064 true 00:39:39.064 17:55:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 420715 00:39:39.064 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh: line 44: kill: (420715) - No such process 00:39:39.064 17:55:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@53 -- # wait 420715 00:39:39.064 17:55:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:39.322 17:55:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:39:39.580 17:55:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@58 -- # nthreads=8 00:39:39.580 17:55:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@58 -- # pids=() 00:39:39.580 17:55:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i = 0 )) 00:39:39.580 17:55:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:39:39.580 17:55:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null0 100 4096 00:39:39.838 null0 00:39:39.838 17:55:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:39:39.838 17:55:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:39:39.838 17:55:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null1 100 4096 00:39:40.096 null1 00:39:40.354 17:55:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:39:40.354 17:55:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:39:40.354 17:55:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null2 100 4096 00:39:40.612 null2 00:39:40.612 17:55:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:39:40.612 17:55:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:39:40.612 17:55:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null3 100 4096 00:39:40.870 null3 00:39:40.870 17:55:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:39:40.870 17:55:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:39:40.870 17:55:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null4 100 4096 00:39:41.128 null4 00:39:41.128 17:55:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:39:41.128 17:55:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:39:41.128 17:55:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null5 100 4096 00:39:41.386 null5 00:39:41.386 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:39:41.386 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:39:41.386 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null6 100 4096 00:39:41.645 null6 00:39:41.645 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:39:41.645 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:39:41.645 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null7 100 4096 00:39:41.904 null7 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i = 0 )) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 1 null0 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=1 bdev=null0 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 2 null1 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=2 bdev=null1 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 3 null2 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=3 bdev=null2 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 4 null3 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=4 bdev=null3 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 5 null4 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=5 bdev=null4 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 6 null5 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=6 bdev=null5 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:39:41.904 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:39:41.905 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 7 null6 00:39:41.905 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:39:41.905 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=7 bdev=null6 00:39:41.905 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:39:41.905 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:39:41.905 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:41.905 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:39:41.905 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:39:41.905 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 8 null7 00:39:41.905 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:39:41.905 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=8 bdev=null7 00:39:41.905 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:39:41.905 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:39:41.905 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@66 -- # wait 424715 424716 424718 424720 424722 424724 424726 424728 00:39:41.905 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:41.905 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:39:42.163 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:39:42.163 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:39:42.163 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:39:42.163 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:42.163 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:39:42.163 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:39:42.163 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:39:42.163 17:55:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:39:42.422 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:42.422 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:42.422 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:39:42.422 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:42.422 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:42.422 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:39:42.422 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:42.422 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:42.422 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:39:42.422 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:42.422 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:42.422 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:39:42.422 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:42.422 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:42.422 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:39:42.422 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:42.422 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:42.422 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:42.422 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:39:42.422 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:42.422 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:39:42.422 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:42.422 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:42.422 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:39:42.680 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:42.680 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:39:42.680 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:39:42.680 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:39:42.680 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:39:42.680 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:39:42.680 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:39:42.680 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:39:42.948 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:42.948 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:42.948 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:39:42.948 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:42.948 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:42.948 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:39:42.948 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:42.948 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:42.948 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:39:42.948 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:42.948 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:42.948 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:39:42.948 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:42.948 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:42.948 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:39:43.205 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:43.205 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:43.205 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:39:43.205 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:43.205 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:43.205 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:39:43.205 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:43.205 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:43.205 17:55:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:39:43.462 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:39:43.462 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:43.462 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:39:43.462 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:39:43.462 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:39:43.462 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:39:43.462 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:39:43.462 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:39:43.719 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:43.719 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:43.719 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:39:43.719 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:43.719 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:43.719 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:39:43.719 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:43.719 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:43.719 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:39:43.719 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:43.719 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:43.719 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:39:43.719 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:43.719 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:43.719 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:39:43.719 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:43.719 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:43.719 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:39:43.719 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:43.719 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:43.719 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:39:43.719 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:43.719 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:43.719 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:39:43.977 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:39:43.977 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:43.977 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:39:43.977 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:39:43.977 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:39:43.977 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:39:43.977 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:39:43.977 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:39:44.235 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:44.235 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:44.235 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:44.235 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:44.235 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:39:44.235 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:39:44.235 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:44.235 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:44.235 17:55:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:39:44.235 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:44.235 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:44.235 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:39:44.235 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:44.235 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:44.235 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:39:44.235 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:44.235 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:44.235 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:39:44.235 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:44.235 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:44.235 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:39:44.235 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:44.235 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:44.235 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:39:44.492 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:39:44.492 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:39:44.492 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:44.492 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:39:44.492 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:39:44.492 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:39:44.492 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:39:44.492 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:39:44.749 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:44.749 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:44.749 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:39:44.749 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:44.749 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:44.749 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:39:45.006 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:45.006 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:45.006 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:39:45.006 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:45.006 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:45.006 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:39:45.006 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:45.006 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:45.006 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:39:45.006 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:45.006 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:45.006 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:39:45.006 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:45.006 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:45.006 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:39:45.006 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:45.006 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:45.006 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:39:45.264 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:39:45.264 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:39:45.264 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:39:45.264 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:45.264 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:39:45.264 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:39:45.264 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:39:45.264 17:55:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:39:45.521 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:45.521 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:45.521 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:39:45.521 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:45.521 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:45.521 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:39:45.521 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:45.521 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:45.521 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:39:45.521 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:45.521 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:45.521 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:39:45.521 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:45.521 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:45.521 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:39:45.521 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:45.521 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:45.521 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:39:45.521 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:45.521 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:45.521 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:39:45.521 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:45.521 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:45.521 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:39:45.778 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:39:45.778 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:39:45.778 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:39:45.778 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:45.778 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:39:45.778 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:39:45.778 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:39:45.778 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:39:46.035 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:46.035 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:46.035 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:39:46.035 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:46.035 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:46.035 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:39:46.035 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:46.035 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:46.035 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:39:46.035 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:46.035 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:46.035 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:46.035 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:39:46.035 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:46.035 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:39:46.035 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:46.035 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:46.035 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:39:46.035 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:46.035 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:46.035 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:39:46.035 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:46.035 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:46.035 17:55:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:39:46.294 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:39:46.294 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:39:46.294 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:39:46.294 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:46.294 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:39:46.294 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:39:46.294 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:39:46.294 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:39:46.551 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:46.551 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:46.551 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:39:46.551 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:46.551 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:46.551 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:39:46.551 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:46.551 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:46.551 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:39:46.551 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:46.552 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:46.552 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:39:46.809 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:46.809 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:46.809 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:39:46.809 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:46.809 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:46.809 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:46.809 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:46.809 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:39:46.809 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:39:46.809 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:46.809 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:46.809 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:39:47.067 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:39:47.067 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:39:47.067 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:39:47.067 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:47.067 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:39:47.067 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:39:47.067 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:39:47.067 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:39:47.327 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:47.327 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:47.327 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:39:47.327 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:47.327 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:47.327 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:39:47.327 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:47.327 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:47.327 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:39:47.327 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:47.327 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:47.327 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:39:47.327 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:47.327 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:47.327 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:39:47.327 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:47.327 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:47.327 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:39:47.327 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:47.327 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:47.327 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:39:47.327 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:47.327 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:47.327 17:55:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:39:47.624 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:39:47.624 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:39:47.624 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:39:47.624 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:39:47.624 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:39:47.624 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:39:47.624 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:39:47.624 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@68 -- # trap - SIGINT SIGTERM EXIT 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@70 -- # nvmftestfini 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@516 -- # nvmfcleanup 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@121 -- # sync 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@124 -- # set +e 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@125 -- # for i in {1..20} 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:39:47.907 rmmod nvme_tcp 00:39:47.907 rmmod nvme_fabrics 00:39:47.907 rmmod nvme_keyring 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@128 -- # set -e 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@129 -- # return 0 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@517 -- # '[' -n 420301 ']' 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@518 -- # killprocess 420301 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@954 -- # '[' -z 420301 ']' 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@958 -- # kill -0 420301 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@959 -- # uname 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 420301 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@972 -- # echo 'killing process with pid 420301' 00:39:47.907 killing process with pid 420301 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@973 -- # kill 420301 00:39:47.907 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@978 -- # wait 420301 00:39:48.165 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:39:48.165 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:39:48.165 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:39:48.165 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@297 -- # iptr 00:39:48.165 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@791 -- # iptables-save 00:39:48.165 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:39:48.165 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@791 -- # iptables-restore 00:39:48.165 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:39:48.165 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@302 -- # remove_spdk_ns 00:39:48.165 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:39:48.165 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:39:48.165 17:55:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:39:50.702 17:55:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:39:50.702 00:39:50.702 real 0m47.361s 00:39:50.702 user 3m18.238s 00:39:50.702 sys 0m21.512s 00:39:50.702 17:55:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1130 -- # xtrace_disable 00:39:50.702 17:55:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:39:50.702 ************************************ 00:39:50.702 END TEST nvmf_ns_hotplug_stress 00:39:50.702 ************************************ 00:39:50.702 17:55:31 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@23 -- # run_test nvmf_delete_subsystem /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh --transport=tcp --interrupt-mode 00:39:50.702 17:55:31 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:39:50.702 17:55:31 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1111 -- # xtrace_disable 00:39:50.702 17:55:31 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:39:50.702 ************************************ 00:39:50.702 START TEST nvmf_delete_subsystem 00:39:50.702 ************************************ 00:39:50.702 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh --transport=tcp --interrupt-mode 00:39:50.702 * Looking for test storage... 00:39:50.702 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:39:50.702 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:39:50.702 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1711 -- # lcov --version 00:39:50.702 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:39:50.702 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:39:50.702 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:39:50.702 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@333 -- # local ver1 ver1_l 00:39:50.702 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@334 -- # local ver2 ver2_l 00:39:50.702 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@336 -- # IFS=.-: 00:39:50.702 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@336 -- # read -ra ver1 00:39:50.702 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@337 -- # IFS=.-: 00:39:50.702 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@337 -- # read -ra ver2 00:39:50.702 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@338 -- # local 'op=<' 00:39:50.702 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@340 -- # ver1_l=2 00:39:50.702 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@341 -- # ver2_l=1 00:39:50.702 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:39:50.702 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@344 -- # case "$op" in 00:39:50.702 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@345 -- # : 1 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@364 -- # (( v = 0 )) 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@365 -- # decimal 1 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@353 -- # local d=1 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@355 -- # echo 1 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@365 -- # ver1[v]=1 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@366 -- # decimal 2 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@353 -- # local d=2 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@355 -- # echo 2 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@366 -- # ver2[v]=2 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@368 -- # return 0 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:39:50.703 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:39:50.703 --rc genhtml_branch_coverage=1 00:39:50.703 --rc genhtml_function_coverage=1 00:39:50.703 --rc genhtml_legend=1 00:39:50.703 --rc geninfo_all_blocks=1 00:39:50.703 --rc geninfo_unexecuted_blocks=1 00:39:50.703 00:39:50.703 ' 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:39:50.703 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:39:50.703 --rc genhtml_branch_coverage=1 00:39:50.703 --rc genhtml_function_coverage=1 00:39:50.703 --rc genhtml_legend=1 00:39:50.703 --rc geninfo_all_blocks=1 00:39:50.703 --rc geninfo_unexecuted_blocks=1 00:39:50.703 00:39:50.703 ' 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:39:50.703 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:39:50.703 --rc genhtml_branch_coverage=1 00:39:50.703 --rc genhtml_function_coverage=1 00:39:50.703 --rc genhtml_legend=1 00:39:50.703 --rc geninfo_all_blocks=1 00:39:50.703 --rc geninfo_unexecuted_blocks=1 00:39:50.703 00:39:50.703 ' 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:39:50.703 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:39:50.703 --rc genhtml_branch_coverage=1 00:39:50.703 --rc genhtml_function_coverage=1 00:39:50.703 --rc genhtml_legend=1 00:39:50.703 --rc geninfo_all_blocks=1 00:39:50.703 --rc geninfo_unexecuted_blocks=1 00:39:50.703 00:39:50.703 ' 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@7 -- # uname -s 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@15 -- # shopt -s extglob 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- paths/export.sh@5 -- # export PATH 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@51 -- # : 0 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@55 -- # have_pci_nics=0 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@12 -- # nvmftestinit 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@476 -- # prepare_net_devs 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@438 -- # local -g is_hw=no 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@440 -- # remove_spdk_ns 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:39:50.703 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:39:50.704 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:39:50.704 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:39:50.704 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@309 -- # xtrace_disable 00:39:50.704 17:55:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@315 -- # pci_devs=() 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@315 -- # local -a pci_devs 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@316 -- # pci_net_devs=() 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@317 -- # pci_drivers=() 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@317 -- # local -A pci_drivers 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@319 -- # net_devs=() 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@319 -- # local -ga net_devs 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@320 -- # e810=() 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@320 -- # local -ga e810 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@321 -- # x722=() 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@321 -- # local -ga x722 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@322 -- # mlx=() 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@322 -- # local -ga mlx 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:39:52.608 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:39:52.608 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@418 -- # [[ up == up ]] 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:39:52.608 Found net devices under 0000:0a:00.0: cvl_0_0 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:39:52.608 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@418 -- # [[ up == up ]] 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:39:52.609 Found net devices under 0000:0a:00.1: cvl_0_1 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@442 -- # is_hw=yes 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:39:52.609 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:39:52.609 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.244 ms 00:39:52.609 00:39:52.609 --- 10.0.0.2 ping statistics --- 00:39:52.609 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:39:52.609 rtt min/avg/max/mdev = 0.244/0.244/0.244/0.000 ms 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:39:52.609 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:39:52.609 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.128 ms 00:39:52.609 00:39:52.609 --- 10.0.0.1 ping statistics --- 00:39:52.609 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:39:52.609 rtt min/avg/max/mdev = 0.128/0.128/0.128/0.000 ms 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@450 -- # return 0 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@13 -- # nvmfappstart -m 0x3 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@726 -- # xtrace_disable 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@509 -- # nvmfpid=427477 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x3 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@510 -- # waitforlisten 427477 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@835 -- # '[' -z 427477 ']' 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@840 -- # local max_retries=100 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:39:52.609 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@844 -- # xtrace_disable 00:39:52.609 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:39:52.609 [2024-12-06 17:55:34.435570] thread.c:3005:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:39:52.609 [2024-12-06 17:55:34.436656] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:39:52.609 [2024-12-06 17:55:34.436731] [ DPDK EAL parameters: nvmf -c 0x3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:39:52.869 [2024-12-06 17:55:34.508030] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:39:52.869 [2024-12-06 17:55:34.552349] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:39:52.869 [2024-12-06 17:55:34.552412] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:39:52.869 [2024-12-06 17:55:34.552440] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:39:52.869 [2024-12-06 17:55:34.552451] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:39:52.869 [2024-12-06 17:55:34.552462] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:39:52.869 [2024-12-06 17:55:34.553960] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:39:52.869 [2024-12-06 17:55:34.553980] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:39:52.869 [2024-12-06 17:55:34.640159] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:39:52.869 [2024-12-06 17:55:34.640159] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:39:52.869 [2024-12-06 17:55:34.640417] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:39:52.869 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:39:52.869 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@868 -- # return 0 00:39:52.869 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:39:52.869 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@732 -- # xtrace_disable 00:39:52.869 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:39:52.869 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:39:52.869 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@15 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:39:52.869 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:39:52.869 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:39:52.869 [2024-12-06 17:55:34.690619] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:39:52.869 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:39:52.869 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@16 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:39:52.869 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:39:52.869 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:39:52.869 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:39:52.869 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@17 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:39:52.869 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:39:52.869 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:39:53.127 [2024-12-06 17:55:34.706961] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:39:53.127 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:39:53.127 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@18 -- # rpc_cmd bdev_null_create NULL1 1000 512 00:39:53.127 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:39:53.127 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:39:53.127 NULL1 00:39:53.127 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:39:53.127 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@23 -- # rpc_cmd bdev_delay_create -b NULL1 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:39:53.127 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:39:53.127 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:39:53.127 Delay0 00:39:53.127 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:39:53.127 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:53.127 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:39:53.128 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:39:53.128 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:39:53.128 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@28 -- # perf_pid=427570 00:39:53.128 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@30 -- # sleep 2 00:39:53.128 17:55:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0xC -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 5 -q 128 -w randrw -M 70 -o 512 -P 4 00:39:53.128 [2024-12-06 17:55:34.791702] subsystem.c:1641:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:39:55.026 17:55:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@32 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:39:55.026 17:55:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:39:55.026 17:55:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 starting I/O failed: -6 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 starting I/O failed: -6 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 starting I/O failed: -6 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 starting I/O failed: -6 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 starting I/O failed: -6 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 starting I/O failed: -6 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 starting I/O failed: -6 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 starting I/O failed: -6 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 starting I/O failed: -6 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 starting I/O failed: -6 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 starting I/O failed: -6 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 starting I/O failed: -6 00:39:55.285 [2024-12-06 17:55:36.877768] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7fa464000c40 is same with the state(6) to be set 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 starting I/O failed: -6 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 starting I/O failed: -6 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 starting I/O failed: -6 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 starting I/O failed: -6 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 starting I/O failed: -6 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 starting I/O failed: -6 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Read completed with error (sct=0, sc=8) 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.285 Write completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Write completed with error (sct=0, sc=8) 00:39:55.286 Write completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Write completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Write completed with error (sct=0, sc=8) 00:39:55.286 Write completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Write completed with error (sct=0, sc=8) 00:39:55.286 Write completed with error (sct=0, sc=8) 00:39:55.286 Write completed with error (sct=0, sc=8) 00:39:55.286 Write completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Write completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Write completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Write completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Write completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Write completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 [2024-12-06 17:55:36.878920] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7fa46400d4b0 is same with the state(6) to be set 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Write completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Write completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Write completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Write completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Write completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Write completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Write completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Write completed with error (sct=0, sc=8) 00:39:55.286 Write completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Write completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Write completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Write completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 Read completed with error (sct=0, sc=8) 00:39:55.286 starting I/O failed: -6 00:39:55.286 starting I/O failed: -6 00:39:55.286 starting I/O failed: -6 00:39:55.286 starting I/O failed: -6 00:39:55.286 starting I/O failed: -6 00:39:55.286 starting I/O failed: -6 00:39:55.286 starting I/O failed: -6 00:39:55.286 starting I/O failed: -6 00:39:56.220 [2024-12-06 17:55:37.846959] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x642190 is same with the state(6) to be set 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 [2024-12-06 17:55:37.879873] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x643f70 is same with the state(6) to be set 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 [2024-12-06 17:55:37.881808] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x644330 is same with the state(6) to be set 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 [2024-12-06 17:55:37.882204] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7fa46400d020 is same with the state(6) to be set 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Write completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.220 Read completed with error (sct=0, sc=8) 00:39:56.221 Read completed with error (sct=0, sc=8) 00:39:56.221 Read completed with error (sct=0, sc=8) 00:39:56.221 Write completed with error (sct=0, sc=8) 00:39:56.221 Read completed with error (sct=0, sc=8) 00:39:56.221 Read completed with error (sct=0, sc=8) 00:39:56.221 Read completed with error (sct=0, sc=8) 00:39:56.221 Read completed with error (sct=0, sc=8) 00:39:56.221 Read completed with error (sct=0, sc=8) 00:39:56.221 Read completed with error (sct=0, sc=8) 00:39:56.221 Read completed with error (sct=0, sc=8) 00:39:56.221 Read completed with error (sct=0, sc=8) 00:39:56.221 Read completed with error (sct=0, sc=8) 00:39:56.221 Read completed with error (sct=0, sc=8) 00:39:56.221 Write completed with error (sct=0, sc=8) 00:39:56.221 Read completed with error (sct=0, sc=8) 00:39:56.221 Read completed with error (sct=0, sc=8) 00:39:56.221 Read completed with error (sct=0, sc=8) 00:39:56.221 Read completed with error (sct=0, sc=8) 00:39:56.221 [2024-12-06 17:55:37.882385] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7fa46400d7e0 is same with the state(6) to be set 00:39:56.221 Initializing NVMe Controllers 00:39:56.221 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:39:56.221 Controller IO queue size 128, less than required. 00:39:56.221 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:39:56.221 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 2 00:39:56.221 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 3 00:39:56.221 Initialization complete. Launching workers. 00:39:56.221 ======================================================== 00:39:56.221 Latency(us) 00:39:56.221 Device Information : IOPS MiB/s Average min max 00:39:56.221 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 2: 185.63 0.09 908666.05 632.09 1011861.15 00:39:56.221 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 3: 171.24 0.08 892684.93 682.84 1043208.78 00:39:56.221 ======================================================== 00:39:56.221 Total : 356.87 0.17 900997.78 632.09 1043208.78 00:39:56.221 00:39:56.221 17:55:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:39:56.221 [2024-12-06 17:55:37.883153] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x642190 (9): Bad file descriptor 00:39:56.221 17:55:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@34 -- # delay=0 00:39:56.221 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf: errors occurred 00:39:56.221 17:55:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@35 -- # kill -0 427570 00:39:56.221 17:55:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@36 -- # sleep 0.5 00:39:56.787 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@38 -- # (( delay++ > 30 )) 00:39:56.787 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@35 -- # kill -0 427570 00:39:56.787 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh: line 35: kill: (427570) - No such process 00:39:56.787 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@45 -- # NOT wait 427570 00:39:56.787 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@652 -- # local es=0 00:39:56.787 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@654 -- # valid_exec_arg wait 427570 00:39:56.787 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@640 -- # local arg=wait 00:39:56.787 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:39:56.787 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@644 -- # type -t wait 00:39:56.787 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:39:56.787 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@655 -- # wait 427570 00:39:56.787 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@655 -- # es=1 00:39:56.787 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:39:56.787 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:39:56.787 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:39:56.787 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@48 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:39:56.787 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:39:56.787 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:39:56.787 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:39:56.787 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@49 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:39:56.787 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:39:56.787 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:39:56.787 [2024-12-06 17:55:38.402837] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:39:56.787 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:39:56.787 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@50 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:39:56.788 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@563 -- # xtrace_disable 00:39:56.788 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:39:56.788 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:39:56.788 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@54 -- # perf_pid=428022 00:39:56.788 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@56 -- # delay=0 00:39:56.788 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0xC -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 3 -q 128 -w randrw -M 70 -o 512 -P 4 00:39:56.788 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 428022 00:39:56.788 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:39:56.788 [2024-12-06 17:55:38.461809] subsystem.c:1641:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:39:57.353 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:39:57.353 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 428022 00:39:57.353 17:55:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:39:57.611 17:55:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:39:57.611 17:55:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 428022 00:39:57.611 17:55:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:39:58.176 17:55:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:39:58.176 17:55:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 428022 00:39:58.176 17:55:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:39:58.742 17:55:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:39:58.742 17:55:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 428022 00:39:58.742 17:55:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:39:59.307 17:55:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:39:59.307 17:55:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 428022 00:39:59.307 17:55:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:39:59.872 17:55:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:39:59.872 17:55:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 428022 00:39:59.872 17:55:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:39:59.872 Initializing NVMe Controllers 00:39:59.872 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:39:59.872 Controller IO queue size 128, less than required. 00:39:59.872 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:39:59.872 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 2 00:39:59.872 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 3 00:39:59.872 Initialization complete. Launching workers. 00:39:59.872 ======================================================== 00:39:59.872 Latency(us) 00:39:59.872 Device Information : IOPS MiB/s Average min max 00:39:59.872 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 2: 128.00 0.06 1004611.77 1000192.00 1011118.26 00:39:59.872 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 3: 128.00 0.06 1004480.96 1000181.45 1011583.07 00:39:59.872 ======================================================== 00:39:59.872 Total : 256.00 0.12 1004546.37 1000181.45 1011583.07 00:39:59.872 00:40:00.130 17:55:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:40:00.130 17:55:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 428022 00:40:00.130 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh: line 57: kill: (428022) - No such process 00:40:00.130 17:55:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@67 -- # wait 428022 00:40:00.130 17:55:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:40:00.130 17:55:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@71 -- # nvmftestfini 00:40:00.130 17:55:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@516 -- # nvmfcleanup 00:40:00.130 17:55:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@121 -- # sync 00:40:00.130 17:55:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:40:00.130 17:55:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@124 -- # set +e 00:40:00.130 17:55:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@125 -- # for i in {1..20} 00:40:00.130 17:55:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:40:00.130 rmmod nvme_tcp 00:40:00.130 rmmod nvme_fabrics 00:40:00.130 rmmod nvme_keyring 00:40:00.389 17:55:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:40:00.389 17:55:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@128 -- # set -e 00:40:00.389 17:55:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@129 -- # return 0 00:40:00.389 17:55:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@517 -- # '[' -n 427477 ']' 00:40:00.389 17:55:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@518 -- # killprocess 427477 00:40:00.389 17:55:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@954 -- # '[' -z 427477 ']' 00:40:00.389 17:55:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@958 -- # kill -0 427477 00:40:00.389 17:55:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@959 -- # uname 00:40:00.389 17:55:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:40:00.389 17:55:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 427477 00:40:00.389 17:55:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:40:00.389 17:55:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:40:00.389 17:55:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@972 -- # echo 'killing process with pid 427477' 00:40:00.389 killing process with pid 427477 00:40:00.389 17:55:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@973 -- # kill 427477 00:40:00.389 17:55:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@978 -- # wait 427477 00:40:00.389 17:55:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:40:00.389 17:55:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:40:00.389 17:55:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:40:00.389 17:55:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@297 -- # iptr 00:40:00.389 17:55:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@791 -- # iptables-save 00:40:00.389 17:55:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:40:00.389 17:55:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@791 -- # iptables-restore 00:40:00.647 17:55:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:40:00.647 17:55:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@302 -- # remove_spdk_ns 00:40:00.647 17:55:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:40:00.647 17:55:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:40:00.647 17:55:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:40:02.551 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:40:02.551 00:40:02.551 real 0m12.265s 00:40:02.551 user 0m24.490s 00:40:02.551 sys 0m3.706s 00:40:02.551 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1130 -- # xtrace_disable 00:40:02.551 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:40:02.551 ************************************ 00:40:02.551 END TEST nvmf_delete_subsystem 00:40:02.551 ************************************ 00:40:02.551 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@26 -- # run_test nvmf_host_management /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh --transport=tcp --interrupt-mode 00:40:02.551 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:40:02.551 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1111 -- # xtrace_disable 00:40:02.551 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:40:02.551 ************************************ 00:40:02.551 START TEST nvmf_host_management 00:40:02.551 ************************************ 00:40:02.551 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh --transport=tcp --interrupt-mode 00:40:02.551 * Looking for test storage... 00:40:02.551 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:40:02.551 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:40:02.551 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1711 -- # lcov --version 00:40:02.551 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:40:02.809 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:40:02.809 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:40:02.809 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@333 -- # local ver1 ver1_l 00:40:02.809 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@334 -- # local ver2 ver2_l 00:40:02.809 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@336 -- # IFS=.-: 00:40:02.809 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@336 -- # read -ra ver1 00:40:02.809 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@337 -- # IFS=.-: 00:40:02.809 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@337 -- # read -ra ver2 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@338 -- # local 'op=<' 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@340 -- # ver1_l=2 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@341 -- # ver2_l=1 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@344 -- # case "$op" in 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@345 -- # : 1 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@364 -- # (( v = 0 )) 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@365 -- # decimal 1 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@353 -- # local d=1 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@355 -- # echo 1 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@365 -- # ver1[v]=1 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@366 -- # decimal 2 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@353 -- # local d=2 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@355 -- # echo 2 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@366 -- # ver2[v]=2 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@368 -- # return 0 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:40:02.810 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:40:02.810 --rc genhtml_branch_coverage=1 00:40:02.810 --rc genhtml_function_coverage=1 00:40:02.810 --rc genhtml_legend=1 00:40:02.810 --rc geninfo_all_blocks=1 00:40:02.810 --rc geninfo_unexecuted_blocks=1 00:40:02.810 00:40:02.810 ' 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:40:02.810 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:40:02.810 --rc genhtml_branch_coverage=1 00:40:02.810 --rc genhtml_function_coverage=1 00:40:02.810 --rc genhtml_legend=1 00:40:02.810 --rc geninfo_all_blocks=1 00:40:02.810 --rc geninfo_unexecuted_blocks=1 00:40:02.810 00:40:02.810 ' 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:40:02.810 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:40:02.810 --rc genhtml_branch_coverage=1 00:40:02.810 --rc genhtml_function_coverage=1 00:40:02.810 --rc genhtml_legend=1 00:40:02.810 --rc geninfo_all_blocks=1 00:40:02.810 --rc geninfo_unexecuted_blocks=1 00:40:02.810 00:40:02.810 ' 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:40:02.810 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:40:02.810 --rc genhtml_branch_coverage=1 00:40:02.810 --rc genhtml_function_coverage=1 00:40:02.810 --rc genhtml_legend=1 00:40:02.810 --rc geninfo_all_blocks=1 00:40:02.810 --rc geninfo_unexecuted_blocks=1 00:40:02.810 00:40:02.810 ' 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@7 -- # uname -s 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@15 -- # shopt -s extglob 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- paths/export.sh@5 -- # export PATH 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@51 -- # : 0 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@55 -- # have_pci_nics=0 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@11 -- # MALLOC_BDEV_SIZE=64 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@105 -- # nvmftestinit 00:40:02.810 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:40:02.811 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:40:02.811 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@476 -- # prepare_net_devs 00:40:02.811 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@438 -- # local -g is_hw=no 00:40:02.811 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@440 -- # remove_spdk_ns 00:40:02.811 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:40:02.811 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:40:02.811 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:40:02.811 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:40:02.811 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:40:02.811 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@309 -- # xtrace_disable 00:40:02.811 17:55:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:40:04.710 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@315 -- # pci_devs=() 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@315 -- # local -a pci_devs 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@316 -- # pci_net_devs=() 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@317 -- # pci_drivers=() 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@317 -- # local -A pci_drivers 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@319 -- # net_devs=() 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@319 -- # local -ga net_devs 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@320 -- # e810=() 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@320 -- # local -ga e810 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@321 -- # x722=() 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@321 -- # local -ga x722 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@322 -- # mlx=() 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@322 -- # local -ga mlx 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:40:04.969 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:40:04.969 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:40:04.969 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@418 -- # [[ up == up ]] 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:40:04.970 Found net devices under 0000:0a:00.0: cvl_0_0 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@418 -- # [[ up == up ]] 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:40:04.970 Found net devices under 0000:0a:00.1: cvl_0_1 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@442 -- # is_hw=yes 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:40:04.970 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:40:04.970 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.201 ms 00:40:04.970 00:40:04.970 --- 10.0.0.2 ping statistics --- 00:40:04.970 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:40:04.970 rtt min/avg/max/mdev = 0.201/0.201/0.201/0.000 ms 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:40:04.970 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:40:04.970 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.161 ms 00:40:04.970 00:40:04.970 --- 10.0.0.1 ping statistics --- 00:40:04.970 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:40:04.970 rtt min/avg/max/mdev = 0.161/0.161/0.161/0.000 ms 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@450 -- # return 0 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@107 -- # nvmf_host_management 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@69 -- # starttarget 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@16 -- # nvmfappstart -m 0x1E 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@726 -- # xtrace_disable 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@509 -- # nvmfpid=430359 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@510 -- # waitforlisten 430359 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x1E 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@835 -- # '[' -z 430359 ']' 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@840 -- # local max_retries=100 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:40:04.970 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@844 -- # xtrace_disable 00:40:04.970 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:40:04.970 [2024-12-06 17:55:46.770120] thread.c:3005:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:40:04.970 [2024-12-06 17:55:46.771201] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:40:04.970 [2024-12-06 17:55:46.771255] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:40:05.228 [2024-12-06 17:55:46.842738] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:40:05.228 [2024-12-06 17:55:46.889043] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:40:05.228 [2024-12-06 17:55:46.889094] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:40:05.228 [2024-12-06 17:55:46.889118] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:40:05.228 [2024-12-06 17:55:46.889130] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:40:05.228 [2024-12-06 17:55:46.889140] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:40:05.228 [2024-12-06 17:55:46.890759] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:40:05.228 [2024-12-06 17:55:46.890819] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:40:05.228 [2024-12-06 17:55:46.890816] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:40:05.228 [2024-12-06 17:55:46.890792] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:40:05.228 [2024-12-06 17:55:46.975595] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:40:05.228 [2024-12-06 17:55:46.975844] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:40:05.228 [2024-12-06 17:55:46.976156] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:40:05.228 [2024-12-06 17:55:46.976849] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:40:05.228 [2024-12-06 17:55:46.977089] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_003) to intr mode from intr mode. 00:40:05.228 17:55:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:40:05.228 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@868 -- # return 0 00:40:05.228 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:40:05.228 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@732 -- # xtrace_disable 00:40:05.228 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:40:05.228 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:40:05.228 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:40:05.228 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@563 -- # xtrace_disable 00:40:05.229 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:40:05.229 [2024-12-06 17:55:47.031501] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:40:05.229 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:40:05.229 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@20 -- # timing_enter create_subsystem 00:40:05.229 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@726 -- # xtrace_disable 00:40:05.229 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:40:05.229 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@22 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:40:05.229 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@23 -- # cat 00:40:05.229 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@30 -- # rpc_cmd 00:40:05.229 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@563 -- # xtrace_disable 00:40:05.229 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:40:05.486 Malloc0 00:40:05.486 [2024-12-06 17:55:47.099739] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:40:05.486 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:40:05.486 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@31 -- # timing_exit create_subsystems 00:40:05.486 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@732 -- # xtrace_disable 00:40:05.486 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:40:05.486 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@73 -- # perfpid=430402 00:40:05.486 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@74 -- # waitforlisten 430402 /var/tmp/bdevperf.sock 00:40:05.486 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@835 -- # '[' -z 430402 ']' 00:40:05.486 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:40:05.486 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@72 -- # gen_nvmf_target_json 0 00:40:05.486 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock --json /dev/fd/63 -q 64 -o 65536 -w verify -t 10 00:40:05.486 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@840 -- # local max_retries=100 00:40:05.486 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:40:05.486 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:40:05.486 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@560 -- # config=() 00:40:05.486 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@844 -- # xtrace_disable 00:40:05.486 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@560 -- # local subsystem config 00:40:05.486 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:40:05.486 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:40:05.486 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:40:05.486 { 00:40:05.486 "params": { 00:40:05.486 "name": "Nvme$subsystem", 00:40:05.486 "trtype": "$TEST_TRANSPORT", 00:40:05.486 "traddr": "$NVMF_FIRST_TARGET_IP", 00:40:05.486 "adrfam": "ipv4", 00:40:05.486 "trsvcid": "$NVMF_PORT", 00:40:05.486 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:40:05.486 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:40:05.486 "hdgst": ${hdgst:-false}, 00:40:05.486 "ddgst": ${ddgst:-false} 00:40:05.486 }, 00:40:05.486 "method": "bdev_nvme_attach_controller" 00:40:05.486 } 00:40:05.486 EOF 00:40:05.486 )") 00:40:05.486 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@582 -- # cat 00:40:05.486 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@584 -- # jq . 00:40:05.486 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@585 -- # IFS=, 00:40:05.486 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:40:05.486 "params": { 00:40:05.486 "name": "Nvme0", 00:40:05.486 "trtype": "tcp", 00:40:05.486 "traddr": "10.0.0.2", 00:40:05.486 "adrfam": "ipv4", 00:40:05.486 "trsvcid": "4420", 00:40:05.486 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:40:05.486 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:40:05.486 "hdgst": false, 00:40:05.486 "ddgst": false 00:40:05.486 }, 00:40:05.486 "method": "bdev_nvme_attach_controller" 00:40:05.486 }' 00:40:05.486 [2024-12-06 17:55:47.182096] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:40:05.486 [2024-12-06 17:55:47.182177] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid430402 ] 00:40:05.486 [2024-12-06 17:55:47.256548] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:40:05.486 [2024-12-06 17:55:47.303786] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:40:06.052 Running I/O for 10 seconds... 00:40:06.052 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:40:06.052 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@868 -- # return 0 00:40:06.052 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@75 -- # rpc_cmd -s /var/tmp/bdevperf.sock framework_wait_init 00:40:06.052 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@563 -- # xtrace_disable 00:40:06.052 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:40:06.052 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:40:06.052 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@78 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; kill -9 $perfpid || true; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:40:06.052 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@80 -- # waitforio /var/tmp/bdevperf.sock Nvme0n1 00:40:06.052 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@45 -- # '[' -z /var/tmp/bdevperf.sock ']' 00:40:06.052 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@49 -- # '[' -z Nvme0n1 ']' 00:40:06.052 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@52 -- # local ret=1 00:40:06.052 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@53 -- # local i 00:40:06.052 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@54 -- # (( i = 10 )) 00:40:06.052 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@54 -- # (( i != 0 )) 00:40:06.052 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@55 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme0n1 00:40:06.052 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@55 -- # jq -r '.bdevs[0].num_read_ops' 00:40:06.052 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@563 -- # xtrace_disable 00:40:06.052 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:40:06.052 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:40:06.052 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@55 -- # read_io_count=67 00:40:06.052 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@58 -- # '[' 67 -ge 100 ']' 00:40:06.052 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@62 -- # sleep 0.25 00:40:06.310 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@54 -- # (( i-- )) 00:40:06.310 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@54 -- # (( i != 0 )) 00:40:06.310 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@55 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme0n1 00:40:06.310 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@55 -- # jq -r '.bdevs[0].num_read_ops' 00:40:06.311 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@563 -- # xtrace_disable 00:40:06.311 17:55:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:40:06.311 17:55:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:40:06.311 17:55:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@55 -- # read_io_count=515 00:40:06.311 17:55:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@58 -- # '[' 515 -ge 100 ']' 00:40:06.311 17:55:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@59 -- # ret=0 00:40:06.311 17:55:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@60 -- # break 00:40:06.311 17:55:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@64 -- # return 0 00:40:06.311 17:55:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@84 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2016-06.io.spdk:cnode0 nqn.2016-06.io.spdk:host0 00:40:06.311 17:55:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@563 -- # xtrace_disable 00:40:06.311 17:55:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:40:06.311 [2024-12-06 17:55:48.031907] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:76800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.031965] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.031996] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:76928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032013] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032029] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:77056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032044] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032060] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:77184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032074] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032090] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:77312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032104] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032119] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:77440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032133] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032148] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:77568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032162] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032178] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:77696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032192] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032207] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:77824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032222] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032238] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:77952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032252] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032268] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:78080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032282] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032298] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:78208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032312] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032337] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:78336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032353] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032368] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:78464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032383] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032398] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:78592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032412] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032427] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:78720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032441] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032457] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:78848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032471] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032487] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:78976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032501] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032516] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:79104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032530] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032546] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:79232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032560] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032576] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:79360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032590] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032605] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:79488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032620] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032635] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:79616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032660] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032686] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:79744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032701] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032717] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:79872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032735] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032751] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:80000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032765] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032780] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:80128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032794] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032809] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:80256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032823] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032838] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:80384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032852] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032867] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:80512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032881] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032896] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:80640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032911] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032926] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:80768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032940] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032965] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:80896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.032979] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.032994] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:81024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.311 [2024-12-06 17:55:48.033008] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.311 [2024-12-06 17:55:48.033023] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:81152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033037] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033052] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:81280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033066] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033081] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:81408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033095] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033114] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:81536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033129] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033144] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:81664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033158] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033174] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:81792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033188] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033203] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:73728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033217] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033233] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:73856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033247] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033262] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:73984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033276] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033291] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:74112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033306] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033321] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:74240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033335] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033351] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:74368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033366] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033381] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:74496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033395] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033410] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:74624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033425] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033440] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:74752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033454] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033469] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:74880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033487] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033504] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:75008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033519] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033534] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:75136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033548] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033564] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:75264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033578] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033594] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:75392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033609] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033624] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:75520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033638] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033657] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:75648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033683] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033700] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:75776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033716] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033731] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:75904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033745] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033760] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:76032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033774] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033789] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:76160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033804] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033819] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:76288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033834] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033849] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:76416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033863] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033882] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:76544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033897] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.033912] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:76672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:40:06.312 [2024-12-06 17:55:48.033927] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.035137] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:40:06.312 17:55:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:40:06.312 17:55:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@85 -- # rpc_cmd nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode0 nqn.2016-06.io.spdk:host0 00:40:06.312 17:55:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@563 -- # xtrace_disable 00:40:06.312 17:55:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:40:06.312 task offset: 76800 on job bdev=Nvme0n1 fails 00:40:06.312 00:40:06.312 Latency(us) 00:40:06.312 [2024-12-06T16:55:48.151Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:40:06.312 Job: Nvme0n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:40:06.312 Job: Nvme0n1 ended in about 0.39 seconds with error 00:40:06.312 Verification LBA range: start 0x0 length 0x400 00:40:06.312 Nvme0n1 : 0.39 1472.32 92.02 163.59 0.00 37989.77 2730.67 37476.88 00:40:06.312 [2024-12-06T16:55:48.151Z] =================================================================================================================== 00:40:06.312 [2024-12-06T16:55:48.151Z] Total : 1472.32 92.02 163.59 0.00 37989.77 2730.67 37476.88 00:40:06.312 [2024-12-06 17:55:48.037070] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:40:06.312 [2024-12-06 17:55:48.037111] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x130e980 (9): Bad file descriptor 00:40:06.312 [2024-12-06 17:55:48.038226] ctrlr.c: 825:nvmf_qpair_access_allowed: *ERROR*: Subsystem 'nqn.2016-06.io.spdk:cnode0' does not allow host 'nqn.2016-06.io.spdk:host0' 00:40:06.312 [2024-12-06 17:55:48.038322] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:0 cid:3 SGL DATA BLOCK OFFSET 0x0 len:0x400 00:40:06.312 [2024-12-06 17:55:48.038350] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND SPECIFIC (01/84) qid:0 cid:3 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:40:06.312 [2024-12-06 17:55:48.038377] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode0 00:40:06.312 [2024-12-06 17:55:48.038394] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 132 00:40:06.312 [2024-12-06 17:55:48.038407] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:40:06.312 [2024-12-06 17:55:48.038419] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x130e980 00:40:06.312 [2024-12-06 17:55:48.038453] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x130e980 (9): Bad file descriptor 00:40:06.313 [2024-12-06 17:55:48.038478] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:40:06.313 [2024-12-06 17:55:48.038493] nvme_ctrlr.c:1826:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:40:06.313 [2024-12-06 17:55:48.038509] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:40:06.313 [2024-12-06 17:55:48.038525] bdev_nvme.c:2284:bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:40:06.313 17:55:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:40:06.313 17:55:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@87 -- # sleep 1 00:40:07.246 17:55:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@91 -- # kill -9 430402 00:40:07.246 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh: line 91: kill: (430402) - No such process 00:40:07.246 17:55:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@91 -- # true 00:40:07.246 17:55:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@97 -- # rm -f /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 /var/tmp/spdk_cpu_lock_003 /var/tmp/spdk_cpu_lock_004 00:40:07.246 17:55:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -q 64 -o 65536 -w verify -t 1 00:40:07.246 17:55:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@100 -- # gen_nvmf_target_json 0 00:40:07.246 17:55:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@560 -- # config=() 00:40:07.246 17:55:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@560 -- # local subsystem config 00:40:07.246 17:55:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:40:07.246 17:55:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:40:07.246 { 00:40:07.246 "params": { 00:40:07.246 "name": "Nvme$subsystem", 00:40:07.246 "trtype": "$TEST_TRANSPORT", 00:40:07.246 "traddr": "$NVMF_FIRST_TARGET_IP", 00:40:07.246 "adrfam": "ipv4", 00:40:07.246 "trsvcid": "$NVMF_PORT", 00:40:07.246 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:40:07.246 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:40:07.246 "hdgst": ${hdgst:-false}, 00:40:07.246 "ddgst": ${ddgst:-false} 00:40:07.246 }, 00:40:07.246 "method": "bdev_nvme_attach_controller" 00:40:07.246 } 00:40:07.246 EOF 00:40:07.246 )") 00:40:07.246 17:55:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@582 -- # cat 00:40:07.246 17:55:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@584 -- # jq . 00:40:07.246 17:55:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@585 -- # IFS=, 00:40:07.246 17:55:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:40:07.246 "params": { 00:40:07.246 "name": "Nvme0", 00:40:07.246 "trtype": "tcp", 00:40:07.246 "traddr": "10.0.0.2", 00:40:07.246 "adrfam": "ipv4", 00:40:07.246 "trsvcid": "4420", 00:40:07.246 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:40:07.246 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:40:07.246 "hdgst": false, 00:40:07.246 "ddgst": false 00:40:07.246 }, 00:40:07.246 "method": "bdev_nvme_attach_controller" 00:40:07.246 }' 00:40:07.504 [2024-12-06 17:55:49.095277] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:40:07.504 [2024-12-06 17:55:49.095370] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid430672 ] 00:40:07.504 [2024-12-06 17:55:49.164431] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:40:07.504 [2024-12-06 17:55:49.211530] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:40:07.762 Running I/O for 1 seconds... 00:40:08.696 1657.00 IOPS, 103.56 MiB/s 00:40:08.696 Latency(us) 00:40:08.696 [2024-12-06T16:55:50.535Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:40:08.696 Job: Nvme0n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:40:08.696 Verification LBA range: start 0x0 length 0x400 00:40:08.696 Nvme0n1 : 1.02 1686.76 105.42 0.00 0.00 37306.40 2924.85 34175.81 00:40:08.696 [2024-12-06T16:55:50.535Z] =================================================================================================================== 00:40:08.696 [2024-12-06T16:55:50.536Z] Total : 1686.76 105.42 0.00 0.00 37306.40 2924.85 34175.81 00:40:08.955 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@102 -- # stoptarget 00:40:08.955 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@36 -- # rm -f ./local-job0-0-verify.state 00:40:08.955 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@37 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:40:08.955 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@38 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:40:08.955 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@40 -- # nvmftestfini 00:40:08.955 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@516 -- # nvmfcleanup 00:40:08.955 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@121 -- # sync 00:40:08.955 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:40:08.955 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@124 -- # set +e 00:40:08.955 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@125 -- # for i in {1..20} 00:40:08.955 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:40:08.955 rmmod nvme_tcp 00:40:08.955 rmmod nvme_fabrics 00:40:08.955 rmmod nvme_keyring 00:40:08.955 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:40:08.955 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@128 -- # set -e 00:40:08.955 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@129 -- # return 0 00:40:08.955 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@517 -- # '[' -n 430359 ']' 00:40:08.955 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@518 -- # killprocess 430359 00:40:08.955 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@954 -- # '[' -z 430359 ']' 00:40:08.955 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@958 -- # kill -0 430359 00:40:08.955 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@959 -- # uname 00:40:08.955 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:40:08.955 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 430359 00:40:08.955 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:40:08.955 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:40:08.955 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@972 -- # echo 'killing process with pid 430359' 00:40:08.955 killing process with pid 430359 00:40:08.955 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@973 -- # kill 430359 00:40:08.955 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@978 -- # wait 430359 00:40:09.213 [2024-12-06 17:55:50.948794] app.c: 721:unclaim_cpu_cores: *ERROR*: Failed to unlink lock fd for core 1, errno: 2 00:40:09.213 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:40:09.213 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:40:09.213 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:40:09.213 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@297 -- # iptr 00:40:09.213 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@791 -- # iptables-save 00:40:09.214 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:40:09.214 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@791 -- # iptables-restore 00:40:09.214 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:40:09.214 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@302 -- # remove_spdk_ns 00:40:09.214 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:40:09.214 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:40:09.214 17:55:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@109 -- # trap - SIGINT SIGTERM EXIT 00:40:11.753 00:40:11.753 real 0m8.716s 00:40:11.753 user 0m17.263s 00:40:11.753 sys 0m3.733s 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1130 -- # xtrace_disable 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:40:11.753 ************************************ 00:40:11.753 END TEST nvmf_host_management 00:40:11.753 ************************************ 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@27 -- # run_test nvmf_lvol /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvol.sh --transport=tcp --interrupt-mode 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1111 -- # xtrace_disable 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:40:11.753 ************************************ 00:40:11.753 START TEST nvmf_lvol 00:40:11.753 ************************************ 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvol.sh --transport=tcp --interrupt-mode 00:40:11.753 * Looking for test storage... 00:40:11.753 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1711 -- # lcov --version 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@333 -- # local ver1 ver1_l 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@334 -- # local ver2 ver2_l 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@336 -- # IFS=.-: 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@336 -- # read -ra ver1 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@337 -- # IFS=.-: 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@337 -- # read -ra ver2 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@338 -- # local 'op=<' 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@340 -- # ver1_l=2 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@341 -- # ver2_l=1 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@344 -- # case "$op" in 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@345 -- # : 1 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@364 -- # (( v = 0 )) 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@365 -- # decimal 1 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@353 -- # local d=1 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:40:11.753 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@355 -- # echo 1 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@365 -- # ver1[v]=1 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@366 -- # decimal 2 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@353 -- # local d=2 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@355 -- # echo 2 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@366 -- # ver2[v]=2 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@368 -- # return 0 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:40:11.754 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:40:11.754 --rc genhtml_branch_coverage=1 00:40:11.754 --rc genhtml_function_coverage=1 00:40:11.754 --rc genhtml_legend=1 00:40:11.754 --rc geninfo_all_blocks=1 00:40:11.754 --rc geninfo_unexecuted_blocks=1 00:40:11.754 00:40:11.754 ' 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:40:11.754 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:40:11.754 --rc genhtml_branch_coverage=1 00:40:11.754 --rc genhtml_function_coverage=1 00:40:11.754 --rc genhtml_legend=1 00:40:11.754 --rc geninfo_all_blocks=1 00:40:11.754 --rc geninfo_unexecuted_blocks=1 00:40:11.754 00:40:11.754 ' 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:40:11.754 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:40:11.754 --rc genhtml_branch_coverage=1 00:40:11.754 --rc genhtml_function_coverage=1 00:40:11.754 --rc genhtml_legend=1 00:40:11.754 --rc geninfo_all_blocks=1 00:40:11.754 --rc geninfo_unexecuted_blocks=1 00:40:11.754 00:40:11.754 ' 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:40:11.754 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:40:11.754 --rc genhtml_branch_coverage=1 00:40:11.754 --rc genhtml_function_coverage=1 00:40:11.754 --rc genhtml_legend=1 00:40:11.754 --rc geninfo_all_blocks=1 00:40:11.754 --rc geninfo_unexecuted_blocks=1 00:40:11.754 00:40:11.754 ' 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@7 -- # uname -s 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@15 -- # shopt -s extglob 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- paths/export.sh@5 -- # export PATH 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@51 -- # : 0 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@55 -- # have_pci_nics=0 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@11 -- # MALLOC_BDEV_SIZE=64 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@13 -- # LVOL_BDEV_INIT_SIZE=20 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@14 -- # LVOL_BDEV_FINAL_SIZE=30 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@16 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@18 -- # nvmftestinit 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@476 -- # prepare_net_devs 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@438 -- # local -g is_hw=no 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@440 -- # remove_spdk_ns 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@309 -- # xtrace_disable 00:40:11.754 17:55:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@315 -- # pci_devs=() 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@315 -- # local -a pci_devs 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@316 -- # pci_net_devs=() 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@317 -- # pci_drivers=() 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@317 -- # local -A pci_drivers 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@319 -- # net_devs=() 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@319 -- # local -ga net_devs 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@320 -- # e810=() 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@320 -- # local -ga e810 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@321 -- # x722=() 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@321 -- # local -ga x722 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@322 -- # mlx=() 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@322 -- # local -ga mlx 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:40:13.658 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:40:13.658 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:40:13.658 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@418 -- # [[ up == up ]] 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:40:13.659 Found net devices under 0000:0a:00.0: cvl_0_0 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@418 -- # [[ up == up ]] 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:40:13.659 Found net devices under 0000:0a:00.1: cvl_0_1 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@442 -- # is_hw=yes 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:40:13.659 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:40:13.659 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.270 ms 00:40:13.659 00:40:13.659 --- 10.0.0.2 ping statistics --- 00:40:13.659 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:40:13.659 rtt min/avg/max/mdev = 0.270/0.270/0.270/0.000 ms 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:40:13.659 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:40:13.659 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.114 ms 00:40:13.659 00:40:13.659 --- 10.0.0.1 ping statistics --- 00:40:13.659 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:40:13.659 rtt min/avg/max/mdev = 0.114/0.114/0.114/0.000 ms 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@450 -- # return 0 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@19 -- # nvmfappstart -m 0x7 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@726 -- # xtrace_disable 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@509 -- # nvmfpid=432875 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x7 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@510 -- # waitforlisten 432875 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@835 -- # '[' -z 432875 ']' 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@840 -- # local max_retries=100 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:40:13.659 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@844 -- # xtrace_disable 00:40:13.659 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:40:13.917 [2024-12-06 17:55:55.504573] thread.c:3005:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:40:13.917 [2024-12-06 17:55:55.505751] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:40:13.917 [2024-12-06 17:55:55.505808] [ DPDK EAL parameters: nvmf -c 0x7 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:40:13.917 [2024-12-06 17:55:55.578884] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:40:13.917 [2024-12-06 17:55:55.627139] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:40:13.917 [2024-12-06 17:55:55.627217] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:40:13.917 [2024-12-06 17:55:55.627230] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:40:13.917 [2024-12-06 17:55:55.627242] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:40:13.917 [2024-12-06 17:55:55.627251] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:40:13.917 [2024-12-06 17:55:55.628898] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:40:13.917 [2024-12-06 17:55:55.628963] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:40:13.917 [2024-12-06 17:55:55.628966] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:40:13.917 [2024-12-06 17:55:55.724746] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:40:13.917 [2024-12-06 17:55:55.724959] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:40:13.917 [2024-12-06 17:55:55.725027] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:40:13.917 [2024-12-06 17:55:55.725255] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:40:13.917 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:40:13.917 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@868 -- # return 0 00:40:13.917 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:40:13.917 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@732 -- # xtrace_disable 00:40:13.917 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:40:14.175 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:40:14.175 17:55:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:40:14.433 [2024-12-06 17:55:56.013660] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:40:14.433 17:55:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:40:14.691 17:55:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@24 -- # base_bdevs='Malloc0 ' 00:40:14.691 17:55:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:40:14.950 17:55:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@25 -- # base_bdevs+=Malloc1 00:40:14.950 17:55:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n raid0 -z 64 -r 0 -b 'Malloc0 Malloc1' 00:40:15.209 17:55:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore raid0 lvs 00:40:15.467 17:55:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@29 -- # lvs=fef6a412-9acf-4010-b743-918fca4d6c57 00:40:15.467 17:55:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u fef6a412-9acf-4010-b743-918fca4d6c57 lvol 20 00:40:15.725 17:55:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@32 -- # lvol=6e9411f0-4440-467c-bcc4-96e3e9d9b77e 00:40:15.725 17:55:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:40:15.983 17:55:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 6e9411f0-4440-467c-bcc4-96e3e9d9b77e 00:40:16.241 17:55:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:40:16.498 [2024-12-06 17:55:58.273865] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:40:16.498 17:55:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:40:16.756 17:55:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@42 -- # perf_pid=433215 00:40:16.756 17:55:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -o 4096 -q 128 -s 512 -w randwrite -t 10 -c 0x18 00:40:16.756 17:55:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@44 -- # sleep 1 00:40:18.130 17:55:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_snapshot 6e9411f0-4440-467c-bcc4-96e3e9d9b77e MY_SNAPSHOT 00:40:18.130 17:55:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@47 -- # snapshot=11a1d847-4ccb-47a1-ad92-bce0345b3920 00:40:18.130 17:55:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_resize 6e9411f0-4440-467c-bcc4-96e3e9d9b77e 30 00:40:18.696 17:56:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@49 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_clone 11a1d847-4ccb-47a1-ad92-bce0345b3920 MY_CLONE 00:40:18.696 17:56:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@49 -- # clone=d7e8a5ef-2bb2-459c-a2f8-2c0923dd4a7e 00:40:18.696 17:56:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_inflate d7e8a5ef-2bb2-459c-a2f8-2c0923dd4a7e 00:40:19.628 17:56:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@53 -- # wait 433215 00:40:27.813 Initializing NVMe Controllers 00:40:27.813 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode0 00:40:27.813 Controller IO queue size 128, less than required. 00:40:27.813 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:40:27.813 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 3 00:40:27.813 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 4 00:40:27.813 Initialization complete. Launching workers. 00:40:27.813 ======================================================== 00:40:27.813 Latency(us) 00:40:27.813 Device Information : IOPS MiB/s Average min max 00:40:27.813 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 from core 3: 10539.49 41.17 12147.87 4639.41 63463.67 00:40:27.813 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 from core 4: 10309.99 40.27 12419.05 6054.74 75177.97 00:40:27.813 ======================================================== 00:40:27.813 Total : 20849.49 81.44 12281.97 4639.41 75177.97 00:40:27.813 00:40:27.813 17:56:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:40:27.813 17:56:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete 6e9411f0-4440-467c-bcc4-96e3e9d9b77e 00:40:27.813 17:56:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u fef6a412-9acf-4010-b743-918fca4d6c57 00:40:28.070 17:56:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@60 -- # rm -f 00:40:28.070 17:56:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@62 -- # trap - SIGINT SIGTERM EXIT 00:40:28.070 17:56:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@64 -- # nvmftestfini 00:40:28.070 17:56:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@516 -- # nvmfcleanup 00:40:28.070 17:56:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@121 -- # sync 00:40:28.070 17:56:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:40:28.070 17:56:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@124 -- # set +e 00:40:28.070 17:56:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@125 -- # for i in {1..20} 00:40:28.070 17:56:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:40:28.070 rmmod nvme_tcp 00:40:28.328 rmmod nvme_fabrics 00:40:28.328 rmmod nvme_keyring 00:40:28.328 17:56:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:40:28.328 17:56:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@128 -- # set -e 00:40:28.328 17:56:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@129 -- # return 0 00:40:28.328 17:56:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@517 -- # '[' -n 432875 ']' 00:40:28.328 17:56:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@518 -- # killprocess 432875 00:40:28.328 17:56:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@954 -- # '[' -z 432875 ']' 00:40:28.328 17:56:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@958 -- # kill -0 432875 00:40:28.328 17:56:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@959 -- # uname 00:40:28.328 17:56:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:40:28.328 17:56:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 432875 00:40:28.328 17:56:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:40:28.328 17:56:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:40:28.328 17:56:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@972 -- # echo 'killing process with pid 432875' 00:40:28.328 killing process with pid 432875 00:40:28.328 17:56:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@973 -- # kill 432875 00:40:28.328 17:56:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@978 -- # wait 432875 00:40:28.587 17:56:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:40:28.587 17:56:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:40:28.587 17:56:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:40:28.587 17:56:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@297 -- # iptr 00:40:28.587 17:56:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@791 -- # iptables-save 00:40:28.587 17:56:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:40:28.587 17:56:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@791 -- # iptables-restore 00:40:28.587 17:56:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:40:28.587 17:56:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@302 -- # remove_spdk_ns 00:40:28.587 17:56:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:40:28.587 17:56:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:40:28.587 17:56:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:40:30.488 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:40:30.488 00:40:30.488 real 0m19.229s 00:40:30.488 user 0m56.733s 00:40:30.488 sys 0m7.848s 00:40:30.488 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1130 -- # xtrace_disable 00:40:30.488 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:40:30.488 ************************************ 00:40:30.488 END TEST nvmf_lvol 00:40:30.488 ************************************ 00:40:30.748 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@28 -- # run_test nvmf_lvs_grow /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh --transport=tcp --interrupt-mode 00:40:30.748 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:40:30.748 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1111 -- # xtrace_disable 00:40:30.748 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:40:30.748 ************************************ 00:40:30.748 START TEST nvmf_lvs_grow 00:40:30.748 ************************************ 00:40:30.748 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh --transport=tcp --interrupt-mode 00:40:30.748 * Looking for test storage... 00:40:30.748 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:40:30.748 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:40:30.748 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1711 -- # lcov --version 00:40:30.748 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:40:30.748 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:40:30.748 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:40:30.748 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@333 -- # local ver1 ver1_l 00:40:30.748 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@334 -- # local ver2 ver2_l 00:40:30.748 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@336 -- # IFS=.-: 00:40:30.748 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@336 -- # read -ra ver1 00:40:30.748 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@337 -- # IFS=.-: 00:40:30.748 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@337 -- # read -ra ver2 00:40:30.748 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@338 -- # local 'op=<' 00:40:30.748 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@340 -- # ver1_l=2 00:40:30.748 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@341 -- # ver2_l=1 00:40:30.748 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:40:30.748 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@344 -- # case "$op" in 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@345 -- # : 1 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@364 -- # (( v = 0 )) 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@365 -- # decimal 1 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@353 -- # local d=1 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@355 -- # echo 1 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@365 -- # ver1[v]=1 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@366 -- # decimal 2 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@353 -- # local d=2 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@355 -- # echo 2 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@366 -- # ver2[v]=2 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@368 -- # return 0 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:40:30.749 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:40:30.749 --rc genhtml_branch_coverage=1 00:40:30.749 --rc genhtml_function_coverage=1 00:40:30.749 --rc genhtml_legend=1 00:40:30.749 --rc geninfo_all_blocks=1 00:40:30.749 --rc geninfo_unexecuted_blocks=1 00:40:30.749 00:40:30.749 ' 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:40:30.749 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:40:30.749 --rc genhtml_branch_coverage=1 00:40:30.749 --rc genhtml_function_coverage=1 00:40:30.749 --rc genhtml_legend=1 00:40:30.749 --rc geninfo_all_blocks=1 00:40:30.749 --rc geninfo_unexecuted_blocks=1 00:40:30.749 00:40:30.749 ' 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:40:30.749 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:40:30.749 --rc genhtml_branch_coverage=1 00:40:30.749 --rc genhtml_function_coverage=1 00:40:30.749 --rc genhtml_legend=1 00:40:30.749 --rc geninfo_all_blocks=1 00:40:30.749 --rc geninfo_unexecuted_blocks=1 00:40:30.749 00:40:30.749 ' 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:40:30.749 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:40:30.749 --rc genhtml_branch_coverage=1 00:40:30.749 --rc genhtml_function_coverage=1 00:40:30.749 --rc genhtml_legend=1 00:40:30.749 --rc geninfo_all_blocks=1 00:40:30.749 --rc geninfo_unexecuted_blocks=1 00:40:30.749 00:40:30.749 ' 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@7 -- # uname -s 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@15 -- # shopt -s extglob 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- paths/export.sh@5 -- # export PATH 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@51 -- # : 0 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@55 -- # have_pci_nics=0 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@11 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@12 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@98 -- # nvmftestinit 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@476 -- # prepare_net_devs 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@438 -- # local -g is_hw=no 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@440 -- # remove_spdk_ns 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:40:30.749 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:40:30.750 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:40:30.750 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:40:30.750 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@309 -- # xtrace_disable 00:40:30.750 17:56:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:40:33.279 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:40:33.279 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@315 -- # pci_devs=() 00:40:33.279 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@315 -- # local -a pci_devs 00:40:33.279 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@316 -- # pci_net_devs=() 00:40:33.279 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:40:33.279 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@317 -- # pci_drivers=() 00:40:33.279 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@317 -- # local -A pci_drivers 00:40:33.279 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@319 -- # net_devs=() 00:40:33.279 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@319 -- # local -ga net_devs 00:40:33.279 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@320 -- # e810=() 00:40:33.279 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@320 -- # local -ga e810 00:40:33.279 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@321 -- # x722=() 00:40:33.279 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@321 -- # local -ga x722 00:40:33.279 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@322 -- # mlx=() 00:40:33.279 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@322 -- # local -ga mlx 00:40:33.279 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:40:33.279 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:40:33.279 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:40:33.280 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:40:33.280 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@418 -- # [[ up == up ]] 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:40:33.280 Found net devices under 0000:0a:00.0: cvl_0_0 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@418 -- # [[ up == up ]] 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:40:33.280 Found net devices under 0000:0a:00.1: cvl_0_1 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@442 -- # is_hw=yes 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:40:33.280 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:40:33.280 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.110 ms 00:40:33.280 00:40:33.280 --- 10.0.0.2 ping statistics --- 00:40:33.280 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:40:33.280 rtt min/avg/max/mdev = 0.110/0.110/0.110/0.000 ms 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:40:33.280 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:40:33.280 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.113 ms 00:40:33.280 00:40:33.280 --- 10.0.0.1 ping statistics --- 00:40:33.280 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:40:33.280 rtt min/avg/max/mdev = 0.113/0.113/0.113/0.000 ms 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@450 -- # return 0 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@99 -- # nvmfappstart -m 0x1 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:40:33.280 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@726 -- # xtrace_disable 00:40:33.281 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:40:33.281 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@509 -- # nvmfpid=436551 00:40:33.281 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x1 00:40:33.281 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@510 -- # waitforlisten 436551 00:40:33.281 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@835 -- # '[' -z 436551 ']' 00:40:33.281 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:40:33.281 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@840 -- # local max_retries=100 00:40:33.281 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:40:33.281 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:40:33.281 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@844 -- # xtrace_disable 00:40:33.281 17:56:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:40:33.281 [2024-12-06 17:56:14.859744] thread.c:3005:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:40:33.281 [2024-12-06 17:56:14.860781] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:40:33.281 [2024-12-06 17:56:14.860844] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:40:33.281 [2024-12-06 17:56:14.927772] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:40:33.281 [2024-12-06 17:56:14.969781] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:40:33.281 [2024-12-06 17:56:14.969834] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:40:33.281 [2024-12-06 17:56:14.969848] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:40:33.281 [2024-12-06 17:56:14.969860] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:40:33.281 [2024-12-06 17:56:14.969870] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:40:33.281 [2024-12-06 17:56:14.970391] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:40:33.281 [2024-12-06 17:56:15.050370] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:40:33.281 [2024-12-06 17:56:15.050646] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:40:33.281 17:56:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:40:33.281 17:56:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@868 -- # return 0 00:40:33.281 17:56:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:40:33.281 17:56:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@732 -- # xtrace_disable 00:40:33.281 17:56:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:40:33.281 17:56:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:40:33.281 17:56:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:40:33.538 [2024-12-06 17:56:15.350933] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:40:33.538 17:56:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@102 -- # run_test lvs_grow_clean lvs_grow 00:40:33.538 17:56:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:40:33.538 17:56:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1111 -- # xtrace_disable 00:40:33.538 17:56:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:40:33.797 ************************************ 00:40:33.797 START TEST lvs_grow_clean 00:40:33.797 ************************************ 00:40:33.797 17:56:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@1129 -- # lvs_grow 00:40:33.797 17:56:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@15 -- # local aio_bdev lvs lvol 00:40:33.797 17:56:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@16 -- # local data_clusters free_clusters 00:40:33.797 17:56:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@17 -- # local bdevperf_pid run_test_pid 00:40:33.797 17:56:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@18 -- # local aio_init_size_mb=200 00:40:33.797 17:56:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@19 -- # local aio_final_size_mb=400 00:40:33.797 17:56:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@20 -- # local lvol_bdev_size_mb=150 00:40:33.797 17:56:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@23 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:40:33.797 17:56:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@24 -- # truncate -s 200M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:40:33.797 17:56:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:40:34.056 17:56:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@25 -- # aio_bdev=aio_bdev 00:40:34.056 17:56:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --cluster-sz 4194304 --md-pages-per-cluster-ratio 300 aio_bdev lvs 00:40:34.314 17:56:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@28 -- # lvs=7470aff4-d5ae-4ffd-82ec-cc8b806dbb45 00:40:34.314 17:56:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 7470aff4-d5ae-4ffd-82ec-cc8b806dbb45 00:40:34.314 17:56:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@29 -- # jq -r '.[0].total_data_clusters' 00:40:34.572 17:56:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@29 -- # data_clusters=49 00:40:34.572 17:56:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@30 -- # (( data_clusters == 49 )) 00:40:34.572 17:56:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u 7470aff4-d5ae-4ffd-82ec-cc8b806dbb45 lvol 150 00:40:34.830 17:56:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@33 -- # lvol=d942617e-dbfa-4318-8bcb-d450871cffc2 00:40:34.830 17:56:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@36 -- # truncate -s 400M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:40:34.830 17:56:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_rescan aio_bdev 00:40:35.088 [2024-12-06 17:56:16.782858] bdev_aio.c:1053:bdev_aio_rescan: *NOTICE*: AIO device is resized: bdev name /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev, old block count 51200, new block count 102400 00:40:35.088 [2024-12-06 17:56:16.782946] vbdev_lvol.c: 165:vbdev_lvs_base_bdev_event_cb: *NOTICE*: Unsupported bdev event: type 1 00:40:35.088 true 00:40:35.088 17:56:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 7470aff4-d5ae-4ffd-82ec-cc8b806dbb45 00:40:35.088 17:56:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@38 -- # jq -r '.[0].total_data_clusters' 00:40:35.346 17:56:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@38 -- # (( data_clusters == 49 )) 00:40:35.346 17:56:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:40:35.605 17:56:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 d942617e-dbfa-4318-8bcb-d450871cffc2 00:40:35.864 17:56:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:40:36.123 [2024-12-06 17:56:17.867199] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:40:36.123 17:56:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:40:36.381 17:56:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@48 -- # bdevperf_pid=436984 00:40:36.381 17:56:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock -m 0x2 -o 4096 -q 128 -w randwrite -t 10 -S 1 -z 00:40:36.381 17:56:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@49 -- # trap 'killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:40:36.381 17:56:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@50 -- # waitforlisten 436984 /var/tmp/bdevperf.sock 00:40:36.381 17:56:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@835 -- # '[' -z 436984 ']' 00:40:36.381 17:56:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:40:36.381 17:56:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@840 -- # local max_retries=100 00:40:36.381 17:56:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:40:36.381 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:40:36.381 17:56:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@844 -- # xtrace_disable 00:40:36.381 17:56:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@10 -- # set +x 00:40:36.381 [2024-12-06 17:56:18.203300] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:40:36.381 [2024-12-06 17:56:18.203385] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid436984 ] 00:40:36.640 [2024-12-06 17:56:18.270147] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:40:36.640 [2024-12-06 17:56:18.315239] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:40:36.640 17:56:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:40:36.640 17:56:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@868 -- # return 0 00:40:36.640 17:56:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 00:40:37.206 Nvme0n1 00:40:37.206 17:56:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_get_bdevs -b Nvme0n1 -t 3000 00:40:37.465 [ 00:40:37.465 { 00:40:37.465 "name": "Nvme0n1", 00:40:37.465 "aliases": [ 00:40:37.465 "d942617e-dbfa-4318-8bcb-d450871cffc2" 00:40:37.465 ], 00:40:37.465 "product_name": "NVMe disk", 00:40:37.465 "block_size": 4096, 00:40:37.465 "num_blocks": 38912, 00:40:37.465 "uuid": "d942617e-dbfa-4318-8bcb-d450871cffc2", 00:40:37.465 "numa_id": 0, 00:40:37.465 "assigned_rate_limits": { 00:40:37.465 "rw_ios_per_sec": 0, 00:40:37.465 "rw_mbytes_per_sec": 0, 00:40:37.465 "r_mbytes_per_sec": 0, 00:40:37.465 "w_mbytes_per_sec": 0 00:40:37.465 }, 00:40:37.465 "claimed": false, 00:40:37.465 "zoned": false, 00:40:37.465 "supported_io_types": { 00:40:37.465 "read": true, 00:40:37.465 "write": true, 00:40:37.465 "unmap": true, 00:40:37.465 "flush": true, 00:40:37.465 "reset": true, 00:40:37.465 "nvme_admin": true, 00:40:37.465 "nvme_io": true, 00:40:37.465 "nvme_io_md": false, 00:40:37.465 "write_zeroes": true, 00:40:37.465 "zcopy": false, 00:40:37.465 "get_zone_info": false, 00:40:37.465 "zone_management": false, 00:40:37.465 "zone_append": false, 00:40:37.465 "compare": true, 00:40:37.465 "compare_and_write": true, 00:40:37.465 "abort": true, 00:40:37.465 "seek_hole": false, 00:40:37.465 "seek_data": false, 00:40:37.465 "copy": true, 00:40:37.465 "nvme_iov_md": false 00:40:37.465 }, 00:40:37.465 "memory_domains": [ 00:40:37.465 { 00:40:37.465 "dma_device_id": "system", 00:40:37.465 "dma_device_type": 1 00:40:37.465 } 00:40:37.465 ], 00:40:37.465 "driver_specific": { 00:40:37.465 "nvme": [ 00:40:37.465 { 00:40:37.465 "trid": { 00:40:37.465 "trtype": "TCP", 00:40:37.465 "adrfam": "IPv4", 00:40:37.465 "traddr": "10.0.0.2", 00:40:37.465 "trsvcid": "4420", 00:40:37.465 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:40:37.465 }, 00:40:37.465 "ctrlr_data": { 00:40:37.465 "cntlid": 1, 00:40:37.465 "vendor_id": "0x8086", 00:40:37.465 "model_number": "SPDK bdev Controller", 00:40:37.465 "serial_number": "SPDK0", 00:40:37.465 "firmware_revision": "25.01", 00:40:37.465 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:40:37.465 "oacs": { 00:40:37.465 "security": 0, 00:40:37.465 "format": 0, 00:40:37.465 "firmware": 0, 00:40:37.465 "ns_manage": 0 00:40:37.465 }, 00:40:37.465 "multi_ctrlr": true, 00:40:37.465 "ana_reporting": false 00:40:37.465 }, 00:40:37.465 "vs": { 00:40:37.465 "nvme_version": "1.3" 00:40:37.465 }, 00:40:37.465 "ns_data": { 00:40:37.465 "id": 1, 00:40:37.465 "can_share": true 00:40:37.465 } 00:40:37.465 } 00:40:37.465 ], 00:40:37.465 "mp_policy": "active_passive" 00:40:37.465 } 00:40:37.465 } 00:40:37.465 ] 00:40:37.465 17:56:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@56 -- # run_test_pid=437113 00:40:37.465 17:56:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:40:37.465 17:56:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@57 -- # sleep 2 00:40:37.465 Running I/O for 10 seconds... 00:40:38.840 Latency(us) 00:40:38.840 [2024-12-06T16:56:20.679Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:40:38.840 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:40:38.840 Nvme0n1 : 1.00 14923.00 58.29 0.00 0.00 0.00 0.00 0.00 00:40:38.840 [2024-12-06T16:56:20.679Z] =================================================================================================================== 00:40:38.840 [2024-12-06T16:56:20.679Z] Total : 14923.00 58.29 0.00 0.00 0.00 0.00 0.00 00:40:38.840 00:40:39.406 17:56:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_grow_lvstore -u 7470aff4-d5ae-4ffd-82ec-cc8b806dbb45 00:40:39.664 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:40:39.664 Nvme0n1 : 2.00 15193.50 59.35 0.00 0.00 0.00 0.00 0.00 00:40:39.664 [2024-12-06T16:56:21.503Z] =================================================================================================================== 00:40:39.664 [2024-12-06T16:56:21.503Z] Total : 15193.50 59.35 0.00 0.00 0.00 0.00 0.00 00:40:39.664 00:40:39.664 true 00:40:39.664 17:56:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 7470aff4-d5ae-4ffd-82ec-cc8b806dbb45 00:40:39.664 17:56:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@61 -- # jq -r '.[0].total_data_clusters' 00:40:39.923 17:56:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@61 -- # data_clusters=99 00:40:39.923 17:56:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@62 -- # (( data_clusters == 99 )) 00:40:39.923 17:56:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@65 -- # wait 437113 00:40:40.490 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:40:40.490 Nvme0n1 : 3.00 15272.67 59.66 0.00 0.00 0.00 0.00 0.00 00:40:40.490 [2024-12-06T16:56:22.329Z] =================================================================================================================== 00:40:40.490 [2024-12-06T16:56:22.329Z] Total : 15272.67 59.66 0.00 0.00 0.00 0.00 0.00 00:40:40.490 00:40:41.863 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:40:41.863 Nvme0n1 : 4.00 15352.25 59.97 0.00 0.00 0.00 0.00 0.00 00:40:41.863 [2024-12-06T16:56:23.702Z] =================================================================================================================== 00:40:41.863 [2024-12-06T16:56:23.702Z] Total : 15352.25 59.97 0.00 0.00 0.00 0.00 0.00 00:40:41.863 00:40:42.799 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:40:42.799 Nvme0n1 : 5.00 15434.80 60.29 0.00 0.00 0.00 0.00 0.00 00:40:42.799 [2024-12-06T16:56:24.638Z] =================================================================================================================== 00:40:42.799 [2024-12-06T16:56:24.638Z] Total : 15434.80 60.29 0.00 0.00 0.00 0.00 0.00 00:40:42.799 00:40:43.730 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:40:43.730 Nvme0n1 : 6.00 15508.17 60.58 0.00 0.00 0.00 0.00 0.00 00:40:43.730 [2024-12-06T16:56:25.569Z] =================================================================================================================== 00:40:43.730 [2024-12-06T16:56:25.569Z] Total : 15508.17 60.58 0.00 0.00 0.00 0.00 0.00 00:40:43.730 00:40:44.664 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:40:44.664 Nvme0n1 : 7.00 15551.57 60.75 0.00 0.00 0.00 0.00 0.00 00:40:44.664 [2024-12-06T16:56:26.503Z] =================================================================================================================== 00:40:44.664 [2024-12-06T16:56:26.503Z] Total : 15551.57 60.75 0.00 0.00 0.00 0.00 0.00 00:40:44.664 00:40:45.598 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:40:45.598 Nvme0n1 : 8.00 15588.25 60.89 0.00 0.00 0.00 0.00 0.00 00:40:45.598 [2024-12-06T16:56:27.437Z] =================================================================================================================== 00:40:45.598 [2024-12-06T16:56:27.437Z] Total : 15588.25 60.89 0.00 0.00 0.00 0.00 0.00 00:40:45.598 00:40:46.531 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:40:46.531 Nvme0n1 : 9.00 15620.11 61.02 0.00 0.00 0.00 0.00 0.00 00:40:46.531 [2024-12-06T16:56:28.370Z] =================================================================================================================== 00:40:46.531 [2024-12-06T16:56:28.370Z] Total : 15620.11 61.02 0.00 0.00 0.00 0.00 0.00 00:40:46.531 00:40:47.464 00:40:47.464 Latency(us) 00:40:47.464 [2024-12-06T16:56:29.303Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:40:47.464 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:40:47.464 Nvme0n1 : 10.00 15642.15 61.10 0.00 0.00 8178.53 4296.25 18738.44 00:40:47.464 [2024-12-06T16:56:29.303Z] =================================================================================================================== 00:40:47.464 [2024-12-06T16:56:29.303Z] Total : 15642.15 61.10 0.00 0.00 8178.53 4296.25 18738.44 00:40:47.464 { 00:40:47.464 "results": [ 00:40:47.464 { 00:40:47.464 "job": "Nvme0n1", 00:40:47.464 "core_mask": "0x2", 00:40:47.464 "workload": "randwrite", 00:40:47.464 "status": "finished", 00:40:47.464 "queue_depth": 128, 00:40:47.464 "io_size": 4096, 00:40:47.464 "runtime": 10.002272, 00:40:47.464 "iops": 15642.14610440508, 00:40:47.464 "mibps": 61.10213322033234, 00:40:47.464 "io_failed": 0, 00:40:47.464 "io_timeout": 0, 00:40:47.464 "avg_latency_us": 8178.531291849446, 00:40:47.464 "min_latency_us": 4296.248888888889, 00:40:47.464 "max_latency_us": 18738.44148148148 00:40:47.464 } 00:40:47.464 ], 00:40:47.464 "core_count": 1 00:40:47.464 } 00:40:47.723 17:56:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@66 -- # killprocess 436984 00:40:47.723 17:56:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@954 -- # '[' -z 436984 ']' 00:40:47.723 17:56:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@958 -- # kill -0 436984 00:40:47.723 17:56:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@959 -- # uname 00:40:47.723 17:56:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:40:47.723 17:56:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 436984 00:40:47.723 17:56:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:40:47.723 17:56:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:40:47.723 17:56:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@972 -- # echo 'killing process with pid 436984' 00:40:47.723 killing process with pid 436984 00:40:47.723 17:56:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@973 -- # kill 436984 00:40:47.723 Received shutdown signal, test time was about 10.000000 seconds 00:40:47.723 00:40:47.723 Latency(us) 00:40:47.723 [2024-12-06T16:56:29.562Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:40:47.723 [2024-12-06T16:56:29.562Z] =================================================================================================================== 00:40:47.723 [2024-12-06T16:56:29.562Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:40:47.723 17:56:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@978 -- # wait 436984 00:40:47.723 17:56:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@68 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:40:47.981 17:56:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@69 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:40:48.550 17:56:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@70 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 7470aff4-d5ae-4ffd-82ec-cc8b806dbb45 00:40:48.550 17:56:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@70 -- # jq -r '.[0].free_clusters' 00:40:48.550 17:56:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@70 -- # free_clusters=61 00:40:48.550 17:56:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@72 -- # [[ '' == \d\i\r\t\y ]] 00:40:48.550 17:56:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:40:48.809 [2024-12-06 17:56:30.610893] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev aio_bdev being removed: closing lvstore lvs 00:40:48.809 17:56:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@85 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 7470aff4-d5ae-4ffd-82ec-cc8b806dbb45 00:40:48.809 17:56:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@652 -- # local es=0 00:40:48.809 17:56:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@654 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 7470aff4-d5ae-4ffd-82ec-cc8b806dbb45 00:40:48.809 17:56:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@640 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:40:48.809 17:56:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:40:48.809 17:56:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@644 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:40:48.809 17:56:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:40:48.809 17:56:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@646 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:40:48.809 17:56:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:40:48.809 17:56:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@646 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:40:48.809 17:56:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@646 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:40:48.809 17:56:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@655 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 7470aff4-d5ae-4ffd-82ec-cc8b806dbb45 00:40:49.068 request: 00:40:49.068 { 00:40:49.068 "uuid": "7470aff4-d5ae-4ffd-82ec-cc8b806dbb45", 00:40:49.068 "method": "bdev_lvol_get_lvstores", 00:40:49.068 "req_id": 1 00:40:49.068 } 00:40:49.068 Got JSON-RPC error response 00:40:49.068 response: 00:40:49.068 { 00:40:49.068 "code": -19, 00:40:49.068 "message": "No such device" 00:40:49.068 } 00:40:49.326 17:56:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@655 -- # es=1 00:40:49.326 17:56:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:40:49.326 17:56:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:40:49.326 17:56:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:40:49.326 17:56:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:40:49.585 aio_bdev 00:40:49.585 17:56:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@87 -- # waitforbdev d942617e-dbfa-4318-8bcb-d450871cffc2 00:40:49.585 17:56:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@903 -- # local bdev_name=d942617e-dbfa-4318-8bcb-d450871cffc2 00:40:49.585 17:56:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:40:49.585 17:56:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@905 -- # local i 00:40:49.585 17:56:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:40:49.585 17:56:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:40:49.585 17:56:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@908 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:40:49.844 17:56:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@910 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b d942617e-dbfa-4318-8bcb-d450871cffc2 -t 2000 00:40:50.102 [ 00:40:50.102 { 00:40:50.102 "name": "d942617e-dbfa-4318-8bcb-d450871cffc2", 00:40:50.102 "aliases": [ 00:40:50.102 "lvs/lvol" 00:40:50.102 ], 00:40:50.102 "product_name": "Logical Volume", 00:40:50.102 "block_size": 4096, 00:40:50.102 "num_blocks": 38912, 00:40:50.102 "uuid": "d942617e-dbfa-4318-8bcb-d450871cffc2", 00:40:50.102 "assigned_rate_limits": { 00:40:50.102 "rw_ios_per_sec": 0, 00:40:50.102 "rw_mbytes_per_sec": 0, 00:40:50.102 "r_mbytes_per_sec": 0, 00:40:50.102 "w_mbytes_per_sec": 0 00:40:50.102 }, 00:40:50.102 "claimed": false, 00:40:50.102 "zoned": false, 00:40:50.102 "supported_io_types": { 00:40:50.102 "read": true, 00:40:50.102 "write": true, 00:40:50.102 "unmap": true, 00:40:50.102 "flush": false, 00:40:50.102 "reset": true, 00:40:50.102 "nvme_admin": false, 00:40:50.102 "nvme_io": false, 00:40:50.102 "nvme_io_md": false, 00:40:50.102 "write_zeroes": true, 00:40:50.102 "zcopy": false, 00:40:50.102 "get_zone_info": false, 00:40:50.102 "zone_management": false, 00:40:50.102 "zone_append": false, 00:40:50.102 "compare": false, 00:40:50.102 "compare_and_write": false, 00:40:50.102 "abort": false, 00:40:50.102 "seek_hole": true, 00:40:50.102 "seek_data": true, 00:40:50.102 "copy": false, 00:40:50.102 "nvme_iov_md": false 00:40:50.102 }, 00:40:50.102 "driver_specific": { 00:40:50.102 "lvol": { 00:40:50.102 "lvol_store_uuid": "7470aff4-d5ae-4ffd-82ec-cc8b806dbb45", 00:40:50.102 "base_bdev": "aio_bdev", 00:40:50.102 "thin_provision": false, 00:40:50.102 "num_allocated_clusters": 38, 00:40:50.102 "snapshot": false, 00:40:50.102 "clone": false, 00:40:50.102 "esnap_clone": false 00:40:50.102 } 00:40:50.102 } 00:40:50.102 } 00:40:50.102 ] 00:40:50.102 17:56:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@911 -- # return 0 00:40:50.102 17:56:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 7470aff4-d5ae-4ffd-82ec-cc8b806dbb45 00:40:50.102 17:56:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@88 -- # jq -r '.[0].free_clusters' 00:40:50.360 17:56:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@88 -- # (( free_clusters == 61 )) 00:40:50.360 17:56:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 7470aff4-d5ae-4ffd-82ec-cc8b806dbb45 00:40:50.360 17:56:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@89 -- # jq -r '.[0].total_data_clusters' 00:40:50.618 17:56:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@89 -- # (( data_clusters == 99 )) 00:40:50.618 17:56:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@92 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete d942617e-dbfa-4318-8bcb-d450871cffc2 00:40:50.877 17:56:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@93 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u 7470aff4-d5ae-4ffd-82ec-cc8b806dbb45 00:40:51.136 17:56:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@94 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:40:51.395 17:56:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@95 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:40:51.395 00:40:51.395 real 0m17.729s 00:40:51.395 user 0m17.341s 00:40:51.395 sys 0m1.775s 00:40:51.395 17:56:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@1130 -- # xtrace_disable 00:40:51.395 17:56:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@10 -- # set +x 00:40:51.395 ************************************ 00:40:51.395 END TEST lvs_grow_clean 00:40:51.395 ************************************ 00:40:51.395 17:56:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@103 -- # run_test lvs_grow_dirty lvs_grow dirty 00:40:51.395 17:56:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:40:51.395 17:56:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1111 -- # xtrace_disable 00:40:51.395 17:56:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:40:51.395 ************************************ 00:40:51.395 START TEST lvs_grow_dirty 00:40:51.395 ************************************ 00:40:51.395 17:56:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@1129 -- # lvs_grow dirty 00:40:51.395 17:56:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@15 -- # local aio_bdev lvs lvol 00:40:51.395 17:56:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@16 -- # local data_clusters free_clusters 00:40:51.395 17:56:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@17 -- # local bdevperf_pid run_test_pid 00:40:51.395 17:56:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@18 -- # local aio_init_size_mb=200 00:40:51.395 17:56:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@19 -- # local aio_final_size_mb=400 00:40:51.395 17:56:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@20 -- # local lvol_bdev_size_mb=150 00:40:51.395 17:56:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@23 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:40:51.395 17:56:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@24 -- # truncate -s 200M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:40:51.395 17:56:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:40:51.961 17:56:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@25 -- # aio_bdev=aio_bdev 00:40:51.961 17:56:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --cluster-sz 4194304 --md-pages-per-cluster-ratio 300 aio_bdev lvs 00:40:52.220 17:56:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@28 -- # lvs=867482d1-d25d-4e44-93d5-518b5790f9c8 00:40:52.220 17:56:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 867482d1-d25d-4e44-93d5-518b5790f9c8 00:40:52.220 17:56:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@29 -- # jq -r '.[0].total_data_clusters' 00:40:52.479 17:56:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@29 -- # data_clusters=49 00:40:52.479 17:56:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@30 -- # (( data_clusters == 49 )) 00:40:52.479 17:56:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u 867482d1-d25d-4e44-93d5-518b5790f9c8 lvol 150 00:40:52.737 17:56:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@33 -- # lvol=cf239f6e-c23e-4b1e-9134-87a332537b1d 00:40:52.737 17:56:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@36 -- # truncate -s 400M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:40:52.737 17:56:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_rescan aio_bdev 00:40:52.995 [2024-12-06 17:56:34.626850] bdev_aio.c:1053:bdev_aio_rescan: *NOTICE*: AIO device is resized: bdev name /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev, old block count 51200, new block count 102400 00:40:52.995 [2024-12-06 17:56:34.626934] vbdev_lvol.c: 165:vbdev_lvs_base_bdev_event_cb: *NOTICE*: Unsupported bdev event: type 1 00:40:52.995 true 00:40:52.995 17:56:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 867482d1-d25d-4e44-93d5-518b5790f9c8 00:40:52.995 17:56:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@38 -- # jq -r '.[0].total_data_clusters' 00:40:53.253 17:56:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@38 -- # (( data_clusters == 49 )) 00:40:53.253 17:56:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:40:53.511 17:56:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 cf239f6e-c23e-4b1e-9134-87a332537b1d 00:40:53.768 17:56:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:40:54.025 [2024-12-06 17:56:35.723129] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:40:54.025 17:56:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:40:54.284 17:56:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@48 -- # bdevperf_pid=439033 00:40:54.284 17:56:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock -m 0x2 -o 4096 -q 128 -w randwrite -t 10 -S 1 -z 00:40:54.284 17:56:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@49 -- # trap 'killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:40:54.284 17:56:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@50 -- # waitforlisten 439033 /var/tmp/bdevperf.sock 00:40:54.284 17:56:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@835 -- # '[' -z 439033 ']' 00:40:54.284 17:56:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:40:54.284 17:56:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@840 -- # local max_retries=100 00:40:54.284 17:56:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:40:54.284 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:40:54.284 17:56:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@844 -- # xtrace_disable 00:40:54.284 17:56:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:40:54.284 [2024-12-06 17:56:36.051458] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:40:54.284 [2024-12-06 17:56:36.051533] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid439033 ] 00:40:54.284 [2024-12-06 17:56:36.118095] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:40:54.542 [2024-12-06 17:56:36.166478] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:40:54.542 17:56:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:40:54.542 17:56:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@868 -- # return 0 00:40:54.542 17:56:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 00:40:55.108 Nvme0n1 00:40:55.108 17:56:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_get_bdevs -b Nvme0n1 -t 3000 00:40:55.367 [ 00:40:55.367 { 00:40:55.367 "name": "Nvme0n1", 00:40:55.367 "aliases": [ 00:40:55.367 "cf239f6e-c23e-4b1e-9134-87a332537b1d" 00:40:55.367 ], 00:40:55.367 "product_name": "NVMe disk", 00:40:55.367 "block_size": 4096, 00:40:55.367 "num_blocks": 38912, 00:40:55.367 "uuid": "cf239f6e-c23e-4b1e-9134-87a332537b1d", 00:40:55.367 "numa_id": 0, 00:40:55.367 "assigned_rate_limits": { 00:40:55.367 "rw_ios_per_sec": 0, 00:40:55.367 "rw_mbytes_per_sec": 0, 00:40:55.367 "r_mbytes_per_sec": 0, 00:40:55.367 "w_mbytes_per_sec": 0 00:40:55.367 }, 00:40:55.367 "claimed": false, 00:40:55.367 "zoned": false, 00:40:55.367 "supported_io_types": { 00:40:55.367 "read": true, 00:40:55.367 "write": true, 00:40:55.367 "unmap": true, 00:40:55.367 "flush": true, 00:40:55.367 "reset": true, 00:40:55.367 "nvme_admin": true, 00:40:55.367 "nvme_io": true, 00:40:55.367 "nvme_io_md": false, 00:40:55.367 "write_zeroes": true, 00:40:55.367 "zcopy": false, 00:40:55.367 "get_zone_info": false, 00:40:55.367 "zone_management": false, 00:40:55.367 "zone_append": false, 00:40:55.367 "compare": true, 00:40:55.367 "compare_and_write": true, 00:40:55.367 "abort": true, 00:40:55.367 "seek_hole": false, 00:40:55.367 "seek_data": false, 00:40:55.367 "copy": true, 00:40:55.367 "nvme_iov_md": false 00:40:55.367 }, 00:40:55.367 "memory_domains": [ 00:40:55.367 { 00:40:55.367 "dma_device_id": "system", 00:40:55.367 "dma_device_type": 1 00:40:55.367 } 00:40:55.367 ], 00:40:55.367 "driver_specific": { 00:40:55.367 "nvme": [ 00:40:55.367 { 00:40:55.367 "trid": { 00:40:55.367 "trtype": "TCP", 00:40:55.367 "adrfam": "IPv4", 00:40:55.367 "traddr": "10.0.0.2", 00:40:55.367 "trsvcid": "4420", 00:40:55.367 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:40:55.367 }, 00:40:55.367 "ctrlr_data": { 00:40:55.367 "cntlid": 1, 00:40:55.367 "vendor_id": "0x8086", 00:40:55.367 "model_number": "SPDK bdev Controller", 00:40:55.367 "serial_number": "SPDK0", 00:40:55.367 "firmware_revision": "25.01", 00:40:55.367 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:40:55.367 "oacs": { 00:40:55.367 "security": 0, 00:40:55.367 "format": 0, 00:40:55.367 "firmware": 0, 00:40:55.367 "ns_manage": 0 00:40:55.367 }, 00:40:55.367 "multi_ctrlr": true, 00:40:55.367 "ana_reporting": false 00:40:55.367 }, 00:40:55.367 "vs": { 00:40:55.367 "nvme_version": "1.3" 00:40:55.367 }, 00:40:55.367 "ns_data": { 00:40:55.367 "id": 1, 00:40:55.367 "can_share": true 00:40:55.367 } 00:40:55.367 } 00:40:55.367 ], 00:40:55.367 "mp_policy": "active_passive" 00:40:55.367 } 00:40:55.367 } 00:40:55.367 ] 00:40:55.367 17:56:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@56 -- # run_test_pid=439168 00:40:55.367 17:56:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:40:55.367 17:56:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@57 -- # sleep 2 00:40:55.367 Running I/O for 10 seconds... 00:40:56.743 Latency(us) 00:40:56.743 [2024-12-06T16:56:38.582Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:40:56.743 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:40:56.743 Nvme0n1 : 1.00 15113.00 59.04 0.00 0.00 0.00 0.00 0.00 00:40:56.743 [2024-12-06T16:56:38.582Z] =================================================================================================================== 00:40:56.743 [2024-12-06T16:56:38.582Z] Total : 15113.00 59.04 0.00 0.00 0.00 0.00 0.00 00:40:56.743 00:40:57.309 17:56:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_grow_lvstore -u 867482d1-d25d-4e44-93d5-518b5790f9c8 00:40:57.584 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:40:57.584 Nvme0n1 : 2.00 15240.00 59.53 0.00 0.00 0.00 0.00 0.00 00:40:57.584 [2024-12-06T16:56:39.423Z] =================================================================================================================== 00:40:57.584 [2024-12-06T16:56:39.423Z] Total : 15240.00 59.53 0.00 0.00 0.00 0.00 0.00 00:40:57.584 00:40:57.584 true 00:40:57.584 17:56:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 867482d1-d25d-4e44-93d5-518b5790f9c8 00:40:57.585 17:56:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@61 -- # jq -r '.[0].total_data_clusters' 00:40:57.927 17:56:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@61 -- # data_clusters=99 00:40:57.927 17:56:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@62 -- # (( data_clusters == 99 )) 00:40:57.927 17:56:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@65 -- # wait 439168 00:40:58.495 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:40:58.495 Nvme0n1 : 3.00 15324.67 59.86 0.00 0.00 0.00 0.00 0.00 00:40:58.495 [2024-12-06T16:56:40.334Z] =================================================================================================================== 00:40:58.495 [2024-12-06T16:56:40.334Z] Total : 15324.67 59.86 0.00 0.00 0.00 0.00 0.00 00:40:58.495 00:40:59.430 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:40:59.430 Nvme0n1 : 4.00 15367.00 60.03 0.00 0.00 0.00 0.00 0.00 00:40:59.430 [2024-12-06T16:56:41.269Z] =================================================================================================================== 00:40:59.430 [2024-12-06T16:56:41.269Z] Total : 15367.00 60.03 0.00 0.00 0.00 0.00 0.00 00:40:59.430 00:41:00.362 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:41:00.362 Nvme0n1 : 5.00 15468.60 60.42 0.00 0.00 0.00 0.00 0.00 00:41:00.362 [2024-12-06T16:56:42.201Z] =================================================================================================================== 00:41:00.362 [2024-12-06T16:56:42.201Z] Total : 15468.60 60.42 0.00 0.00 0.00 0.00 0.00 00:41:00.362 00:41:01.739 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:41:01.739 Nvme0n1 : 6.00 15557.50 60.77 0.00 0.00 0.00 0.00 0.00 00:41:01.739 [2024-12-06T16:56:43.578Z] =================================================================================================================== 00:41:01.739 [2024-12-06T16:56:43.578Z] Total : 15557.50 60.77 0.00 0.00 0.00 0.00 0.00 00:41:01.739 00:41:02.674 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:41:02.674 Nvme0n1 : 7.00 15621.00 61.02 0.00 0.00 0.00 0.00 0.00 00:41:02.674 [2024-12-06T16:56:44.513Z] =================================================================================================================== 00:41:02.674 [2024-12-06T16:56:44.513Z] Total : 15621.00 61.02 0.00 0.00 0.00 0.00 0.00 00:41:02.674 00:41:03.607 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:41:03.607 Nvme0n1 : 8.00 15668.62 61.21 0.00 0.00 0.00 0.00 0.00 00:41:03.607 [2024-12-06T16:56:45.446Z] =================================================================================================================== 00:41:03.607 [2024-12-06T16:56:45.447Z] Total : 15668.62 61.21 0.00 0.00 0.00 0.00 0.00 00:41:03.608 00:41:04.540 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:41:04.540 Nvme0n1 : 9.00 15705.67 61.35 0.00 0.00 0.00 0.00 0.00 00:41:04.540 [2024-12-06T16:56:46.379Z] =================================================================================================================== 00:41:04.540 [2024-12-06T16:56:46.379Z] Total : 15705.67 61.35 0.00 0.00 0.00 0.00 0.00 00:41:04.540 00:41:05.471 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:41:05.471 Nvme0n1 : 10.00 15735.30 61.47 0.00 0.00 0.00 0.00 0.00 00:41:05.471 [2024-12-06T16:56:47.310Z] =================================================================================================================== 00:41:05.471 [2024-12-06T16:56:47.310Z] Total : 15735.30 61.47 0.00 0.00 0.00 0.00 0.00 00:41:05.471 00:41:05.471 00:41:05.471 Latency(us) 00:41:05.471 [2024-12-06T16:56:47.310Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:41:05.471 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:41:05.471 Nvme0n1 : 10.00 15743.66 61.50 0.00 0.00 8125.71 6990.51 17961.72 00:41:05.471 [2024-12-06T16:56:47.310Z] =================================================================================================================== 00:41:05.471 [2024-12-06T16:56:47.310Z] Total : 15743.66 61.50 0.00 0.00 8125.71 6990.51 17961.72 00:41:05.471 { 00:41:05.471 "results": [ 00:41:05.471 { 00:41:05.471 "job": "Nvme0n1", 00:41:05.471 "core_mask": "0x2", 00:41:05.471 "workload": "randwrite", 00:41:05.471 "status": "finished", 00:41:05.471 "queue_depth": 128, 00:41:05.471 "io_size": 4096, 00:41:05.471 "runtime": 10.002823, 00:41:05.471 "iops": 15743.655566033709, 00:41:05.471 "mibps": 61.498654554819176, 00:41:05.471 "io_failed": 0, 00:41:05.471 "io_timeout": 0, 00:41:05.471 "avg_latency_us": 8125.709254783706, 00:41:05.471 "min_latency_us": 6990.506666666667, 00:41:05.471 "max_latency_us": 17961.71851851852 00:41:05.471 } 00:41:05.471 ], 00:41:05.471 "core_count": 1 00:41:05.471 } 00:41:05.471 17:56:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@66 -- # killprocess 439033 00:41:05.471 17:56:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@954 -- # '[' -z 439033 ']' 00:41:05.471 17:56:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@958 -- # kill -0 439033 00:41:05.471 17:56:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@959 -- # uname 00:41:05.471 17:56:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:41:05.471 17:56:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 439033 00:41:05.471 17:56:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:41:05.471 17:56:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:41:05.471 17:56:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@972 -- # echo 'killing process with pid 439033' 00:41:05.471 killing process with pid 439033 00:41:05.471 17:56:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@973 -- # kill 439033 00:41:05.471 Received shutdown signal, test time was about 10.000000 seconds 00:41:05.471 00:41:05.471 Latency(us) 00:41:05.471 [2024-12-06T16:56:47.310Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:41:05.471 [2024-12-06T16:56:47.310Z] =================================================================================================================== 00:41:05.471 [2024-12-06T16:56:47.311Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:41:05.472 17:56:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@978 -- # wait 439033 00:41:05.729 17:56:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@68 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:41:05.987 17:56:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@69 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:41:06.244 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@70 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 867482d1-d25d-4e44-93d5-518b5790f9c8 00:41:06.244 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@70 -- # jq -r '.[0].free_clusters' 00:41:06.501 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@70 -- # free_clusters=61 00:41:06.501 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@72 -- # [[ dirty == \d\i\r\t\y ]] 00:41:06.501 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@74 -- # kill -9 436551 00:41:06.501 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@75 -- # wait 436551 00:41:06.759 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh: line 75: 436551 Killed "${NVMF_APP[@]}" "$@" 00:41:06.759 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@75 -- # true 00:41:06.759 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@76 -- # nvmfappstart -m 0x1 00:41:06.759 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:41:06.759 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@726 -- # xtrace_disable 00:41:06.759 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:41:06.759 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@509 -- # nvmfpid=440488 00:41:06.759 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x1 00:41:06.759 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@510 -- # waitforlisten 440488 00:41:06.759 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@835 -- # '[' -z 440488 ']' 00:41:06.759 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:41:06.759 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@840 -- # local max_retries=100 00:41:06.759 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:41:06.759 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:41:06.759 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@844 -- # xtrace_disable 00:41:06.759 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:41:06.759 [2024-12-06 17:56:48.430385] thread.c:3005:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:41:06.759 [2024-12-06 17:56:48.431429] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:41:06.759 [2024-12-06 17:56:48.431484] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:41:06.759 [2024-12-06 17:56:48.503717] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:41:06.759 [2024-12-06 17:56:48.544942] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:41:06.759 [2024-12-06 17:56:48.545015] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:41:06.759 [2024-12-06 17:56:48.545036] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:41:06.759 [2024-12-06 17:56:48.545046] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:41:06.759 [2024-12-06 17:56:48.545055] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:41:06.759 [2024-12-06 17:56:48.545573] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:41:07.017 [2024-12-06 17:56:48.628053] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:41:07.017 [2024-12-06 17:56:48.628317] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:41:07.017 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:41:07.017 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@868 -- # return 0 00:41:07.017 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:41:07.017 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@732 -- # xtrace_disable 00:41:07.017 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:41:07.017 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:41:07.017 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@77 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:41:07.274 [2024-12-06 17:56:48.932459] blobstore.c:4899:bs_recover: *NOTICE*: Performing recovery on blobstore 00:41:07.274 [2024-12-06 17:56:48.932634] blobstore.c:4846:bs_load_replay_md_cpl: *NOTICE*: Recover: blob 0x0 00:41:07.274 [2024-12-06 17:56:48.932726] blobstore.c:4846:bs_load_replay_md_cpl: *NOTICE*: Recover: blob 0x1 00:41:07.274 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@77 -- # aio_bdev=aio_bdev 00:41:07.274 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@78 -- # waitforbdev cf239f6e-c23e-4b1e-9134-87a332537b1d 00:41:07.274 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@903 -- # local bdev_name=cf239f6e-c23e-4b1e-9134-87a332537b1d 00:41:07.274 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:41:07.274 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@905 -- # local i 00:41:07.274 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:41:07.274 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:41:07.274 17:56:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@908 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:41:07.532 17:56:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@910 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b cf239f6e-c23e-4b1e-9134-87a332537b1d -t 2000 00:41:07.789 [ 00:41:07.789 { 00:41:07.789 "name": "cf239f6e-c23e-4b1e-9134-87a332537b1d", 00:41:07.789 "aliases": [ 00:41:07.789 "lvs/lvol" 00:41:07.789 ], 00:41:07.789 "product_name": "Logical Volume", 00:41:07.789 "block_size": 4096, 00:41:07.789 "num_blocks": 38912, 00:41:07.789 "uuid": "cf239f6e-c23e-4b1e-9134-87a332537b1d", 00:41:07.789 "assigned_rate_limits": { 00:41:07.789 "rw_ios_per_sec": 0, 00:41:07.789 "rw_mbytes_per_sec": 0, 00:41:07.789 "r_mbytes_per_sec": 0, 00:41:07.789 "w_mbytes_per_sec": 0 00:41:07.789 }, 00:41:07.789 "claimed": false, 00:41:07.789 "zoned": false, 00:41:07.789 "supported_io_types": { 00:41:07.789 "read": true, 00:41:07.789 "write": true, 00:41:07.789 "unmap": true, 00:41:07.789 "flush": false, 00:41:07.789 "reset": true, 00:41:07.789 "nvme_admin": false, 00:41:07.789 "nvme_io": false, 00:41:07.789 "nvme_io_md": false, 00:41:07.789 "write_zeroes": true, 00:41:07.789 "zcopy": false, 00:41:07.789 "get_zone_info": false, 00:41:07.789 "zone_management": false, 00:41:07.789 "zone_append": false, 00:41:07.789 "compare": false, 00:41:07.789 "compare_and_write": false, 00:41:07.789 "abort": false, 00:41:07.789 "seek_hole": true, 00:41:07.789 "seek_data": true, 00:41:07.789 "copy": false, 00:41:07.789 "nvme_iov_md": false 00:41:07.789 }, 00:41:07.789 "driver_specific": { 00:41:07.789 "lvol": { 00:41:07.789 "lvol_store_uuid": "867482d1-d25d-4e44-93d5-518b5790f9c8", 00:41:07.789 "base_bdev": "aio_bdev", 00:41:07.789 "thin_provision": false, 00:41:07.789 "num_allocated_clusters": 38, 00:41:07.789 "snapshot": false, 00:41:07.789 "clone": false, 00:41:07.789 "esnap_clone": false 00:41:07.789 } 00:41:07.789 } 00:41:07.789 } 00:41:07.789 ] 00:41:07.789 17:56:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@911 -- # return 0 00:41:07.789 17:56:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 867482d1-d25d-4e44-93d5-518b5790f9c8 00:41:07.789 17:56:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@79 -- # jq -r '.[0].free_clusters' 00:41:08.046 17:56:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@79 -- # (( free_clusters == 61 )) 00:41:08.046 17:56:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@80 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 867482d1-d25d-4e44-93d5-518b5790f9c8 00:41:08.046 17:56:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@80 -- # jq -r '.[0].total_data_clusters' 00:41:08.304 17:56:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@80 -- # (( data_clusters == 99 )) 00:41:08.304 17:56:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:41:08.562 [2024-12-06 17:56:50.298140] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev aio_bdev being removed: closing lvstore lvs 00:41:08.562 17:56:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@85 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 867482d1-d25d-4e44-93d5-518b5790f9c8 00:41:08.562 17:56:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@652 -- # local es=0 00:41:08.562 17:56:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@654 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 867482d1-d25d-4e44-93d5-518b5790f9c8 00:41:08.562 17:56:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@640 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:41:08.562 17:56:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:41:08.562 17:56:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@644 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:41:08.562 17:56:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:41:08.562 17:56:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@646 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:41:08.562 17:56:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:41:08.562 17:56:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@646 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:41:08.562 17:56:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@646 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:41:08.562 17:56:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@655 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 867482d1-d25d-4e44-93d5-518b5790f9c8 00:41:08.821 request: 00:41:08.821 { 00:41:08.821 "uuid": "867482d1-d25d-4e44-93d5-518b5790f9c8", 00:41:08.821 "method": "bdev_lvol_get_lvstores", 00:41:08.821 "req_id": 1 00:41:08.821 } 00:41:08.821 Got JSON-RPC error response 00:41:08.821 response: 00:41:08.821 { 00:41:08.821 "code": -19, 00:41:08.821 "message": "No such device" 00:41:08.821 } 00:41:08.821 17:56:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@655 -- # es=1 00:41:08.821 17:56:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:41:08.821 17:56:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:41:08.821 17:56:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:41:08.821 17:56:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:41:09.087 aio_bdev 00:41:09.087 17:56:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@87 -- # waitforbdev cf239f6e-c23e-4b1e-9134-87a332537b1d 00:41:09.087 17:56:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@903 -- # local bdev_name=cf239f6e-c23e-4b1e-9134-87a332537b1d 00:41:09.087 17:56:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@904 -- # local bdev_timeout= 00:41:09.087 17:56:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@905 -- # local i 00:41:09.087 17:56:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@906 -- # [[ -z '' ]] 00:41:09.087 17:56:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@906 -- # bdev_timeout=2000 00:41:09.087 17:56:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@908 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:41:09.346 17:56:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@910 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b cf239f6e-c23e-4b1e-9134-87a332537b1d -t 2000 00:41:09.605 [ 00:41:09.605 { 00:41:09.605 "name": "cf239f6e-c23e-4b1e-9134-87a332537b1d", 00:41:09.605 "aliases": [ 00:41:09.605 "lvs/lvol" 00:41:09.605 ], 00:41:09.605 "product_name": "Logical Volume", 00:41:09.605 "block_size": 4096, 00:41:09.605 "num_blocks": 38912, 00:41:09.605 "uuid": "cf239f6e-c23e-4b1e-9134-87a332537b1d", 00:41:09.605 "assigned_rate_limits": { 00:41:09.605 "rw_ios_per_sec": 0, 00:41:09.605 "rw_mbytes_per_sec": 0, 00:41:09.605 "r_mbytes_per_sec": 0, 00:41:09.605 "w_mbytes_per_sec": 0 00:41:09.605 }, 00:41:09.605 "claimed": false, 00:41:09.605 "zoned": false, 00:41:09.605 "supported_io_types": { 00:41:09.605 "read": true, 00:41:09.605 "write": true, 00:41:09.605 "unmap": true, 00:41:09.605 "flush": false, 00:41:09.605 "reset": true, 00:41:09.605 "nvme_admin": false, 00:41:09.605 "nvme_io": false, 00:41:09.605 "nvme_io_md": false, 00:41:09.605 "write_zeroes": true, 00:41:09.605 "zcopy": false, 00:41:09.605 "get_zone_info": false, 00:41:09.605 "zone_management": false, 00:41:09.605 "zone_append": false, 00:41:09.605 "compare": false, 00:41:09.605 "compare_and_write": false, 00:41:09.605 "abort": false, 00:41:09.605 "seek_hole": true, 00:41:09.605 "seek_data": true, 00:41:09.605 "copy": false, 00:41:09.605 "nvme_iov_md": false 00:41:09.605 }, 00:41:09.605 "driver_specific": { 00:41:09.605 "lvol": { 00:41:09.605 "lvol_store_uuid": "867482d1-d25d-4e44-93d5-518b5790f9c8", 00:41:09.605 "base_bdev": "aio_bdev", 00:41:09.605 "thin_provision": false, 00:41:09.605 "num_allocated_clusters": 38, 00:41:09.605 "snapshot": false, 00:41:09.605 "clone": false, 00:41:09.605 "esnap_clone": false 00:41:09.605 } 00:41:09.605 } 00:41:09.605 } 00:41:09.605 ] 00:41:09.864 17:56:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@911 -- # return 0 00:41:09.864 17:56:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 867482d1-d25d-4e44-93d5-518b5790f9c8 00:41:09.864 17:56:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@88 -- # jq -r '.[0].free_clusters' 00:41:10.122 17:56:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@88 -- # (( free_clusters == 61 )) 00:41:10.122 17:56:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 867482d1-d25d-4e44-93d5-518b5790f9c8 00:41:10.122 17:56:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@89 -- # jq -r '.[0].total_data_clusters' 00:41:10.381 17:56:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@89 -- # (( data_clusters == 99 )) 00:41:10.381 17:56:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@92 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete cf239f6e-c23e-4b1e-9134-87a332537b1d 00:41:10.639 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@93 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u 867482d1-d25d-4e44-93d5-518b5790f9c8 00:41:10.897 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@94 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@95 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:41:11.155 00:41:11.155 real 0m19.693s 00:41:11.155 user 0m36.593s 00:41:11.155 sys 0m4.725s 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@1130 -- # xtrace_disable 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:41:11.155 ************************************ 00:41:11.155 END TEST lvs_grow_dirty 00:41:11.155 ************************************ 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@1 -- # process_shm --id 0 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@812 -- # type=--id 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@813 -- # id=0 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@814 -- # '[' --id = --pid ']' 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@818 -- # find /dev/shm -name '*.0' -printf '%f\n' 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@818 -- # shm_files=nvmf_trace.0 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@820 -- # [[ -z nvmf_trace.0 ]] 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@824 -- # for n in $shm_files 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@825 -- # tar -C /dev/shm/ -cvzf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf_trace.0_shm.tar.gz nvmf_trace.0 00:41:11.155 nvmf_trace.0 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@827 -- # return 0 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@1 -- # nvmftestfini 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@516 -- # nvmfcleanup 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@121 -- # sync 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@124 -- # set +e 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@125 -- # for i in {1..20} 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:41:11.155 rmmod nvme_tcp 00:41:11.155 rmmod nvme_fabrics 00:41:11.155 rmmod nvme_keyring 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@128 -- # set -e 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@129 -- # return 0 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@517 -- # '[' -n 440488 ']' 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@518 -- # killprocess 440488 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@954 -- # '[' -z 440488 ']' 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@958 -- # kill -0 440488 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@959 -- # uname 00:41:11.155 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:41:11.414 17:56:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 440488 00:41:11.414 17:56:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:41:11.414 17:56:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:41:11.414 17:56:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@972 -- # echo 'killing process with pid 440488' 00:41:11.414 killing process with pid 440488 00:41:11.414 17:56:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@973 -- # kill 440488 00:41:11.414 17:56:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@978 -- # wait 440488 00:41:11.414 17:56:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:41:11.414 17:56:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:41:11.414 17:56:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:41:11.414 17:56:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@297 -- # iptr 00:41:11.414 17:56:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@791 -- # iptables-save 00:41:11.414 17:56:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:41:11.414 17:56:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@791 -- # iptables-restore 00:41:11.414 17:56:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:41:11.414 17:56:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@302 -- # remove_spdk_ns 00:41:11.414 17:56:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:41:11.414 17:56:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:41:11.414 17:56:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:41:13.950 00:41:13.950 real 0m42.901s 00:41:13.950 user 0m55.709s 00:41:13.950 sys 0m8.495s 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1130 -- # xtrace_disable 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:41:13.950 ************************************ 00:41:13.950 END TEST nvmf_lvs_grow 00:41:13.950 ************************************ 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@29 -- # run_test nvmf_bdev_io_wait /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdev_io_wait.sh --transport=tcp --interrupt-mode 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1111 -- # xtrace_disable 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:41:13.950 ************************************ 00:41:13.950 START TEST nvmf_bdev_io_wait 00:41:13.950 ************************************ 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdev_io_wait.sh --transport=tcp --interrupt-mode 00:41:13.950 * Looking for test storage... 00:41:13.950 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1711 -- # lcov --version 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@333 -- # local ver1 ver1_l 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@334 -- # local ver2 ver2_l 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@336 -- # IFS=.-: 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@336 -- # read -ra ver1 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@337 -- # IFS=.-: 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@337 -- # read -ra ver2 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@338 -- # local 'op=<' 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@340 -- # ver1_l=2 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@341 -- # ver2_l=1 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@344 -- # case "$op" in 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@345 -- # : 1 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@364 -- # (( v = 0 )) 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@365 -- # decimal 1 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@353 -- # local d=1 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@355 -- # echo 1 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@365 -- # ver1[v]=1 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@366 -- # decimal 2 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@353 -- # local d=2 00:41:13.950 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@355 -- # echo 2 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@366 -- # ver2[v]=2 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@368 -- # return 0 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:41:13.951 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:41:13.951 --rc genhtml_branch_coverage=1 00:41:13.951 --rc genhtml_function_coverage=1 00:41:13.951 --rc genhtml_legend=1 00:41:13.951 --rc geninfo_all_blocks=1 00:41:13.951 --rc geninfo_unexecuted_blocks=1 00:41:13.951 00:41:13.951 ' 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:41:13.951 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:41:13.951 --rc genhtml_branch_coverage=1 00:41:13.951 --rc genhtml_function_coverage=1 00:41:13.951 --rc genhtml_legend=1 00:41:13.951 --rc geninfo_all_blocks=1 00:41:13.951 --rc geninfo_unexecuted_blocks=1 00:41:13.951 00:41:13.951 ' 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:41:13.951 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:41:13.951 --rc genhtml_branch_coverage=1 00:41:13.951 --rc genhtml_function_coverage=1 00:41:13.951 --rc genhtml_legend=1 00:41:13.951 --rc geninfo_all_blocks=1 00:41:13.951 --rc geninfo_unexecuted_blocks=1 00:41:13.951 00:41:13.951 ' 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:41:13.951 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:41:13.951 --rc genhtml_branch_coverage=1 00:41:13.951 --rc genhtml_function_coverage=1 00:41:13.951 --rc genhtml_legend=1 00:41:13.951 --rc geninfo_all_blocks=1 00:41:13.951 --rc geninfo_unexecuted_blocks=1 00:41:13.951 00:41:13.951 ' 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@7 -- # uname -s 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@15 -- # shopt -s extglob 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- paths/export.sh@5 -- # export PATH 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@51 -- # : 0 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@55 -- # have_pci_nics=0 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@11 -- # MALLOC_BDEV_SIZE=64 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@14 -- # nvmftestinit 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@476 -- # prepare_net_devs 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@438 -- # local -g is_hw=no 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@440 -- # remove_spdk_ns 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@309 -- # xtrace_disable 00:41:13.951 17:56:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@315 -- # pci_devs=() 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@315 -- # local -a pci_devs 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@316 -- # pci_net_devs=() 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@317 -- # pci_drivers=() 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@317 -- # local -A pci_drivers 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@319 -- # net_devs=() 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@319 -- # local -ga net_devs 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@320 -- # e810=() 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@320 -- # local -ga e810 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@321 -- # x722=() 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@321 -- # local -ga x722 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@322 -- # mlx=() 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@322 -- # local -ga mlx 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:41:15.864 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:41:15.864 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:41:15.865 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@418 -- # [[ up == up ]] 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:41:15.865 Found net devices under 0000:0a:00.0: cvl_0_0 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@418 -- # [[ up == up ]] 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:41:15.865 Found net devices under 0000:0a:00.1: cvl_0_1 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@442 -- # is_hw=yes 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:41:15.865 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:41:16.124 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:41:16.124 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:41:16.124 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:41:16.124 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:41:16.124 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:41:16.124 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.291 ms 00:41:16.124 00:41:16.124 --- 10.0.0.2 ping statistics --- 00:41:16.124 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:41:16.124 rtt min/avg/max/mdev = 0.291/0.291/0.291/0.000 ms 00:41:16.124 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:41:16.124 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:41:16.124 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.163 ms 00:41:16.124 00:41:16.124 --- 10.0.0.1 ping statistics --- 00:41:16.124 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:41:16.124 rtt min/avg/max/mdev = 0.163/0.163/0.163/0.000 ms 00:41:16.124 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:41:16.124 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@450 -- # return 0 00:41:16.124 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:41:16.124 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:41:16.124 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:41:16.124 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:41:16.124 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:41:16.124 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:41:16.124 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:41:16.124 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@15 -- # nvmfappstart -m 0xF --wait-for-rpc 00:41:16.124 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:41:16.124 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@726 -- # xtrace_disable 00:41:16.124 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:41:16.124 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@509 -- # nvmfpid=443005 00:41:16.124 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@510 -- # waitforlisten 443005 00:41:16.124 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@835 -- # '[' -z 443005 ']' 00:41:16.124 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:41:16.124 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0xF --wait-for-rpc 00:41:16.124 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@840 -- # local max_retries=100 00:41:16.124 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:41:16.124 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:41:16.124 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@844 -- # xtrace_disable 00:41:16.124 17:56:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:41:16.124 [2024-12-06 17:56:57.842612] thread.c:3005:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:41:16.124 [2024-12-06 17:56:57.843790] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:41:16.124 [2024-12-06 17:56:57.843851] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:41:16.124 [2024-12-06 17:56:57.918458] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:41:16.384 [2024-12-06 17:56:57.970739] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:41:16.384 [2024-12-06 17:56:57.970789] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:41:16.384 [2024-12-06 17:56:57.970805] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:41:16.384 [2024-12-06 17:56:57.970818] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:41:16.384 [2024-12-06 17:56:57.970828] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:41:16.384 [2024-12-06 17:56:57.972307] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:41:16.384 [2024-12-06 17:56:57.972369] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:41:16.384 [2024-12-06 17:56:57.972433] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:41:16.384 [2024-12-06 17:56:57.972436] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:41:16.384 [2024-12-06 17:56:57.972915] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:41:16.384 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:41:16.384 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@868 -- # return 0 00:41:16.384 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:41:16.384 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@732 -- # xtrace_disable 00:41:16.384 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:41:16.384 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:41:16.384 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@18 -- # rpc_cmd bdev_set_options -p 5 -c 1 00:41:16.384 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:41:16.384 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:41:16.384 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:41:16.384 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@19 -- # rpc_cmd framework_start_init 00:41:16.384 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:41:16.384 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:41:16.384 [2024-12-06 17:56:58.172785] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:41:16.384 [2024-12-06 17:56:58.172994] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:41:16.384 [2024-12-06 17:56:58.173958] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:41:16.384 [2024-12-06 17:56:58.174739] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_003) to intr mode from intr mode. 00:41:16.384 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:41:16.384 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@20 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:41:16.384 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:41:16.384 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:41:16.384 [2024-12-06 17:56:58.181110] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:41:16.384 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:41:16.384 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@22 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:41:16.384 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:41:16.384 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:41:16.384 Malloc0 00:41:16.384 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:41:16.384 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@23 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:41:16.384 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:41:16.384 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:41:16.643 [2024-12-06 17:56:58.237299] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@28 -- # WRITE_PID=443148 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@27 -- # gen_nvmf_target_json 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x10 -i 1 --json /dev/fd/63 -q 128 -o 4096 -w write -t 1 -s 256 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@30 -- # READ_PID=443150 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # config=() 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # local subsystem config 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:41:16.643 { 00:41:16.643 "params": { 00:41:16.643 "name": "Nvme$subsystem", 00:41:16.643 "trtype": "$TEST_TRANSPORT", 00:41:16.643 "traddr": "$NVMF_FIRST_TARGET_IP", 00:41:16.643 "adrfam": "ipv4", 00:41:16.643 "trsvcid": "$NVMF_PORT", 00:41:16.643 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:41:16.643 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:41:16.643 "hdgst": ${hdgst:-false}, 00:41:16.643 "ddgst": ${ddgst:-false} 00:41:16.643 }, 00:41:16.643 "method": "bdev_nvme_attach_controller" 00:41:16.643 } 00:41:16.643 EOF 00:41:16.643 )") 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@29 -- # gen_nvmf_target_json 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x20 -i 2 --json /dev/fd/63 -q 128 -o 4096 -w read -t 1 -s 256 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@32 -- # FLUSH_PID=443152 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # config=() 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # local subsystem config 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:41:16.643 { 00:41:16.643 "params": { 00:41:16.643 "name": "Nvme$subsystem", 00:41:16.643 "trtype": "$TEST_TRANSPORT", 00:41:16.643 "traddr": "$NVMF_FIRST_TARGET_IP", 00:41:16.643 "adrfam": "ipv4", 00:41:16.643 "trsvcid": "$NVMF_PORT", 00:41:16.643 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:41:16.643 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:41:16.643 "hdgst": ${hdgst:-false}, 00:41:16.643 "ddgst": ${ddgst:-false} 00:41:16.643 }, 00:41:16.643 "method": "bdev_nvme_attach_controller" 00:41:16.643 } 00:41:16.643 EOF 00:41:16.643 )") 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@34 -- # UNMAP_PID=443155 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@31 -- # gen_nvmf_target_json 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x40 -i 3 --json /dev/fd/63 -q 128 -o 4096 -w flush -t 1 -s 256 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # cat 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@35 -- # sync 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # config=() 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # local subsystem config 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:41:16.643 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:41:16.643 { 00:41:16.643 "params": { 00:41:16.643 "name": "Nvme$subsystem", 00:41:16.643 "trtype": "$TEST_TRANSPORT", 00:41:16.643 "traddr": "$NVMF_FIRST_TARGET_IP", 00:41:16.643 "adrfam": "ipv4", 00:41:16.643 "trsvcid": "$NVMF_PORT", 00:41:16.643 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:41:16.643 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:41:16.643 "hdgst": ${hdgst:-false}, 00:41:16.643 "ddgst": ${ddgst:-false} 00:41:16.643 }, 00:41:16.643 "method": "bdev_nvme_attach_controller" 00:41:16.643 } 00:41:16.643 EOF 00:41:16.643 )") 00:41:16.644 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x80 -i 4 --json /dev/fd/63 -q 128 -o 4096 -w unmap -t 1 -s 256 00:41:16.644 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@33 -- # gen_nvmf_target_json 00:41:16.644 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # config=() 00:41:16.644 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # cat 00:41:16.644 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # local subsystem config 00:41:16.644 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:41:16.644 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:41:16.644 { 00:41:16.644 "params": { 00:41:16.644 "name": "Nvme$subsystem", 00:41:16.644 "trtype": "$TEST_TRANSPORT", 00:41:16.644 "traddr": "$NVMF_FIRST_TARGET_IP", 00:41:16.644 "adrfam": "ipv4", 00:41:16.644 "trsvcid": "$NVMF_PORT", 00:41:16.644 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:41:16.644 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:41:16.644 "hdgst": ${hdgst:-false}, 00:41:16.644 "ddgst": ${ddgst:-false} 00:41:16.644 }, 00:41:16.644 "method": "bdev_nvme_attach_controller" 00:41:16.644 } 00:41:16.644 EOF 00:41:16.644 )") 00:41:16.644 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # cat 00:41:16.644 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@37 -- # wait 443148 00:41:16.644 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # cat 00:41:16.644 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # jq . 00:41:16.644 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # jq . 00:41:16.644 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # jq . 00:41:16.644 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@585 -- # IFS=, 00:41:16.644 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:41:16.644 "params": { 00:41:16.644 "name": "Nvme1", 00:41:16.644 "trtype": "tcp", 00:41:16.644 "traddr": "10.0.0.2", 00:41:16.644 "adrfam": "ipv4", 00:41:16.644 "trsvcid": "4420", 00:41:16.644 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:41:16.644 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:41:16.644 "hdgst": false, 00:41:16.644 "ddgst": false 00:41:16.644 }, 00:41:16.644 "method": "bdev_nvme_attach_controller" 00:41:16.644 }' 00:41:16.644 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # jq . 00:41:16.644 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@585 -- # IFS=, 00:41:16.644 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:41:16.644 "params": { 00:41:16.644 "name": "Nvme1", 00:41:16.644 "trtype": "tcp", 00:41:16.644 "traddr": "10.0.0.2", 00:41:16.644 "adrfam": "ipv4", 00:41:16.644 "trsvcid": "4420", 00:41:16.644 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:41:16.644 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:41:16.644 "hdgst": false, 00:41:16.644 "ddgst": false 00:41:16.644 }, 00:41:16.644 "method": "bdev_nvme_attach_controller" 00:41:16.644 }' 00:41:16.644 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@585 -- # IFS=, 00:41:16.644 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@585 -- # IFS=, 00:41:16.644 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:41:16.644 "params": { 00:41:16.644 "name": "Nvme1", 00:41:16.644 "trtype": "tcp", 00:41:16.644 "traddr": "10.0.0.2", 00:41:16.644 "adrfam": "ipv4", 00:41:16.644 "trsvcid": "4420", 00:41:16.644 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:41:16.644 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:41:16.644 "hdgst": false, 00:41:16.644 "ddgst": false 00:41:16.644 }, 00:41:16.644 "method": "bdev_nvme_attach_controller" 00:41:16.644 }' 00:41:16.644 17:56:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:41:16.644 "params": { 00:41:16.644 "name": "Nvme1", 00:41:16.644 "trtype": "tcp", 00:41:16.644 "traddr": "10.0.0.2", 00:41:16.644 "adrfam": "ipv4", 00:41:16.644 "trsvcid": "4420", 00:41:16.644 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:41:16.644 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:41:16.644 "hdgst": false, 00:41:16.644 "ddgst": false 00:41:16.644 }, 00:41:16.644 "method": "bdev_nvme_attach_controller" 00:41:16.644 }' 00:41:16.644 [2024-12-06 17:56:58.289636] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:41:16.644 [2024-12-06 17:56:58.289636] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:41:16.644 [2024-12-06 17:56:58.289636] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:41:16.644 [2024-12-06 17:56:58.289635] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:41:16.644 [2024-12-06 17:56:58.289756] [ DPDK EAL parameters: bdevperf -c 0x40 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib[2024-12-06 17:56:58.289756] [ DPDK EAL parameters: bdevperf -c 0x80 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib[2024-12-06 17:56:58.289756] [ DPDK EAL parameters: bdevperf -c 0x10 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib[2024-12-06 17:56:58.289757] [ DPDK EAL parameters: bdevperf -c 0x20 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk3 .cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk4 .cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk1 .cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk2 --proc-type=auto ] 00:41:16.644 --proc-type=auto ] 00:41:16.644 --proc-type=auto ] 00:41:16.644 --proc-type=auto ] 00:41:16.644 [2024-12-06 17:56:58.465750] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:41:16.902 [2024-12-06 17:56:58.508098] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 5 00:41:16.902 [2024-12-06 17:56:58.563513] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:41:16.902 [2024-12-06 17:56:58.605020] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 7 00:41:16.902 [2024-12-06 17:56:58.659703] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:41:16.902 [2024-12-06 17:56:58.702488] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 6 00:41:17.160 [2024-12-06 17:56:58.759810] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:41:17.160 [2024-12-06 17:56:58.798855] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:41:17.160 Running I/O for 1 seconds... 00:41:17.419 Running I/O for 1 seconds... 00:41:17.419 Running I/O for 1 seconds... 00:41:17.419 Running I/O for 1 seconds... 00:41:18.353 11453.00 IOPS, 44.74 MiB/s 00:41:18.353 Latency(us) 00:41:18.353 [2024-12-06T16:57:00.192Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:41:18.353 Job: Nvme1n1 (Core Mask 0x80, workload: unmap, depth: 128, IO size: 4096) 00:41:18.353 Nvme1n1 : 1.01 11519.15 45.00 0.00 0.00 11073.60 4126.34 13204.29 00:41:18.353 [2024-12-06T16:57:00.192Z] =================================================================================================================== 00:41:18.353 [2024-12-06T16:57:00.192Z] Total : 11519.15 45.00 0.00 0.00 11073.60 4126.34 13204.29 00:41:18.353 8337.00 IOPS, 32.57 MiB/s 00:41:18.353 Latency(us) 00:41:18.353 [2024-12-06T16:57:00.192Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:41:18.353 Job: Nvme1n1 (Core Mask 0x20, workload: read, depth: 128, IO size: 4096) 00:41:18.353 Nvme1n1 : 1.01 8383.94 32.75 0.00 0.00 15188.84 4660.34 17573.36 00:41:18.353 [2024-12-06T16:57:00.192Z] =================================================================================================================== 00:41:18.353 [2024-12-06T16:57:00.192Z] Total : 8383.94 32.75 0.00 0.00 15188.84 4660.34 17573.36 00:41:18.353 188296.00 IOPS, 735.53 MiB/s 00:41:18.353 Latency(us) 00:41:18.353 [2024-12-06T16:57:00.192Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:41:18.353 Job: Nvme1n1 (Core Mask 0x40, workload: flush, depth: 128, IO size: 4096) 00:41:18.353 Nvme1n1 : 1.00 187931.71 734.11 0.00 0.00 677.27 304.92 1917.53 00:41:18.353 [2024-12-06T16:57:00.192Z] =================================================================================================================== 00:41:18.353 [2024-12-06T16:57:00.193Z] Total : 187931.71 734.11 0.00 0.00 677.27 304.92 1917.53 00:41:18.354 8913.00 IOPS, 34.82 MiB/s 00:41:18.354 Latency(us) 00:41:18.354 [2024-12-06T16:57:00.193Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:41:18.354 Job: Nvme1n1 (Core Mask 0x10, workload: write, depth: 128, IO size: 4096) 00:41:18.354 Nvme1n1 : 1.01 8989.19 35.11 0.00 0.00 14186.11 1929.67 20874.43 00:41:18.354 [2024-12-06T16:57:00.193Z] =================================================================================================================== 00:41:18.354 [2024-12-06T16:57:00.193Z] Total : 8989.19 35.11 0.00 0.00 14186.11 1929.67 20874.43 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@38 -- # wait 443150 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@39 -- # wait 443152 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@40 -- # wait 443155 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@42 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@563 -- # xtrace_disable 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@44 -- # trap - SIGINT SIGTERM EXIT 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@46 -- # nvmftestfini 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@516 -- # nvmfcleanup 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@121 -- # sync 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@124 -- # set +e 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@125 -- # for i in {1..20} 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:41:18.612 rmmod nvme_tcp 00:41:18.612 rmmod nvme_fabrics 00:41:18.612 rmmod nvme_keyring 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@128 -- # set -e 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@129 -- # return 0 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@517 -- # '[' -n 443005 ']' 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@518 -- # killprocess 443005 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@954 -- # '[' -z 443005 ']' 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@958 -- # kill -0 443005 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@959 -- # uname 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 443005 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@972 -- # echo 'killing process with pid 443005' 00:41:18.612 killing process with pid 443005 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@973 -- # kill 443005 00:41:18.612 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@978 -- # wait 443005 00:41:18.871 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:41:18.871 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:41:18.871 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:41:18.871 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@297 -- # iptr 00:41:18.871 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@791 -- # iptables-save 00:41:18.871 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:41:18.871 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@791 -- # iptables-restore 00:41:18.871 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:41:18.871 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@302 -- # remove_spdk_ns 00:41:18.871 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:41:18.871 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:41:18.871 17:57:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:41:20.772 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:41:20.772 00:41:20.772 real 0m7.287s 00:41:20.772 user 0m14.278s 00:41:20.772 sys 0m4.183s 00:41:20.772 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1130 -- # xtrace_disable 00:41:20.772 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:41:20.772 ************************************ 00:41:20.772 END TEST nvmf_bdev_io_wait 00:41:20.772 ************************************ 00:41:21.031 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@30 -- # run_test nvmf_queue_depth /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/queue_depth.sh --transport=tcp --interrupt-mode 00:41:21.031 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:41:21.031 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1111 -- # xtrace_disable 00:41:21.031 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:41:21.031 ************************************ 00:41:21.031 START TEST nvmf_queue_depth 00:41:21.031 ************************************ 00:41:21.031 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/queue_depth.sh --transport=tcp --interrupt-mode 00:41:21.031 * Looking for test storage... 00:41:21.031 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:41:21.031 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:41:21.031 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1711 -- # lcov --version 00:41:21.031 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:41:21.031 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:41:21.031 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:41:21.031 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@333 -- # local ver1 ver1_l 00:41:21.031 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@334 -- # local ver2 ver2_l 00:41:21.031 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@336 -- # IFS=.-: 00:41:21.031 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@336 -- # read -ra ver1 00:41:21.031 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@337 -- # IFS=.-: 00:41:21.031 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@337 -- # read -ra ver2 00:41:21.031 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@338 -- # local 'op=<' 00:41:21.031 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@340 -- # ver1_l=2 00:41:21.031 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@341 -- # ver2_l=1 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@344 -- # case "$op" in 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@345 -- # : 1 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@364 -- # (( v = 0 )) 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@365 -- # decimal 1 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@353 -- # local d=1 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@355 -- # echo 1 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@365 -- # ver1[v]=1 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@366 -- # decimal 2 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@353 -- # local d=2 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@355 -- # echo 2 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@366 -- # ver2[v]=2 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@368 -- # return 0 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:41:21.032 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:41:21.032 --rc genhtml_branch_coverage=1 00:41:21.032 --rc genhtml_function_coverage=1 00:41:21.032 --rc genhtml_legend=1 00:41:21.032 --rc geninfo_all_blocks=1 00:41:21.032 --rc geninfo_unexecuted_blocks=1 00:41:21.032 00:41:21.032 ' 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:41:21.032 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:41:21.032 --rc genhtml_branch_coverage=1 00:41:21.032 --rc genhtml_function_coverage=1 00:41:21.032 --rc genhtml_legend=1 00:41:21.032 --rc geninfo_all_blocks=1 00:41:21.032 --rc geninfo_unexecuted_blocks=1 00:41:21.032 00:41:21.032 ' 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:41:21.032 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:41:21.032 --rc genhtml_branch_coverage=1 00:41:21.032 --rc genhtml_function_coverage=1 00:41:21.032 --rc genhtml_legend=1 00:41:21.032 --rc geninfo_all_blocks=1 00:41:21.032 --rc geninfo_unexecuted_blocks=1 00:41:21.032 00:41:21.032 ' 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:41:21.032 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:41:21.032 --rc genhtml_branch_coverage=1 00:41:21.032 --rc genhtml_function_coverage=1 00:41:21.032 --rc genhtml_legend=1 00:41:21.032 --rc geninfo_all_blocks=1 00:41:21.032 --rc geninfo_unexecuted_blocks=1 00:41:21.032 00:41:21.032 ' 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@12 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@7 -- # uname -s 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@15 -- # shopt -s extglob 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:41:21.032 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- paths/export.sh@5 -- # export PATH 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@51 -- # : 0 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@55 -- # have_pci_nics=0 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@14 -- # MALLOC_BDEV_SIZE=64 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@15 -- # MALLOC_BLOCK_SIZE=512 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@17 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@19 -- # nvmftestinit 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@476 -- # prepare_net_devs 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@438 -- # local -g is_hw=no 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@440 -- # remove_spdk_ns 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@309 -- # xtrace_disable 00:41:21.033 17:57:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@315 -- # pci_devs=() 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@315 -- # local -a pci_devs 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@316 -- # pci_net_devs=() 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@317 -- # pci_drivers=() 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@317 -- # local -A pci_drivers 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@319 -- # net_devs=() 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@319 -- # local -ga net_devs 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@320 -- # e810=() 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@320 -- # local -ga e810 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@321 -- # x722=() 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@321 -- # local -ga x722 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@322 -- # mlx=() 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@322 -- # local -ga mlx 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:41:23.565 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:41:23.565 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@418 -- # [[ up == up ]] 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:41:23.565 Found net devices under 0000:0a:00.0: cvl_0_0 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@418 -- # [[ up == up ]] 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:41:23.565 Found net devices under 0000:0a:00.1: cvl_0_1 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@442 -- # is_hw=yes 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:41:23.565 17:57:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:41:23.565 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:41:23.565 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:41:23.565 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:41:23.565 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:41:23.565 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:41:23.566 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:41:23.566 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.277 ms 00:41:23.566 00:41:23.566 --- 10.0.0.2 ping statistics --- 00:41:23.566 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:41:23.566 rtt min/avg/max/mdev = 0.277/0.277/0.277/0.000 ms 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:41:23.566 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:41:23.566 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.163 ms 00:41:23.566 00:41:23.566 --- 10.0.0.1 ping statistics --- 00:41:23.566 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:41:23.566 rtt min/avg/max/mdev = 0.163/0.163/0.163/0.000 ms 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@450 -- # return 0 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@21 -- # nvmfappstart -m 0x2 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@726 -- # xtrace_disable 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@509 -- # nvmfpid=445480 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@510 -- # waitforlisten 445480 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x2 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@835 -- # '[' -z 445480 ']' 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@840 -- # local max_retries=100 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:41:23.566 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@844 -- # xtrace_disable 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:41:23.566 [2024-12-06 17:57:05.142489] thread.c:3005:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:41:23.566 [2024-12-06 17:57:05.143603] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:41:23.566 [2024-12-06 17:57:05.143684] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:41:23.566 [2024-12-06 17:57:05.219611] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:41:23.566 [2024-12-06 17:57:05.265997] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:41:23.566 [2024-12-06 17:57:05.266053] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:41:23.566 [2024-12-06 17:57:05.266067] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:41:23.566 [2024-12-06 17:57:05.266079] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:41:23.566 [2024-12-06 17:57:05.266088] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:41:23.566 [2024-12-06 17:57:05.266642] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:41:23.566 [2024-12-06 17:57:05.361778] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:41:23.566 [2024-12-06 17:57:05.362061] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@868 -- # return 0 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@732 -- # xtrace_disable 00:41:23.566 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@23 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@563 -- # xtrace_disable 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:41:23.823 [2024-12-06 17:57:05.411195] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@24 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@563 -- # xtrace_disable 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:41:23.823 Malloc0 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@25 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@563 -- # xtrace_disable 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@26 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@563 -- # xtrace_disable 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@563 -- # xtrace_disable 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:41:23.823 [2024-12-06 17:57:05.479338] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@30 -- # bdevperf_pid=445511 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -z -r /var/tmp/bdevperf.sock -q 1024 -o 4096 -w verify -t 10 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@32 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@33 -- # waitforlisten 445511 /var/tmp/bdevperf.sock 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@835 -- # '[' -z 445511 ']' 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@840 -- # local max_retries=100 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:41:23.823 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@844 -- # xtrace_disable 00:41:23.823 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:41:23.823 [2024-12-06 17:57:05.525461] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:41:23.823 [2024-12-06 17:57:05.525527] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid445511 ] 00:41:23.823 [2024-12-06 17:57:05.591040] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:41:23.824 [2024-12-06 17:57:05.635987] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:41:24.081 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:41:24.081 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@868 -- # return 0 00:41:24.081 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@34 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:41:24.081 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@563 -- # xtrace_disable 00:41:24.081 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:41:24.081 NVMe0n1 00:41:24.081 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:41:24.081 17:57:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:41:24.338 Running I/O for 10 seconds... 00:41:26.206 8192.00 IOPS, 32.00 MiB/s [2024-12-06T16:57:09.415Z] 8550.50 IOPS, 33.40 MiB/s [2024-12-06T16:57:10.348Z] 8533.67 IOPS, 33.33 MiB/s [2024-12-06T16:57:11.282Z] 8586.25 IOPS, 33.54 MiB/s [2024-12-06T16:57:12.241Z] 8608.20 IOPS, 33.63 MiB/s [2024-12-06T16:57:13.280Z] 8682.50 IOPS, 33.92 MiB/s [2024-12-06T16:57:14.215Z] 8658.71 IOPS, 33.82 MiB/s [2024-12-06T16:57:15.150Z] 8707.25 IOPS, 34.01 MiB/s [2024-12-06T16:57:16.083Z] 8753.22 IOPS, 34.19 MiB/s [2024-12-06T16:57:16.083Z] 8762.40 IOPS, 34.23 MiB/s 00:41:34.244 Latency(us) 00:41:34.244 [2024-12-06T16:57:16.083Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:41:34.244 Job: NVMe0n1 (Core Mask 0x1, workload: verify, depth: 1024, IO size: 4096) 00:41:34.244 Verification LBA range: start 0x0 length 0x4000 00:41:34.244 NVMe0n1 : 10.08 8788.99 34.33 0.00 0.00 115938.73 22330.79 72623.60 00:41:34.244 [2024-12-06T16:57:16.083Z] =================================================================================================================== 00:41:34.244 [2024-12-06T16:57:16.083Z] Total : 8788.99 34.33 0.00 0.00 115938.73 22330.79 72623.60 00:41:34.244 { 00:41:34.244 "results": [ 00:41:34.244 { 00:41:34.244 "job": "NVMe0n1", 00:41:34.244 "core_mask": "0x1", 00:41:34.244 "workload": "verify", 00:41:34.244 "status": "finished", 00:41:34.244 "verify_range": { 00:41:34.244 "start": 0, 00:41:34.244 "length": 16384 00:41:34.244 }, 00:41:34.244 "queue_depth": 1024, 00:41:34.244 "io_size": 4096, 00:41:34.244 "runtime": 10.077832, 00:41:34.244 "iops": 8788.993505746077, 00:41:34.244 "mibps": 34.33200588182061, 00:41:34.244 "io_failed": 0, 00:41:34.244 "io_timeout": 0, 00:41:34.244 "avg_latency_us": 115938.73210729007, 00:41:34.244 "min_latency_us": 22330.785185185185, 00:41:34.244 "max_latency_us": 72623.59703703703 00:41:34.244 } 00:41:34.244 ], 00:41:34.244 "core_count": 1 00:41:34.244 } 00:41:34.502 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@39 -- # killprocess 445511 00:41:34.502 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@954 -- # '[' -z 445511 ']' 00:41:34.502 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@958 -- # kill -0 445511 00:41:34.502 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@959 -- # uname 00:41:34.502 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:41:34.502 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 445511 00:41:34.502 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:41:34.502 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:41:34.502 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@972 -- # echo 'killing process with pid 445511' 00:41:34.502 killing process with pid 445511 00:41:34.502 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@973 -- # kill 445511 00:41:34.502 Received shutdown signal, test time was about 10.000000 seconds 00:41:34.502 00:41:34.502 Latency(us) 00:41:34.502 [2024-12-06T16:57:16.341Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:41:34.502 [2024-12-06T16:57:16.341Z] =================================================================================================================== 00:41:34.502 [2024-12-06T16:57:16.341Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:41:34.502 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@978 -- # wait 445511 00:41:34.502 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@41 -- # trap - SIGINT SIGTERM EXIT 00:41:34.502 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@43 -- # nvmftestfini 00:41:34.502 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@516 -- # nvmfcleanup 00:41:34.502 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@121 -- # sync 00:41:34.502 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:41:34.502 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@124 -- # set +e 00:41:34.502 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@125 -- # for i in {1..20} 00:41:34.502 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:41:34.502 rmmod nvme_tcp 00:41:34.502 rmmod nvme_fabrics 00:41:34.759 rmmod nvme_keyring 00:41:34.759 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:41:34.759 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@128 -- # set -e 00:41:34.759 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@129 -- # return 0 00:41:34.759 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@517 -- # '[' -n 445480 ']' 00:41:34.759 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@518 -- # killprocess 445480 00:41:34.759 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@954 -- # '[' -z 445480 ']' 00:41:34.759 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@958 -- # kill -0 445480 00:41:34.759 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@959 -- # uname 00:41:34.759 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:41:34.759 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 445480 00:41:34.759 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:41:34.759 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:41:34.759 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@972 -- # echo 'killing process with pid 445480' 00:41:34.759 killing process with pid 445480 00:41:34.759 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@973 -- # kill 445480 00:41:34.759 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@978 -- # wait 445480 00:41:35.019 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:41:35.019 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:41:35.019 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:41:35.019 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@297 -- # iptr 00:41:35.019 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@791 -- # iptables-save 00:41:35.019 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:41:35.019 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@791 -- # iptables-restore 00:41:35.019 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:41:35.019 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@302 -- # remove_spdk_ns 00:41:35.019 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:41:35.019 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:41:35.019 17:57:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:41:36.925 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:41:36.925 00:41:36.925 real 0m16.034s 00:41:36.925 user 0m22.072s 00:41:36.925 sys 0m3.416s 00:41:36.925 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1130 -- # xtrace_disable 00:41:36.925 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:41:36.925 ************************************ 00:41:36.925 END TEST nvmf_queue_depth 00:41:36.925 ************************************ 00:41:36.925 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@31 -- # run_test nvmf_target_multipath /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multipath.sh --transport=tcp --interrupt-mode 00:41:36.925 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:41:36.925 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1111 -- # xtrace_disable 00:41:36.925 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:41:36.925 ************************************ 00:41:36.925 START TEST nvmf_target_multipath 00:41:36.925 ************************************ 00:41:36.925 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multipath.sh --transport=tcp --interrupt-mode 00:41:37.185 * Looking for test storage... 00:41:37.185 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1711 -- # lcov --version 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@333 -- # local ver1 ver1_l 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@334 -- # local ver2 ver2_l 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@336 -- # IFS=.-: 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@336 -- # read -ra ver1 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@337 -- # IFS=.-: 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@337 -- # read -ra ver2 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@338 -- # local 'op=<' 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@340 -- # ver1_l=2 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@341 -- # ver2_l=1 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@344 -- # case "$op" in 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@345 -- # : 1 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@364 -- # (( v = 0 )) 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@365 -- # decimal 1 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@353 -- # local d=1 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@355 -- # echo 1 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@365 -- # ver1[v]=1 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@366 -- # decimal 2 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@353 -- # local d=2 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@355 -- # echo 2 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@366 -- # ver2[v]=2 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@368 -- # return 0 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:41:37.185 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:41:37.185 --rc genhtml_branch_coverage=1 00:41:37.185 --rc genhtml_function_coverage=1 00:41:37.185 --rc genhtml_legend=1 00:41:37.185 --rc geninfo_all_blocks=1 00:41:37.185 --rc geninfo_unexecuted_blocks=1 00:41:37.185 00:41:37.185 ' 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:41:37.185 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:41:37.185 --rc genhtml_branch_coverage=1 00:41:37.185 --rc genhtml_function_coverage=1 00:41:37.185 --rc genhtml_legend=1 00:41:37.185 --rc geninfo_all_blocks=1 00:41:37.185 --rc geninfo_unexecuted_blocks=1 00:41:37.185 00:41:37.185 ' 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:41:37.185 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:41:37.185 --rc genhtml_branch_coverage=1 00:41:37.185 --rc genhtml_function_coverage=1 00:41:37.185 --rc genhtml_legend=1 00:41:37.185 --rc geninfo_all_blocks=1 00:41:37.185 --rc geninfo_unexecuted_blocks=1 00:41:37.185 00:41:37.185 ' 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:41:37.185 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:41:37.185 --rc genhtml_branch_coverage=1 00:41:37.185 --rc genhtml_function_coverage=1 00:41:37.185 --rc genhtml_legend=1 00:41:37.185 --rc geninfo_all_blocks=1 00:41:37.185 --rc geninfo_unexecuted_blocks=1 00:41:37.185 00:41:37.185 ' 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@7 -- # uname -s 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:41:37.185 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@15 -- # shopt -s extglob 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- paths/export.sh@5 -- # export PATH 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@51 -- # : 0 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@55 -- # have_pci_nics=0 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@11 -- # MALLOC_BDEV_SIZE=64 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@13 -- # nqn=nqn.2016-06.io.spdk:cnode1 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@15 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@43 -- # nvmftestinit 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@476 -- # prepare_net_devs 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@438 -- # local -g is_hw=no 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@440 -- # remove_spdk_ns 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@309 -- # xtrace_disable 00:41:37.186 17:57:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@10 -- # set +x 00:41:39.723 17:57:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:41:39.723 17:57:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@315 -- # pci_devs=() 00:41:39.723 17:57:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@315 -- # local -a pci_devs 00:41:39.723 17:57:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@316 -- # pci_net_devs=() 00:41:39.723 17:57:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:41:39.723 17:57:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@317 -- # pci_drivers=() 00:41:39.723 17:57:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@317 -- # local -A pci_drivers 00:41:39.723 17:57:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@319 -- # net_devs=() 00:41:39.723 17:57:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@319 -- # local -ga net_devs 00:41:39.723 17:57:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@320 -- # e810=() 00:41:39.723 17:57:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@320 -- # local -ga e810 00:41:39.723 17:57:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@321 -- # x722=() 00:41:39.723 17:57:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@321 -- # local -ga x722 00:41:39.723 17:57:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@322 -- # mlx=() 00:41:39.723 17:57:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@322 -- # local -ga mlx 00:41:39.723 17:57:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:41:39.723 17:57:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:41:39.723 17:57:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:41:39.723 17:57:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:41:39.723 17:57:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:41:39.723 17:57:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:41:39.723 17:57:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:41:39.723 17:57:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:41:39.723 17:57:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:41:39.723 17:57:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:41:39.723 17:57:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:41:39.723 17:57:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:41:39.723 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:41:39.723 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@418 -- # [[ up == up ]] 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:41:39.723 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:41:39.723 Found net devices under 0000:0a:00.0: cvl_0_0 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@418 -- # [[ up == up ]] 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:41:39.724 Found net devices under 0000:0a:00.1: cvl_0_1 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@442 -- # is_hw=yes 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:41:39.724 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:41:39.724 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.188 ms 00:41:39.724 00:41:39.724 --- 10.0.0.2 ping statistics --- 00:41:39.724 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:41:39.724 rtt min/avg/max/mdev = 0.188/0.188/0.188/0.000 ms 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:41:39.724 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:41:39.724 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.096 ms 00:41:39.724 00:41:39.724 --- 10.0.0.1 ping statistics --- 00:41:39.724 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:41:39.724 rtt min/avg/max/mdev = 0.096/0.096/0.096/0.000 ms 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@450 -- # return 0 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@45 -- # '[' -z ']' 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@46 -- # echo 'only one NIC for nvmf test' 00:41:39.724 only one NIC for nvmf test 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@47 -- # nvmftestfini 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@516 -- # nvmfcleanup 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@121 -- # sync 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@124 -- # set +e 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@125 -- # for i in {1..20} 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:41:39.724 rmmod nvme_tcp 00:41:39.724 rmmod nvme_fabrics 00:41:39.724 rmmod nvme_keyring 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@128 -- # set -e 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@129 -- # return 0 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@517 -- # '[' -n '' ']' 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@297 -- # iptr 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@791 -- # iptables-save 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@791 -- # iptables-restore 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@302 -- # remove_spdk_ns 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:41:39.724 17:57:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@48 -- # exit 0 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@1 -- # nvmftestfini 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@516 -- # nvmfcleanup 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@121 -- # sync 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@124 -- # set +e 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@125 -- # for i in {1..20} 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@128 -- # set -e 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@129 -- # return 0 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@517 -- # '[' -n '' ']' 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@297 -- # iptr 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@791 -- # iptables-save 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@791 -- # iptables-restore 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@302 -- # remove_spdk_ns 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:41:41.627 00:41:41.627 real 0m4.556s 00:41:41.627 user 0m0.915s 00:41:41.627 sys 0m1.647s 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1130 -- # xtrace_disable 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@10 -- # set +x 00:41:41.627 ************************************ 00:41:41.627 END TEST nvmf_target_multipath 00:41:41.627 ************************************ 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@32 -- # run_test nvmf_zcopy /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh --transport=tcp --interrupt-mode 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1111 -- # xtrace_disable 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:41:41.627 ************************************ 00:41:41.627 START TEST nvmf_zcopy 00:41:41.627 ************************************ 00:41:41.627 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh --transport=tcp --interrupt-mode 00:41:41.627 * Looking for test storage... 00:41:41.627 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:41:41.628 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:41:41.628 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1711 -- # lcov --version 00:41:41.628 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@333 -- # local ver1 ver1_l 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@334 -- # local ver2 ver2_l 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@336 -- # IFS=.-: 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@336 -- # read -ra ver1 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@337 -- # IFS=.-: 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@337 -- # read -ra ver2 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@338 -- # local 'op=<' 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@340 -- # ver1_l=2 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@341 -- # ver2_l=1 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@344 -- # case "$op" in 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@345 -- # : 1 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@364 -- # (( v = 0 )) 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@365 -- # decimal 1 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@353 -- # local d=1 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@355 -- # echo 1 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@365 -- # ver1[v]=1 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@366 -- # decimal 2 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@353 -- # local d=2 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@355 -- # echo 2 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@366 -- # ver2[v]=2 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@368 -- # return 0 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:41:41.887 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:41:41.887 --rc genhtml_branch_coverage=1 00:41:41.887 --rc genhtml_function_coverage=1 00:41:41.887 --rc genhtml_legend=1 00:41:41.887 --rc geninfo_all_blocks=1 00:41:41.887 --rc geninfo_unexecuted_blocks=1 00:41:41.887 00:41:41.887 ' 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:41:41.887 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:41:41.887 --rc genhtml_branch_coverage=1 00:41:41.887 --rc genhtml_function_coverage=1 00:41:41.887 --rc genhtml_legend=1 00:41:41.887 --rc geninfo_all_blocks=1 00:41:41.887 --rc geninfo_unexecuted_blocks=1 00:41:41.887 00:41:41.887 ' 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:41:41.887 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:41:41.887 --rc genhtml_branch_coverage=1 00:41:41.887 --rc genhtml_function_coverage=1 00:41:41.887 --rc genhtml_legend=1 00:41:41.887 --rc geninfo_all_blocks=1 00:41:41.887 --rc geninfo_unexecuted_blocks=1 00:41:41.887 00:41:41.887 ' 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:41:41.887 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:41:41.887 --rc genhtml_branch_coverage=1 00:41:41.887 --rc genhtml_function_coverage=1 00:41:41.887 --rc genhtml_legend=1 00:41:41.887 --rc geninfo_all_blocks=1 00:41:41.887 --rc geninfo_unexecuted_blocks=1 00:41:41.887 00:41:41.887 ' 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@7 -- # uname -s 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@15 -- # shopt -s extglob 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:41:41.887 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- paths/export.sh@5 -- # export PATH 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@51 -- # : 0 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@55 -- # have_pci_nics=0 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@12 -- # nvmftestinit 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@476 -- # prepare_net_devs 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@438 -- # local -g is_hw=no 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@440 -- # remove_spdk_ns 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@309 -- # xtrace_disable 00:41:41.888 17:57:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:41:44.419 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:41:44.419 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@315 -- # pci_devs=() 00:41:44.419 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@315 -- # local -a pci_devs 00:41:44.419 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@316 -- # pci_net_devs=() 00:41:44.419 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:41:44.419 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@317 -- # pci_drivers=() 00:41:44.419 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@317 -- # local -A pci_drivers 00:41:44.419 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@319 -- # net_devs=() 00:41:44.419 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@319 -- # local -ga net_devs 00:41:44.419 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@320 -- # e810=() 00:41:44.419 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@320 -- # local -ga e810 00:41:44.419 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@321 -- # x722=() 00:41:44.419 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@321 -- # local -ga x722 00:41:44.419 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@322 -- # mlx=() 00:41:44.419 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@322 -- # local -ga mlx 00:41:44.419 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:41:44.419 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:41:44.419 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:41:44.419 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:41:44.419 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:41:44.419 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:41:44.419 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:41:44.419 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:41:44.420 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:41:44.420 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@418 -- # [[ up == up ]] 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:41:44.420 Found net devices under 0000:0a:00.0: cvl_0_0 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@418 -- # [[ up == up ]] 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:41:44.420 Found net devices under 0000:0a:00.1: cvl_0_1 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@442 -- # is_hw=yes 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:41:44.420 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:41:44.420 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:41:44.420 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.301 ms 00:41:44.420 00:41:44.420 --- 10.0.0.2 ping statistics --- 00:41:44.420 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:41:44.421 rtt min/avg/max/mdev = 0.301/0.301/0.301/0.000 ms 00:41:44.421 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:41:44.421 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:41:44.421 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.171 ms 00:41:44.421 00:41:44.421 --- 10.0.0.1 ping statistics --- 00:41:44.421 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:41:44.421 rtt min/avg/max/mdev = 0.171/0.171/0.171/0.000 ms 00:41:44.421 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:41:44.421 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@450 -- # return 0 00:41:44.421 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:41:44.421 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:41:44.421 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:41:44.421 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:41:44.421 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:41:44.421 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:41:44.421 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:41:44.421 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@13 -- # nvmfappstart -m 0x2 00:41:44.421 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:41:44.421 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@726 -- # xtrace_disable 00:41:44.421 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:41:44.421 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@509 -- # nvmfpid=451191 00:41:44.421 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x2 00:41:44.421 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@510 -- # waitforlisten 451191 00:41:44.421 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@835 -- # '[' -z 451191 ']' 00:41:44.421 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:41:44.421 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@840 -- # local max_retries=100 00:41:44.421 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:41:44.421 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:41:44.421 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@844 -- # xtrace_disable 00:41:44.421 17:57:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:41:44.421 [2024-12-06 17:57:25.972604] thread.c:3005:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:41:44.421 [2024-12-06 17:57:25.973782] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:41:44.421 [2024-12-06 17:57:25.973839] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:41:44.421 [2024-12-06 17:57:26.045870] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:41:44.421 [2024-12-06 17:57:26.093448] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:41:44.421 [2024-12-06 17:57:26.093508] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:41:44.421 [2024-12-06 17:57:26.093521] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:41:44.421 [2024-12-06 17:57:26.093532] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:41:44.421 [2024-12-06 17:57:26.093542] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:41:44.421 [2024-12-06 17:57:26.094162] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:41:44.421 [2024-12-06 17:57:26.190443] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:41:44.421 [2024-12-06 17:57:26.190732] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:41:44.421 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:41:44.421 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@868 -- # return 0 00:41:44.421 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:41:44.421 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@732 -- # xtrace_disable 00:41:44.421 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:41:44.421 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:41:44.421 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@15 -- # '[' tcp '!=' tcp ']' 00:41:44.421 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@22 -- # rpc_cmd nvmf_create_transport -t tcp -o -c 0 --zcopy 00:41:44.421 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:41:44.421 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:41:44.421 [2024-12-06 17:57:26.242764] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:41:44.421 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:41:44.421 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:41:44.421 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:41:44.421 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:41:44.421 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:41:44.421 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:41:44.421 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:41:44.421 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:41:44.680 [2024-12-06 17:57:26.258938] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:41:44.680 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:41:44.680 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:41:44.680 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:41:44.680 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:41:44.680 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:41:44.680 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@29 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc0 00:41:44.680 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:41:44.680 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:41:44.680 malloc0 00:41:44.680 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:41:44.680 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@30 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:41:44.680 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:41:44.680 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:41:44.680 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:41:44.680 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -t 10 -q 128 -w verify -o 8192 00:41:44.680 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@33 -- # gen_nvmf_target_json 00:41:44.680 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@560 -- # config=() 00:41:44.680 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@560 -- # local subsystem config 00:41:44.680 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:41:44.680 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:41:44.680 { 00:41:44.680 "params": { 00:41:44.680 "name": "Nvme$subsystem", 00:41:44.680 "trtype": "$TEST_TRANSPORT", 00:41:44.680 "traddr": "$NVMF_FIRST_TARGET_IP", 00:41:44.680 "adrfam": "ipv4", 00:41:44.680 "trsvcid": "$NVMF_PORT", 00:41:44.680 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:41:44.680 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:41:44.680 "hdgst": ${hdgst:-false}, 00:41:44.680 "ddgst": ${ddgst:-false} 00:41:44.680 }, 00:41:44.680 "method": "bdev_nvme_attach_controller" 00:41:44.680 } 00:41:44.680 EOF 00:41:44.680 )") 00:41:44.680 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@582 -- # cat 00:41:44.680 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@584 -- # jq . 00:41:44.680 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@585 -- # IFS=, 00:41:44.680 17:57:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:41:44.680 "params": { 00:41:44.680 "name": "Nvme1", 00:41:44.680 "trtype": "tcp", 00:41:44.680 "traddr": "10.0.0.2", 00:41:44.680 "adrfam": "ipv4", 00:41:44.680 "trsvcid": "4420", 00:41:44.680 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:41:44.680 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:41:44.680 "hdgst": false, 00:41:44.680 "ddgst": false 00:41:44.680 }, 00:41:44.680 "method": "bdev_nvme_attach_controller" 00:41:44.680 }' 00:41:44.680 [2024-12-06 17:57:26.347298] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:41:44.680 [2024-12-06 17:57:26.347381] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid451213 ] 00:41:44.680 [2024-12-06 17:57:26.415581] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:41:44.680 [2024-12-06 17:57:26.464081] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:41:44.938 Running I/O for 10 seconds... 00:41:47.264 5097.00 IOPS, 39.82 MiB/s [2024-12-06T16:57:30.037Z] 5198.50 IOPS, 40.61 MiB/s [2024-12-06T16:57:30.969Z] 5246.00 IOPS, 40.98 MiB/s [2024-12-06T16:57:31.904Z] 5230.25 IOPS, 40.86 MiB/s [2024-12-06T16:57:32.838Z] 5234.20 IOPS, 40.89 MiB/s [2024-12-06T16:57:33.772Z] 5231.67 IOPS, 40.87 MiB/s [2024-12-06T16:57:35.144Z] 5244.43 IOPS, 40.97 MiB/s [2024-12-06T16:57:36.074Z] 5252.00 IOPS, 41.03 MiB/s [2024-12-06T16:57:37.008Z] 5258.89 IOPS, 41.09 MiB/s [2024-12-06T16:57:37.008Z] 5269.20 IOPS, 41.17 MiB/s 00:41:55.169 Latency(us) 00:41:55.169 [2024-12-06T16:57:37.008Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:41:55.169 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 8192) 00:41:55.169 Verification LBA range: start 0x0 length 0x1000 00:41:55.169 Nvme1n1 : 10.01 5271.85 41.19 0.00 0.00 24217.23 3592.34 32234.00 00:41:55.169 [2024-12-06T16:57:37.008Z] =================================================================================================================== 00:41:55.169 [2024-12-06T16:57:37.008Z] Total : 5271.85 41.19 0.00 0.00 24217.23 3592.34 32234.00 00:41:55.169 17:57:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@39 -- # perfpid=452390 00:41:55.169 17:57:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@41 -- # xtrace_disable 00:41:55.169 17:57:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:41:55.169 17:57:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/63 -t 5 -q 128 -w randrw -M 50 -o 8192 00:41:55.169 17:57:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@37 -- # gen_nvmf_target_json 00:41:55.169 17:57:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@560 -- # config=() 00:41:55.169 17:57:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@560 -- # local subsystem config 00:41:55.169 17:57:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:41:55.169 17:57:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:41:55.169 { 00:41:55.169 "params": { 00:41:55.169 "name": "Nvme$subsystem", 00:41:55.169 "trtype": "$TEST_TRANSPORT", 00:41:55.169 "traddr": "$NVMF_FIRST_TARGET_IP", 00:41:55.169 "adrfam": "ipv4", 00:41:55.169 "trsvcid": "$NVMF_PORT", 00:41:55.169 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:41:55.169 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:41:55.169 "hdgst": ${hdgst:-false}, 00:41:55.169 "ddgst": ${ddgst:-false} 00:41:55.169 }, 00:41:55.169 "method": "bdev_nvme_attach_controller" 00:41:55.169 } 00:41:55.169 EOF 00:41:55.169 )") 00:41:55.169 17:57:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@582 -- # cat 00:41:55.169 17:57:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@584 -- # jq . 00:41:55.169 [2024-12-06 17:57:36.910681] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.169 [2024-12-06 17:57:36.910726] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.169 17:57:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@585 -- # IFS=, 00:41:55.169 17:57:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:41:55.169 "params": { 00:41:55.169 "name": "Nvme1", 00:41:55.169 "trtype": "tcp", 00:41:55.169 "traddr": "10.0.0.2", 00:41:55.169 "adrfam": "ipv4", 00:41:55.169 "trsvcid": "4420", 00:41:55.169 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:41:55.169 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:41:55.169 "hdgst": false, 00:41:55.169 "ddgst": false 00:41:55.169 }, 00:41:55.169 "method": "bdev_nvme_attach_controller" 00:41:55.169 }' 00:41:55.169 [2024-12-06 17:57:36.918583] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.169 [2024-12-06 17:57:36.918605] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.169 [2024-12-06 17:57:36.926581] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.169 [2024-12-06 17:57:36.926602] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.169 [2024-12-06 17:57:36.934584] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.169 [2024-12-06 17:57:36.934605] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.169 [2024-12-06 17:57:36.942585] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.169 [2024-12-06 17:57:36.942606] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.169 [2024-12-06 17:57:36.948741] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:41:55.169 [2024-12-06 17:57:36.948805] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid452390 ] 00:41:55.169 [2024-12-06 17:57:36.950585] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.169 [2024-12-06 17:57:36.950621] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.169 [2024-12-06 17:57:36.958581] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.169 [2024-12-06 17:57:36.958600] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.169 [2024-12-06 17:57:36.966580] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.169 [2024-12-06 17:57:36.966600] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.169 [2024-12-06 17:57:36.974580] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.169 [2024-12-06 17:57:36.974599] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.169 [2024-12-06 17:57:36.982597] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.169 [2024-12-06 17:57:36.982617] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.169 [2024-12-06 17:57:36.990583] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.169 [2024-12-06 17:57:36.990608] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.169 [2024-12-06 17:57:36.998591] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.169 [2024-12-06 17:57:36.998611] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.169 [2024-12-06 17:57:37.006598] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.169 [2024-12-06 17:57:37.006631] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.014586] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.014609] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.020573] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:41:55.428 [2024-12-06 17:57:37.022582] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.022602] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.030623] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.030687] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.038597] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.038630] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.046586] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.046612] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.054582] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.054603] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.062581] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.062602] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.069158] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:41:55.428 [2024-12-06 17:57:37.070581] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.070602] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.078581] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.078601] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.086608] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.086655] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.094611] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.094671] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.102612] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.102661] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.110611] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.110660] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.118610] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.118660] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.126610] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.126659] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.134595] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.134623] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.142584] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.142606] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.150607] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.150673] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.158606] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.158640] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.166598] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.166633] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.174584] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.174606] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.182588] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.182612] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.190588] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.190627] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.198603] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.198627] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.206585] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.206607] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.214586] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.214608] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.222590] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.222613] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.230586] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.230609] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.238586] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.238609] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 [2024-12-06 17:57:37.246589] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.246613] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.428 Running I/O for 5 seconds... 00:41:55.428 [2024-12-06 17:57:37.254604] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.428 [2024-12-06 17:57:37.254643] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.686 [2024-12-06 17:57:37.272877] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.686 [2024-12-06 17:57:37.272907] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.686 [2024-12-06 17:57:37.288717] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.686 [2024-12-06 17:57:37.288760] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.686 [2024-12-06 17:57:37.300060] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.686 [2024-12-06 17:57:37.300087] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.686 [2024-12-06 17:57:37.315746] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.686 [2024-12-06 17:57:37.315774] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.686 [2024-12-06 17:57:37.325804] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.686 [2024-12-06 17:57:37.325831] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.686 [2024-12-06 17:57:37.340653] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.686 [2024-12-06 17:57:37.340700] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.686 [2024-12-06 17:57:37.357433] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.686 [2024-12-06 17:57:37.357460] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.686 [2024-12-06 17:57:37.367462] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.686 [2024-12-06 17:57:37.367490] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.686 [2024-12-06 17:57:37.379850] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.686 [2024-12-06 17:57:37.379879] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.686 [2024-12-06 17:57:37.390309] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.686 [2024-12-06 17:57:37.390335] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.686 [2024-12-06 17:57:37.401753] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.686 [2024-12-06 17:57:37.401782] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.686 [2024-12-06 17:57:37.413608] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.686 [2024-12-06 17:57:37.413636] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.686 [2024-12-06 17:57:37.429200] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.686 [2024-12-06 17:57:37.429228] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.686 [2024-12-06 17:57:37.443236] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.686 [2024-12-06 17:57:37.443265] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.686 [2024-12-06 17:57:37.453032] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.686 [2024-12-06 17:57:37.453059] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.686 [2024-12-06 17:57:37.465315] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.686 [2024-12-06 17:57:37.465341] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.686 [2024-12-06 17:57:37.480548] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.686 [2024-12-06 17:57:37.480576] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.686 [2024-12-06 17:57:37.490620] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.686 [2024-12-06 17:57:37.490648] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.686 [2024-12-06 17:57:37.503202] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.686 [2024-12-06 17:57:37.503229] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.686 [2024-12-06 17:57:37.513716] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.686 [2024-12-06 17:57:37.513744] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.943 [2024-12-06 17:57:37.527794] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.943 [2024-12-06 17:57:37.527823] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.944 [2024-12-06 17:57:37.537834] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.944 [2024-12-06 17:57:37.537863] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.944 [2024-12-06 17:57:37.551058] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.944 [2024-12-06 17:57:37.551086] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.944 [2024-12-06 17:57:37.562699] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.944 [2024-12-06 17:57:37.562736] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.944 [2024-12-06 17:57:37.573980] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.944 [2024-12-06 17:57:37.574028] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.944 [2024-12-06 17:57:37.585756] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.944 [2024-12-06 17:57:37.585784] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.944 [2024-12-06 17:57:37.596923] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.944 [2024-12-06 17:57:37.596951] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.944 [2024-12-06 17:57:37.613979] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.944 [2024-12-06 17:57:37.614019] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.944 [2024-12-06 17:57:37.624289] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.944 [2024-12-06 17:57:37.624316] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.944 [2024-12-06 17:57:37.636715] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.944 [2024-12-06 17:57:37.636743] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.944 [2024-12-06 17:57:37.651369] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.944 [2024-12-06 17:57:37.651397] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.944 [2024-12-06 17:57:37.661371] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.944 [2024-12-06 17:57:37.661399] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.944 [2024-12-06 17:57:37.673627] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.944 [2024-12-06 17:57:37.673654] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.944 [2024-12-06 17:57:37.687687] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.944 [2024-12-06 17:57:37.687715] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.944 [2024-12-06 17:57:37.697893] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.944 [2024-12-06 17:57:37.697921] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.944 [2024-12-06 17:57:37.710223] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.944 [2024-12-06 17:57:37.710250] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.944 [2024-12-06 17:57:37.721514] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.944 [2024-12-06 17:57:37.721540] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.944 [2024-12-06 17:57:37.732834] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.944 [2024-12-06 17:57:37.732863] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.944 [2024-12-06 17:57:37.746485] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.944 [2024-12-06 17:57:37.746515] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.944 [2024-12-06 17:57:37.757169] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.944 [2024-12-06 17:57:37.757196] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.944 [2024-12-06 17:57:37.771450] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.944 [2024-12-06 17:57:37.771477] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:55.944 [2024-12-06 17:57:37.781535] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:55.944 [2024-12-06 17:57:37.781564] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.202 [2024-12-06 17:57:37.794821] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.202 [2024-12-06 17:57:37.794848] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.202 [2024-12-06 17:57:37.806178] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.202 [2024-12-06 17:57:37.806204] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.202 [2024-12-06 17:57:37.818298] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.202 [2024-12-06 17:57:37.818325] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.202 [2024-12-06 17:57:37.830264] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.202 [2024-12-06 17:57:37.830292] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.202 [2024-12-06 17:57:37.841880] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.202 [2024-12-06 17:57:37.841909] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.202 [2024-12-06 17:57:37.853260] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.202 [2024-12-06 17:57:37.853286] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.202 [2024-12-06 17:57:37.867923] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.202 [2024-12-06 17:57:37.867951] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.202 [2024-12-06 17:57:37.878047] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.202 [2024-12-06 17:57:37.878075] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.202 [2024-12-06 17:57:37.890557] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.202 [2024-12-06 17:57:37.890583] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.202 [2024-12-06 17:57:37.902360] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.202 [2024-12-06 17:57:37.902403] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.202 [2024-12-06 17:57:37.913332] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.202 [2024-12-06 17:57:37.913377] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.202 [2024-12-06 17:57:37.928936] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.202 [2024-12-06 17:57:37.928980] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.202 [2024-12-06 17:57:37.942004] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.202 [2024-12-06 17:57:37.942032] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.202 [2024-12-06 17:57:37.952481] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.202 [2024-12-06 17:57:37.952508] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.202 [2024-12-06 17:57:37.964606] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.202 [2024-12-06 17:57:37.964633] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.202 [2024-12-06 17:57:37.980983] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.202 [2024-12-06 17:57:37.981009] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.202 [2024-12-06 17:57:37.996310] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.202 [2024-12-06 17:57:37.996340] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.202 [2024-12-06 17:57:38.006106] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.202 [2024-12-06 17:57:38.006135] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.202 [2024-12-06 17:57:38.019072] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.202 [2024-12-06 17:57:38.019099] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.202 [2024-12-06 17:57:38.030412] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.202 [2024-12-06 17:57:38.030440] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.460 [2024-12-06 17:57:38.042110] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.460 [2024-12-06 17:57:38.042139] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.460 [2024-12-06 17:57:38.053159] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.460 [2024-12-06 17:57:38.053187] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.460 [2024-12-06 17:57:38.066625] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.460 [2024-12-06 17:57:38.066655] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.460 [2024-12-06 17:57:38.076214] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.460 [2024-12-06 17:57:38.076243] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.460 [2024-12-06 17:57:38.088689] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.460 [2024-12-06 17:57:38.088717] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.460 [2024-12-06 17:57:38.102398] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.460 [2024-12-06 17:57:38.102427] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.460 [2024-12-06 17:57:38.112902] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.460 [2024-12-06 17:57:38.112930] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.460 [2024-12-06 17:57:38.125798] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.460 [2024-12-06 17:57:38.125827] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.460 [2024-12-06 17:57:38.138606] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.460 [2024-12-06 17:57:38.138634] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.460 [2024-12-06 17:57:38.148601] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.460 [2024-12-06 17:57:38.148630] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.460 [2024-12-06 17:57:38.161398] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.460 [2024-12-06 17:57:38.161426] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.460 [2024-12-06 17:57:38.172899] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.460 [2024-12-06 17:57:38.172929] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.460 [2024-12-06 17:57:38.186752] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.460 [2024-12-06 17:57:38.186782] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.460 [2024-12-06 17:57:38.196892] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.460 [2024-12-06 17:57:38.196920] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.460 [2024-12-06 17:57:38.212507] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.460 [2024-12-06 17:57:38.212537] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.460 [2024-12-06 17:57:38.222838] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.460 [2024-12-06 17:57:38.222866] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.460 [2024-12-06 17:57:38.235388] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.460 [2024-12-06 17:57:38.235417] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.460 [2024-12-06 17:57:38.247219] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.460 [2024-12-06 17:57:38.247247] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.460 11002.00 IOPS, 85.95 MiB/s [2024-12-06T16:57:38.299Z] [2024-12-06 17:57:38.258465] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.460 [2024-12-06 17:57:38.258501] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.460 [2024-12-06 17:57:38.269701] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.460 [2024-12-06 17:57:38.269743] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.460 [2024-12-06 17:57:38.281257] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.460 [2024-12-06 17:57:38.281284] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.461 [2024-12-06 17:57:38.295547] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.461 [2024-12-06 17:57:38.295575] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.719 [2024-12-06 17:57:38.305705] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.719 [2024-12-06 17:57:38.305733] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.719 [2024-12-06 17:57:38.320334] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.719 [2024-12-06 17:57:38.320361] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.719 [2024-12-06 17:57:38.336750] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.719 [2024-12-06 17:57:38.336780] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.719 [2024-12-06 17:57:38.351197] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.719 [2024-12-06 17:57:38.351226] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.719 [2024-12-06 17:57:38.361140] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.719 [2024-12-06 17:57:38.361168] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.719 [2024-12-06 17:57:38.375296] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.719 [2024-12-06 17:57:38.375339] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.719 [2024-12-06 17:57:38.386010] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.719 [2024-12-06 17:57:38.386037] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.719 [2024-12-06 17:57:38.398215] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.719 [2024-12-06 17:57:38.398243] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.719 [2024-12-06 17:57:38.409509] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.719 [2024-12-06 17:57:38.409537] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.719 [2024-12-06 17:57:38.422987] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.719 [2024-12-06 17:57:38.423016] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.719 [2024-12-06 17:57:38.432722] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.719 [2024-12-06 17:57:38.432751] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.719 [2024-12-06 17:57:38.445267] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.719 [2024-12-06 17:57:38.445296] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.719 [2024-12-06 17:57:38.460659] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.719 [2024-12-06 17:57:38.460698] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.719 [2024-12-06 17:57:38.477099] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.719 [2024-12-06 17:57:38.477128] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.719 [2024-12-06 17:57:38.487761] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.719 [2024-12-06 17:57:38.487790] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.719 [2024-12-06 17:57:38.500773] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.719 [2024-12-06 17:57:38.500809] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.719 [2024-12-06 17:57:38.515840] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.719 [2024-12-06 17:57:38.515869] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.719 [2024-12-06 17:57:38.525327] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.719 [2024-12-06 17:57:38.525354] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.719 [2024-12-06 17:57:38.540433] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.719 [2024-12-06 17:57:38.540460] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.719 [2024-12-06 17:57:38.556420] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.719 [2024-12-06 17:57:38.556449] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.977 [2024-12-06 17:57:38.567093] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.977 [2024-12-06 17:57:38.567119] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.977 [2024-12-06 17:57:38.580057] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.977 [2024-12-06 17:57:38.580087] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.977 [2024-12-06 17:57:38.594220] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.977 [2024-12-06 17:57:38.594249] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.977 [2024-12-06 17:57:38.604478] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.977 [2024-12-06 17:57:38.604506] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.977 [2024-12-06 17:57:38.617648] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.977 [2024-12-06 17:57:38.617686] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.977 [2024-12-06 17:57:38.630959] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.978 [2024-12-06 17:57:38.631003] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.978 [2024-12-06 17:57:38.641024] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.978 [2024-12-06 17:57:38.641052] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.978 [2024-12-06 17:57:38.656776] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.978 [2024-12-06 17:57:38.656814] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.978 [2024-12-06 17:57:38.672566] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.978 [2024-12-06 17:57:38.672595] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.978 [2024-12-06 17:57:38.683197] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.978 [2024-12-06 17:57:38.683225] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.978 [2024-12-06 17:57:38.695923] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.978 [2024-12-06 17:57:38.695951] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.978 [2024-12-06 17:57:38.713075] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.978 [2024-12-06 17:57:38.713102] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.978 [2024-12-06 17:57:38.729132] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.978 [2024-12-06 17:57:38.729161] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.978 [2024-12-06 17:57:38.744800] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.978 [2024-12-06 17:57:38.744828] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.978 [2024-12-06 17:57:38.760896] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.978 [2024-12-06 17:57:38.760948] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.978 [2024-12-06 17:57:38.776198] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.978 [2024-12-06 17:57:38.776242] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.978 [2024-12-06 17:57:38.786912] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.978 [2024-12-06 17:57:38.786941] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.978 [2024-12-06 17:57:38.799625] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.978 [2024-12-06 17:57:38.799654] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:56.978 [2024-12-06 17:57:38.811161] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:56.978 [2024-12-06 17:57:38.811189] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.237 [2024-12-06 17:57:38.822255] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.237 [2024-12-06 17:57:38.822283] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.237 [2024-12-06 17:57:38.832887] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.237 [2024-12-06 17:57:38.832915] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.237 [2024-12-06 17:57:38.845542] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.237 [2024-12-06 17:57:38.845570] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.237 [2024-12-06 17:57:38.858826] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.237 [2024-12-06 17:57:38.858855] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.237 [2024-12-06 17:57:38.868859] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.237 [2024-12-06 17:57:38.868888] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.237 [2024-12-06 17:57:38.881766] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.237 [2024-12-06 17:57:38.881795] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.237 [2024-12-06 17:57:38.893395] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.237 [2024-12-06 17:57:38.893424] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.237 [2024-12-06 17:57:38.904882] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.237 [2024-12-06 17:57:38.904911] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.237 [2024-12-06 17:57:38.919644] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.237 [2024-12-06 17:57:38.919681] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.237 [2024-12-06 17:57:38.929868] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.237 [2024-12-06 17:57:38.929897] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.237 [2024-12-06 17:57:38.942485] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.237 [2024-12-06 17:57:38.942531] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.237 [2024-12-06 17:57:38.954631] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.237 [2024-12-06 17:57:38.954659] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.237 [2024-12-06 17:57:38.966377] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.237 [2024-12-06 17:57:38.966405] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.237 [2024-12-06 17:57:38.978211] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.237 [2024-12-06 17:57:38.978239] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.237 [2024-12-06 17:57:38.989346] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.237 [2024-12-06 17:57:38.989382] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.237 [2024-12-06 17:57:39.002653] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.237 [2024-12-06 17:57:39.002705] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.237 [2024-12-06 17:57:39.012356] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.237 [2024-12-06 17:57:39.012383] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.237 [2024-12-06 17:57:39.024978] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.237 [2024-12-06 17:57:39.025018] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.237 [2024-12-06 17:57:39.040925] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.237 [2024-12-06 17:57:39.040967] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.237 [2024-12-06 17:57:39.056043] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.237 [2024-12-06 17:57:39.056072] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.237 [2024-12-06 17:57:39.065607] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.237 [2024-12-06 17:57:39.065634] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.495 [2024-12-06 17:57:39.078014] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.495 [2024-12-06 17:57:39.078043] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.495 [2024-12-06 17:57:39.089092] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.495 [2024-12-06 17:57:39.089118] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.495 [2024-12-06 17:57:39.104415] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.495 [2024-12-06 17:57:39.104443] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.495 [2024-12-06 17:57:39.114557] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.495 [2024-12-06 17:57:39.114587] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.495 [2024-12-06 17:57:39.127183] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.495 [2024-12-06 17:57:39.127210] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.495 [2024-12-06 17:57:39.138376] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.495 [2024-12-06 17:57:39.138402] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.495 [2024-12-06 17:57:39.149907] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.495 [2024-12-06 17:57:39.149935] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.495 [2024-12-06 17:57:39.161250] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.496 [2024-12-06 17:57:39.161277] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.496 [2024-12-06 17:57:39.174699] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.496 [2024-12-06 17:57:39.174741] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.496 [2024-12-06 17:57:39.184582] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.496 [2024-12-06 17:57:39.184609] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.496 [2024-12-06 17:57:39.197449] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.496 [2024-12-06 17:57:39.197475] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.496 [2024-12-06 17:57:39.212138] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.496 [2024-12-06 17:57:39.212169] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.496 [2024-12-06 17:57:39.222192] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.496 [2024-12-06 17:57:39.222228] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.496 [2024-12-06 17:57:39.234367] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.496 [2024-12-06 17:57:39.234397] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.496 [2024-12-06 17:57:39.244583] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.496 [2024-12-06 17:57:39.244611] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.496 [2024-12-06 17:57:39.260086] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.496 [2024-12-06 17:57:39.260113] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.496 11026.50 IOPS, 86.14 MiB/s [2024-12-06T16:57:39.335Z] [2024-12-06 17:57:39.276812] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.496 [2024-12-06 17:57:39.276840] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.496 [2024-12-06 17:57:39.287534] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.496 [2024-12-06 17:57:39.287561] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.496 [2024-12-06 17:57:39.304297] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.496 [2024-12-06 17:57:39.304326] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.496 [2024-12-06 17:57:39.314782] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.496 [2024-12-06 17:57:39.314811] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.496 [2024-12-06 17:57:39.327606] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.496 [2024-12-06 17:57:39.327634] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.753 [2024-12-06 17:57:39.338966] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.753 [2024-12-06 17:57:39.338993] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.753 [2024-12-06 17:57:39.350178] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.753 [2024-12-06 17:57:39.350204] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.753 [2024-12-06 17:57:39.360915] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.753 [2024-12-06 17:57:39.360942] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.753 [2024-12-06 17:57:39.375425] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.753 [2024-12-06 17:57:39.375451] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.753 [2024-12-06 17:57:39.385969] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.753 [2024-12-06 17:57:39.385996] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.753 [2024-12-06 17:57:39.398160] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.753 [2024-12-06 17:57:39.398186] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.753 [2024-12-06 17:57:39.409526] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.753 [2024-12-06 17:57:39.409551] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.753 [2024-12-06 17:57:39.422224] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.753 [2024-12-06 17:57:39.422252] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.753 [2024-12-06 17:57:39.432130] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.753 [2024-12-06 17:57:39.432157] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.753 [2024-12-06 17:57:39.444323] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.753 [2024-12-06 17:57:39.444349] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.753 [2024-12-06 17:57:39.458795] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.753 [2024-12-06 17:57:39.458822] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.753 [2024-12-06 17:57:39.468682] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.753 [2024-12-06 17:57:39.468710] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.753 [2024-12-06 17:57:39.484060] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.753 [2024-12-06 17:57:39.484086] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.753 [2024-12-06 17:57:39.494534] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.753 [2024-12-06 17:57:39.494563] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.753 [2024-12-06 17:57:39.506998] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.753 [2024-12-06 17:57:39.507024] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.753 [2024-12-06 17:57:39.518179] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.753 [2024-12-06 17:57:39.518206] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.753 [2024-12-06 17:57:39.529696] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.753 [2024-12-06 17:57:39.529743] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.753 [2024-12-06 17:57:39.541339] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.753 [2024-12-06 17:57:39.541381] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.753 [2024-12-06 17:57:39.556212] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.753 [2024-12-06 17:57:39.556240] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.753 [2024-12-06 17:57:39.572458] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.753 [2024-12-06 17:57:39.572487] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:57.753 [2024-12-06 17:57:39.589279] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:57.753 [2024-12-06 17:57:39.589309] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.017 [2024-12-06 17:57:39.599618] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.017 [2024-12-06 17:57:39.599644] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.017 [2024-12-06 17:57:39.612307] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.017 [2024-12-06 17:57:39.612335] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.017 [2024-12-06 17:57:39.629521] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.017 [2024-12-06 17:57:39.629548] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.017 [2024-12-06 17:57:39.639539] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.017 [2024-12-06 17:57:39.639581] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.017 [2024-12-06 17:57:39.651964] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.017 [2024-12-06 17:57:39.651992] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.018 [2024-12-06 17:57:39.668883] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.018 [2024-12-06 17:57:39.668911] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.018 [2024-12-06 17:57:39.684311] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.018 [2024-12-06 17:57:39.684338] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.018 [2024-12-06 17:57:39.694151] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.018 [2024-12-06 17:57:39.694179] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.018 [2024-12-06 17:57:39.706849] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.018 [2024-12-06 17:57:39.706878] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.018 [2024-12-06 17:57:39.718630] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.018 [2024-12-06 17:57:39.718680] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.018 [2024-12-06 17:57:39.730787] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.018 [2024-12-06 17:57:39.730815] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.018 [2024-12-06 17:57:39.742203] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.018 [2024-12-06 17:57:39.742229] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.018 [2024-12-06 17:57:39.754444] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.018 [2024-12-06 17:57:39.754471] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.018 [2024-12-06 17:57:39.765792] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.018 [2024-12-06 17:57:39.765819] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.018 [2024-12-06 17:57:39.779388] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.018 [2024-12-06 17:57:39.779430] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.018 [2024-12-06 17:57:39.789641] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.018 [2024-12-06 17:57:39.789679] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.018 [2024-12-06 17:57:39.802102] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.018 [2024-12-06 17:57:39.802129] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.018 [2024-12-06 17:57:39.813371] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.018 [2024-12-06 17:57:39.813398] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.018 [2024-12-06 17:57:39.824705] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.018 [2024-12-06 17:57:39.824739] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.018 [2024-12-06 17:57:39.838355] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.018 [2024-12-06 17:57:39.838382] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.018 [2024-12-06 17:57:39.848320] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.018 [2024-12-06 17:57:39.848347] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.277 [2024-12-06 17:57:39.860912] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.277 [2024-12-06 17:57:39.860939] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.277 [2024-12-06 17:57:39.871957] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.277 [2024-12-06 17:57:39.871984] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.277 [2024-12-06 17:57:39.883910] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.277 [2024-12-06 17:57:39.883937] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.277 [2024-12-06 17:57:39.900065] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.277 [2024-12-06 17:57:39.900091] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.277 [2024-12-06 17:57:39.918936] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.277 [2024-12-06 17:57:39.918977] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.277 [2024-12-06 17:57:39.930006] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.277 [2024-12-06 17:57:39.930038] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.277 [2024-12-06 17:57:39.941590] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.277 [2024-12-06 17:57:39.941618] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.277 [2024-12-06 17:57:39.953002] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.277 [2024-12-06 17:57:39.953042] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.277 [2024-12-06 17:57:39.968931] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.277 [2024-12-06 17:57:39.968973] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.277 [2024-12-06 17:57:39.985169] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.277 [2024-12-06 17:57:39.985197] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.277 [2024-12-06 17:57:40.000154] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.277 [2024-12-06 17:57:40.000181] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.277 [2024-12-06 17:57:40.011880] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.277 [2024-12-06 17:57:40.011911] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.277 [2024-12-06 17:57:40.024968] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.277 [2024-12-06 17:57:40.024997] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.277 [2024-12-06 17:57:40.042026] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.277 [2024-12-06 17:57:40.042078] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.277 [2024-12-06 17:57:40.054204] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.277 [2024-12-06 17:57:40.054237] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.277 [2024-12-06 17:57:40.069645] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.277 [2024-12-06 17:57:40.069699] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.277 [2024-12-06 17:57:40.084034] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.277 [2024-12-06 17:57:40.084065] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.277 [2024-12-06 17:57:40.098859] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.277 [2024-12-06 17:57:40.098893] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.277 [2024-12-06 17:57:40.110762] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.277 [2024-12-06 17:57:40.110808] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.535 [2024-12-06 17:57:40.122402] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.535 [2024-12-06 17:57:40.122428] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.535 [2024-12-06 17:57:40.134016] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.535 [2024-12-06 17:57:40.134044] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.535 [2024-12-06 17:57:40.145779] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.535 [2024-12-06 17:57:40.145808] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.535 [2024-12-06 17:57:40.157295] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.535 [2024-12-06 17:57:40.157322] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.535 [2024-12-06 17:57:40.171052] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.535 [2024-12-06 17:57:40.171080] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.535 [2024-12-06 17:57:40.181281] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.535 [2024-12-06 17:57:40.181316] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.535 [2024-12-06 17:57:40.197062] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.535 [2024-12-06 17:57:40.197089] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.535 [2024-12-06 17:57:40.212512] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.535 [2024-12-06 17:57:40.212540] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.535 [2024-12-06 17:57:40.227999] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.535 [2024-12-06 17:57:40.228043] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.535 [2024-12-06 17:57:40.238563] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.535 [2024-12-06 17:57:40.238604] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.535 [2024-12-06 17:57:40.251681] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.535 [2024-12-06 17:57:40.251721] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.535 10970.00 IOPS, 85.70 MiB/s [2024-12-06T16:57:40.374Z] [2024-12-06 17:57:40.263598] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.535 [2024-12-06 17:57:40.263623] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.535 [2024-12-06 17:57:40.275387] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.535 [2024-12-06 17:57:40.275413] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.535 [2024-12-06 17:57:40.287149] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.535 [2024-12-06 17:57:40.287174] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.535 [2024-12-06 17:57:40.299622] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.535 [2024-12-06 17:57:40.299649] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.535 [2024-12-06 17:57:40.316746] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.535 [2024-12-06 17:57:40.316773] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.535 [2024-12-06 17:57:40.327411] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.535 [2024-12-06 17:57:40.327438] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.535 [2024-12-06 17:57:40.340446] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.535 [2024-12-06 17:57:40.340475] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.535 [2024-12-06 17:57:40.357348] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.535 [2024-12-06 17:57:40.357378] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.535 [2024-12-06 17:57:40.367712] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.535 [2024-12-06 17:57:40.367755] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.793 [2024-12-06 17:57:40.383840] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.793 [2024-12-06 17:57:40.383868] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.793 [2024-12-06 17:57:40.401277] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.793 [2024-12-06 17:57:40.401304] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.793 [2024-12-06 17:57:40.411616] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.793 [2024-12-06 17:57:40.411659] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.793 [2024-12-06 17:57:40.428478] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.793 [2024-12-06 17:57:40.428503] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.793 [2024-12-06 17:57:40.445190] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.793 [2024-12-06 17:57:40.445226] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.793 [2024-12-06 17:57:40.460474] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.793 [2024-12-06 17:57:40.460501] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.793 [2024-12-06 17:57:40.471286] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.793 [2024-12-06 17:57:40.471312] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.793 [2024-12-06 17:57:40.484002] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.793 [2024-12-06 17:57:40.484028] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.793 [2024-12-06 17:57:40.499170] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.793 [2024-12-06 17:57:40.499199] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.793 [2024-12-06 17:57:40.509274] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.793 [2024-12-06 17:57:40.509315] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.793 [2024-12-06 17:57:40.524014] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.793 [2024-12-06 17:57:40.524040] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.793 [2024-12-06 17:57:40.534427] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.793 [2024-12-06 17:57:40.534454] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.793 [2024-12-06 17:57:40.547577] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.793 [2024-12-06 17:57:40.547604] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.793 [2024-12-06 17:57:40.559417] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.793 [2024-12-06 17:57:40.559444] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.793 [2024-12-06 17:57:40.576256] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.793 [2024-12-06 17:57:40.576282] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.793 [2024-12-06 17:57:40.587031] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.793 [2024-12-06 17:57:40.587058] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.793 [2024-12-06 17:57:40.599235] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.793 [2024-12-06 17:57:40.599260] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.793 [2024-12-06 17:57:40.610148] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.793 [2024-12-06 17:57:40.610172] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:58.793 [2024-12-06 17:57:40.621421] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:58.793 [2024-12-06 17:57:40.621447] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.051 [2024-12-06 17:57:40.635175] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.051 [2024-12-06 17:57:40.635201] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.051 [2024-12-06 17:57:40.645864] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.051 [2024-12-06 17:57:40.645891] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.051 [2024-12-06 17:57:40.658732] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.051 [2024-12-06 17:57:40.658757] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.051 [2024-12-06 17:57:40.670238] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.051 [2024-12-06 17:57:40.670264] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.051 [2024-12-06 17:57:40.681304] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.051 [2024-12-06 17:57:40.681330] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.051 [2024-12-06 17:57:40.696294] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.051 [2024-12-06 17:57:40.696321] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.051 [2024-12-06 17:57:40.712771] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.051 [2024-12-06 17:57:40.712798] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.051 [2024-12-06 17:57:40.728346] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.051 [2024-12-06 17:57:40.728373] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.051 [2024-12-06 17:57:40.738362] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.051 [2024-12-06 17:57:40.738387] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.051 [2024-12-06 17:57:40.751274] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.051 [2024-12-06 17:57:40.751299] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.051 [2024-12-06 17:57:40.762535] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.051 [2024-12-06 17:57:40.762575] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.051 [2024-12-06 17:57:40.774212] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.051 [2024-12-06 17:57:40.774236] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.051 [2024-12-06 17:57:40.785672] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.051 [2024-12-06 17:57:40.785697] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.051 [2024-12-06 17:57:40.796821] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.051 [2024-12-06 17:57:40.796847] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.051 [2024-12-06 17:57:40.813274] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.051 [2024-12-06 17:57:40.813299] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.051 [2024-12-06 17:57:40.828898] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.051 [2024-12-06 17:57:40.828926] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.051 [2024-12-06 17:57:40.844478] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.051 [2024-12-06 17:57:40.844504] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.051 [2024-12-06 17:57:40.860890] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.051 [2024-12-06 17:57:40.860917] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.051 [2024-12-06 17:57:40.876798] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.051 [2024-12-06 17:57:40.876826] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.309 [2024-12-06 17:57:40.893115] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.309 [2024-12-06 17:57:40.893156] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.309 [2024-12-06 17:57:40.908755] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.309 [2024-12-06 17:57:40.908782] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.309 [2024-12-06 17:57:40.924430] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.309 [2024-12-06 17:57:40.924457] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.309 [2024-12-06 17:57:40.934853] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.309 [2024-12-06 17:57:40.934880] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.309 [2024-12-06 17:57:40.947890] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.309 [2024-12-06 17:57:40.947917] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.309 [2024-12-06 17:57:40.959918] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.309 [2024-12-06 17:57:40.959969] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.309 [2024-12-06 17:57:40.976274] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.309 [2024-12-06 17:57:40.976314] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.309 [2024-12-06 17:57:40.986934] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.309 [2024-12-06 17:57:40.986976] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.309 [2024-12-06 17:57:40.999750] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.309 [2024-12-06 17:57:40.999778] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.309 [2024-12-06 17:57:41.016416] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.309 [2024-12-06 17:57:41.016440] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.309 [2024-12-06 17:57:41.033080] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.309 [2024-12-06 17:57:41.033121] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.309 [2024-12-06 17:57:41.048035] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.309 [2024-12-06 17:57:41.048062] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.309 [2024-12-06 17:57:41.058479] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.309 [2024-12-06 17:57:41.058503] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.309 [2024-12-06 17:57:41.070805] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.309 [2024-12-06 17:57:41.070831] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.310 [2024-12-06 17:57:41.082168] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.310 [2024-12-06 17:57:41.082193] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.310 [2024-12-06 17:57:41.094262] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.310 [2024-12-06 17:57:41.094289] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.310 [2024-12-06 17:57:41.106179] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.310 [2024-12-06 17:57:41.106218] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.310 [2024-12-06 17:57:41.118415] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.310 [2024-12-06 17:57:41.118441] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.310 [2024-12-06 17:57:41.130335] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.310 [2024-12-06 17:57:41.130360] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.310 [2024-12-06 17:57:41.141467] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.310 [2024-12-06 17:57:41.141492] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.567 [2024-12-06 17:57:41.155524] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.567 [2024-12-06 17:57:41.155565] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.567 [2024-12-06 17:57:41.165168] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.567 [2024-12-06 17:57:41.165193] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.567 [2024-12-06 17:57:41.180243] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.567 [2024-12-06 17:57:41.180283] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.567 [2024-12-06 17:57:41.190216] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.567 [2024-12-06 17:57:41.190240] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.567 [2024-12-06 17:57:41.202790] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.567 [2024-12-06 17:57:41.202815] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.567 [2024-12-06 17:57:41.214209] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.567 [2024-12-06 17:57:41.214233] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.567 [2024-12-06 17:57:41.225427] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.567 [2024-12-06 17:57:41.225452] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.567 [2024-12-06 17:57:41.236719] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.567 [2024-12-06 17:57:41.236745] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.567 [2024-12-06 17:57:41.252891] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.568 [2024-12-06 17:57:41.252916] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.568 [2024-12-06 17:57:41.263412] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.568 [2024-12-06 17:57:41.263437] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.568 10953.50 IOPS, 85.57 MiB/s [2024-12-06T16:57:41.407Z] [2024-12-06 17:57:41.275839] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.568 [2024-12-06 17:57:41.275866] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.568 [2024-12-06 17:57:41.293563] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.568 [2024-12-06 17:57:41.293587] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.568 [2024-12-06 17:57:41.303857] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.568 [2024-12-06 17:57:41.303884] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.568 [2024-12-06 17:57:41.320637] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.568 [2024-12-06 17:57:41.320685] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.568 [2024-12-06 17:57:41.337604] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.568 [2024-12-06 17:57:41.337629] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.568 [2024-12-06 17:57:41.348370] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.568 [2024-12-06 17:57:41.348396] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.568 [2024-12-06 17:57:41.361224] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.568 [2024-12-06 17:57:41.361251] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.568 [2024-12-06 17:57:41.374919] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.568 [2024-12-06 17:57:41.374947] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.568 [2024-12-06 17:57:41.385594] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.568 [2024-12-06 17:57:41.385619] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.568 [2024-12-06 17:57:41.400071] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.568 [2024-12-06 17:57:41.400096] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.825 [2024-12-06 17:57:41.410917] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.825 [2024-12-06 17:57:41.410960] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.825 [2024-12-06 17:57:41.423958] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.825 [2024-12-06 17:57:41.423991] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.825 [2024-12-06 17:57:41.439979] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.825 [2024-12-06 17:57:41.440005] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.825 [2024-12-06 17:57:41.450340] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.825 [2024-12-06 17:57:41.450365] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.825 [2024-12-06 17:57:41.463133] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.825 [2024-12-06 17:57:41.463158] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.825 [2024-12-06 17:57:41.474283] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.825 [2024-12-06 17:57:41.474309] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.825 [2024-12-06 17:57:41.486334] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.826 [2024-12-06 17:57:41.486358] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.826 [2024-12-06 17:57:41.498185] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.826 [2024-12-06 17:57:41.498212] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.826 [2024-12-06 17:57:41.509903] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.826 [2024-12-06 17:57:41.509932] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.826 [2024-12-06 17:57:41.521462] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.826 [2024-12-06 17:57:41.521489] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.826 [2024-12-06 17:57:41.533552] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.826 [2024-12-06 17:57:41.533577] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.826 [2024-12-06 17:57:41.545088] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.826 [2024-12-06 17:57:41.545114] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.826 [2024-12-06 17:57:41.559919] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.826 [2024-12-06 17:57:41.559961] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.826 [2024-12-06 17:57:41.570021] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.826 [2024-12-06 17:57:41.570060] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.826 [2024-12-06 17:57:41.583040] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.826 [2024-12-06 17:57:41.583064] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.826 [2024-12-06 17:57:41.594923] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.826 [2024-12-06 17:57:41.594971] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.826 [2024-12-06 17:57:41.606976] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.826 [2024-12-06 17:57:41.607022] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.826 [2024-12-06 17:57:41.617864] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.826 [2024-12-06 17:57:41.617890] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.826 [2024-12-06 17:57:41.629292] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.826 [2024-12-06 17:57:41.629317] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.826 [2024-12-06 17:57:41.643285] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.826 [2024-12-06 17:57:41.643312] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:41:59.826 [2024-12-06 17:57:41.653211] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:41:59.826 [2024-12-06 17:57:41.653244] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.084 [2024-12-06 17:57:41.668250] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.084 [2024-12-06 17:57:41.668275] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.084 [2024-12-06 17:57:41.678489] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.084 [2024-12-06 17:57:41.678517] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.084 [2024-12-06 17:57:41.691555] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.084 [2024-12-06 17:57:41.691580] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.084 [2024-12-06 17:57:41.703684] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.084 [2024-12-06 17:57:41.703724] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.084 [2024-12-06 17:57:41.719739] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.084 [2024-12-06 17:57:41.719768] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.084 [2024-12-06 17:57:41.729977] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.084 [2024-12-06 17:57:41.730004] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.084 [2024-12-06 17:57:41.743105] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.084 [2024-12-06 17:57:41.743145] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.084 [2024-12-06 17:57:41.754727] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.084 [2024-12-06 17:57:41.754756] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.084 [2024-12-06 17:57:41.766565] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.084 [2024-12-06 17:57:41.766590] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.084 [2024-12-06 17:57:41.777966] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.084 [2024-12-06 17:57:41.777991] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.084 [2024-12-06 17:57:41.789430] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.084 [2024-12-06 17:57:41.789457] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.084 [2024-12-06 17:57:41.801369] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.084 [2024-12-06 17:57:41.801394] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.084 [2024-12-06 17:57:41.817216] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.084 [2024-12-06 17:57:41.817242] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.084 [2024-12-06 17:57:41.832603] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.084 [2024-12-06 17:57:41.832630] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.084 [2024-12-06 17:57:41.843315] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.084 [2024-12-06 17:57:41.843339] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.084 [2024-12-06 17:57:41.855587] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.084 [2024-12-06 17:57:41.855612] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.084 [2024-12-06 17:57:41.866940] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.084 [2024-12-06 17:57:41.866979] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.084 [2024-12-06 17:57:41.879189] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.084 [2024-12-06 17:57:41.879213] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.084 [2024-12-06 17:57:41.890770] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.084 [2024-12-06 17:57:41.890805] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.084 [2024-12-06 17:57:41.902474] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.084 [2024-12-06 17:57:41.902498] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.084 [2024-12-06 17:57:41.914144] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.084 [2024-12-06 17:57:41.914168] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.342 [2024-12-06 17:57:41.926020] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.342 [2024-12-06 17:57:41.926060] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.342 [2024-12-06 17:57:41.937906] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.342 [2024-12-06 17:57:41.937934] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.342 [2024-12-06 17:57:41.949553] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.342 [2024-12-06 17:57:41.949592] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.342 [2024-12-06 17:57:41.962445] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.342 [2024-12-06 17:57:41.962472] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.342 [2024-12-06 17:57:41.972984] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.342 [2024-12-06 17:57:41.973024] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.342 [2024-12-06 17:57:41.988254] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.342 [2024-12-06 17:57:41.988279] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.342 [2024-12-06 17:57:41.998688] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.342 [2024-12-06 17:57:41.998716] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.342 [2024-12-06 17:57:42.011250] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.342 [2024-12-06 17:57:42.011275] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.342 [2024-12-06 17:57:42.023261] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.342 [2024-12-06 17:57:42.023301] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.342 [2024-12-06 17:57:42.034722] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.343 [2024-12-06 17:57:42.034748] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.343 [2024-12-06 17:57:42.045566] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.343 [2024-12-06 17:57:42.045606] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.343 [2024-12-06 17:57:42.059254] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.343 [2024-12-06 17:57:42.059282] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.343 [2024-12-06 17:57:42.079354] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.343 [2024-12-06 17:57:42.079381] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.343 [2024-12-06 17:57:42.090754] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.343 [2024-12-06 17:57:42.090791] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.343 [2024-12-06 17:57:42.102278] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.343 [2024-12-06 17:57:42.102318] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.343 [2024-12-06 17:57:42.113298] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.343 [2024-12-06 17:57:42.113323] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.343 [2024-12-06 17:57:42.128346] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.343 [2024-12-06 17:57:42.128394] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.343 [2024-12-06 17:57:42.138626] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.343 [2024-12-06 17:57:42.138672] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.343 [2024-12-06 17:57:42.151454] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.343 [2024-12-06 17:57:42.151480] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.343 [2024-12-06 17:57:42.162868] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.343 [2024-12-06 17:57:42.162895] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.343 [2024-12-06 17:57:42.174323] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.343 [2024-12-06 17:57:42.174348] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.601 [2024-12-06 17:57:42.185980] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.601 [2024-12-06 17:57:42.186006] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.601 [2024-12-06 17:57:42.198022] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.601 [2024-12-06 17:57:42.198048] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.601 [2024-12-06 17:57:42.209266] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.601 [2024-12-06 17:57:42.209290] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.601 [2024-12-06 17:57:42.221022] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.601 [2024-12-06 17:57:42.221046] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.601 [2024-12-06 17:57:42.235139] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.601 [2024-12-06 17:57:42.235165] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.601 [2024-12-06 17:57:42.245583] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.601 [2024-12-06 17:57:42.245609] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.601 [2024-12-06 17:57:42.260661] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.601 [2024-12-06 17:57:42.260708] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.601 10938.60 IOPS, 85.46 MiB/s [2024-12-06T16:57:42.440Z] [2024-12-06 17:57:42.273985] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.601 [2024-12-06 17:57:42.274013] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.601 [2024-12-06 17:57:42.278589] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.601 [2024-12-06 17:57:42.278613] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.601 00:42:00.601 Latency(us) 00:42:00.601 [2024-12-06T16:57:42.440Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:42:00.601 Job: Nvme1n1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 128, IO size: 8192) 00:42:00.601 Nvme1n1 : 5.01 10939.88 85.47 0.00 0.00 11683.24 3325.35 21359.88 00:42:00.601 [2024-12-06T16:57:42.440Z] =================================================================================================================== 00:42:00.601 [2024-12-06T16:57:42.440Z] Total : 10939.88 85.47 0.00 0.00 11683.24 3325.35 21359.88 00:42:00.601 [2024-12-06 17:57:42.286585] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.601 [2024-12-06 17:57:42.286608] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.601 [2024-12-06 17:57:42.294594] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.601 [2024-12-06 17:57:42.294620] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.602 [2024-12-06 17:57:42.302650] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.602 [2024-12-06 17:57:42.302709] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.602 [2024-12-06 17:57:42.310652] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.602 [2024-12-06 17:57:42.310711] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.602 [2024-12-06 17:57:42.318640] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.602 [2024-12-06 17:57:42.318694] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.602 [2024-12-06 17:57:42.326639] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.602 [2024-12-06 17:57:42.326693] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.602 [2024-12-06 17:57:42.334637] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.602 [2024-12-06 17:57:42.334690] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.602 [2024-12-06 17:57:42.342636] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.602 [2024-12-06 17:57:42.342691] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.602 [2024-12-06 17:57:42.350639] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.602 [2024-12-06 17:57:42.350693] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.602 [2024-12-06 17:57:42.358638] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.602 [2024-12-06 17:57:42.358694] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.602 [2024-12-06 17:57:42.366639] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.602 [2024-12-06 17:57:42.366690] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.602 [2024-12-06 17:57:42.374643] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.602 [2024-12-06 17:57:42.374696] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.602 [2024-12-06 17:57:42.382642] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.602 [2024-12-06 17:57:42.382698] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.602 [2024-12-06 17:57:42.390643] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.602 [2024-12-06 17:57:42.390696] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.602 [2024-12-06 17:57:42.398639] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.602 [2024-12-06 17:57:42.398712] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.602 [2024-12-06 17:57:42.406638] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.602 [2024-12-06 17:57:42.406689] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.602 [2024-12-06 17:57:42.414639] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.602 [2024-12-06 17:57:42.414691] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.602 [2024-12-06 17:57:42.422618] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.602 [2024-12-06 17:57:42.422681] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.602 [2024-12-06 17:57:42.430588] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.602 [2024-12-06 17:57:42.430608] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.602 [2024-12-06 17:57:42.438611] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.602 [2024-12-06 17:57:42.438646] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.860 [2024-12-06 17:57:42.446640] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.860 [2024-12-06 17:57:42.446694] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.860 [2024-12-06 17:57:42.454641] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.860 [2024-12-06 17:57:42.454696] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.860 [2024-12-06 17:57:42.462581] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.860 [2024-12-06 17:57:42.462601] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.860 [2024-12-06 17:57:42.470581] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.860 [2024-12-06 17:57:42.470599] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.860 [2024-12-06 17:57:42.478580] subsystem.c:2130:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:42:00.860 [2024-12-06 17:57:42.478598] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:00.860 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh: line 42: kill: (452390) - No such process 00:42:00.860 17:57:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@49 -- # wait 452390 00:42:00.860 17:57:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@52 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:42:00.860 17:57:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:42:00.860 17:57:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:42:00.860 17:57:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:42:00.860 17:57:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@53 -- # rpc_cmd bdev_delay_create -b malloc0 -d delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:42:00.860 17:57:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:42:00.860 17:57:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:42:00.860 delay0 00:42:00.860 17:57:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:42:00.860 17:57:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@54 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 delay0 -n 1 00:42:00.860 17:57:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@563 -- # xtrace_disable 00:42:00.860 17:57:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:42:00.860 17:57:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:42:00.860 17:57:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -c 0x1 -t 5 -q 64 -w randrw -M 50 -l warning -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 ns:1' 00:42:00.860 [2024-12-06 17:57:42.556041] nvme_fabric.c: 295:nvme_fabric_discover_probe: *WARNING*: Skipping unsupported current discovery service or discovery service referral 00:42:07.414 Initializing NVMe Controllers 00:42:07.414 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:42:07.414 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:42:07.414 Initialization complete. Launching workers. 00:42:07.414 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 I/O completed: 320, failed: 103 00:42:07.414 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) abort submitted 389, failed to submit 34 00:42:07.414 success 258, unsuccessful 131, failed 0 00:42:07.414 17:57:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@59 -- # trap - SIGINT SIGTERM EXIT 00:42:07.414 17:57:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@60 -- # nvmftestfini 00:42:07.414 17:57:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@516 -- # nvmfcleanup 00:42:07.414 17:57:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@121 -- # sync 00:42:07.414 17:57:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:42:07.414 17:57:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@124 -- # set +e 00:42:07.414 17:57:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@125 -- # for i in {1..20} 00:42:07.414 17:57:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:42:07.414 rmmod nvme_tcp 00:42:07.414 rmmod nvme_fabrics 00:42:07.414 rmmod nvme_keyring 00:42:07.414 17:57:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:42:07.414 17:57:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@128 -- # set -e 00:42:07.414 17:57:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@129 -- # return 0 00:42:07.414 17:57:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@517 -- # '[' -n 451191 ']' 00:42:07.414 17:57:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@518 -- # killprocess 451191 00:42:07.414 17:57:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@954 -- # '[' -z 451191 ']' 00:42:07.414 17:57:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@958 -- # kill -0 451191 00:42:07.414 17:57:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@959 -- # uname 00:42:07.414 17:57:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:42:07.414 17:57:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 451191 00:42:07.414 17:57:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:42:07.414 17:57:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:42:07.414 17:57:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@972 -- # echo 'killing process with pid 451191' 00:42:07.414 killing process with pid 451191 00:42:07.414 17:57:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@973 -- # kill 451191 00:42:07.414 17:57:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@978 -- # wait 451191 00:42:07.414 17:57:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:42:07.414 17:57:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:42:07.414 17:57:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:42:07.414 17:57:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@297 -- # iptr 00:42:07.414 17:57:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@791 -- # iptables-save 00:42:07.414 17:57:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:42:07.414 17:57:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@791 -- # iptables-restore 00:42:07.414 17:57:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:42:07.414 17:57:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@302 -- # remove_spdk_ns 00:42:07.414 17:57:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:42:07.414 17:57:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:42:07.415 17:57:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:42:09.423 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:42:09.423 00:42:09.423 real 0m27.778s 00:42:09.423 user 0m38.037s 00:42:09.423 sys 0m10.097s 00:42:09.423 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1130 -- # xtrace_disable 00:42:09.423 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:42:09.423 ************************************ 00:42:09.423 END TEST nvmf_zcopy 00:42:09.423 ************************************ 00:42:09.423 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@33 -- # run_test nvmf_nmic /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nmic.sh --transport=tcp --interrupt-mode 00:42:09.423 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:42:09.423 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1111 -- # xtrace_disable 00:42:09.423 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:42:09.423 ************************************ 00:42:09.423 START TEST nvmf_nmic 00:42:09.423 ************************************ 00:42:09.423 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nmic.sh --transport=tcp --interrupt-mode 00:42:09.423 * Looking for test storage... 00:42:09.423 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:42:09.424 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:42:09.424 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1711 -- # lcov --version 00:42:09.424 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@333 -- # local ver1 ver1_l 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@334 -- # local ver2 ver2_l 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@336 -- # IFS=.-: 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@336 -- # read -ra ver1 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@337 -- # IFS=.-: 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@337 -- # read -ra ver2 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@338 -- # local 'op=<' 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@340 -- # ver1_l=2 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@341 -- # ver2_l=1 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@344 -- # case "$op" in 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@345 -- # : 1 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@364 -- # (( v = 0 )) 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@365 -- # decimal 1 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@353 -- # local d=1 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@355 -- # echo 1 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@365 -- # ver1[v]=1 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@366 -- # decimal 2 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@353 -- # local d=2 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@355 -- # echo 2 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@366 -- # ver2[v]=2 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@368 -- # return 0 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:42:09.683 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:42:09.683 --rc genhtml_branch_coverage=1 00:42:09.683 --rc genhtml_function_coverage=1 00:42:09.683 --rc genhtml_legend=1 00:42:09.683 --rc geninfo_all_blocks=1 00:42:09.683 --rc geninfo_unexecuted_blocks=1 00:42:09.683 00:42:09.683 ' 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:42:09.683 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:42:09.683 --rc genhtml_branch_coverage=1 00:42:09.683 --rc genhtml_function_coverage=1 00:42:09.683 --rc genhtml_legend=1 00:42:09.683 --rc geninfo_all_blocks=1 00:42:09.683 --rc geninfo_unexecuted_blocks=1 00:42:09.683 00:42:09.683 ' 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:42:09.683 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:42:09.683 --rc genhtml_branch_coverage=1 00:42:09.683 --rc genhtml_function_coverage=1 00:42:09.683 --rc genhtml_legend=1 00:42:09.683 --rc geninfo_all_blocks=1 00:42:09.683 --rc geninfo_unexecuted_blocks=1 00:42:09.683 00:42:09.683 ' 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:42:09.683 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:42:09.683 --rc genhtml_branch_coverage=1 00:42:09.683 --rc genhtml_function_coverage=1 00:42:09.683 --rc genhtml_legend=1 00:42:09.683 --rc geninfo_all_blocks=1 00:42:09.683 --rc geninfo_unexecuted_blocks=1 00:42:09.683 00:42:09.683 ' 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@7 -- # uname -s 00:42:09.683 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@15 -- # shopt -s extglob 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- paths/export.sh@5 -- # export PATH 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@51 -- # : 0 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@55 -- # have_pci_nics=0 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@11 -- # MALLOC_BDEV_SIZE=64 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@14 -- # nvmftestinit 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@476 -- # prepare_net_devs 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@438 -- # local -g is_hw=no 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@440 -- # remove_spdk_ns 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@309 -- # xtrace_disable 00:42:09.684 17:57:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:42:11.585 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:42:11.585 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@315 -- # pci_devs=() 00:42:11.585 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@315 -- # local -a pci_devs 00:42:11.585 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@316 -- # pci_net_devs=() 00:42:11.585 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:42:11.585 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@317 -- # pci_drivers=() 00:42:11.585 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@317 -- # local -A pci_drivers 00:42:11.585 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@319 -- # net_devs=() 00:42:11.585 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@319 -- # local -ga net_devs 00:42:11.585 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@320 -- # e810=() 00:42:11.585 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@320 -- # local -ga e810 00:42:11.585 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@321 -- # x722=() 00:42:11.585 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@321 -- # local -ga x722 00:42:11.585 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@322 -- # mlx=() 00:42:11.585 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@322 -- # local -ga mlx 00:42:11.585 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:42:11.585 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:42:11.585 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:42:11.585 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:42:11.585 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:42:11.585 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:42:11.585 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:42:11.585 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:42:11.585 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:42:11.585 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:42:11.585 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:42:11.586 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:42:11.586 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@418 -- # [[ up == up ]] 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:42:11.586 Found net devices under 0000:0a:00.0: cvl_0_0 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@418 -- # [[ up == up ]] 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:42:11.586 Found net devices under 0000:0a:00.1: cvl_0_1 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@442 -- # is_hw=yes 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:42:11.586 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:42:11.845 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:42:11.845 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.268 ms 00:42:11.845 00:42:11.845 --- 10.0.0.2 ping statistics --- 00:42:11.845 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:42:11.845 rtt min/avg/max/mdev = 0.268/0.268/0.268/0.000 ms 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:42:11.845 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:42:11.845 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.174 ms 00:42:11.845 00:42:11.845 --- 10.0.0.1 ping statistics --- 00:42:11.845 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:42:11.845 rtt min/avg/max/mdev = 0.174/0.174/0.174/0.000 ms 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@450 -- # return 0 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@15 -- # nvmfappstart -m 0xF 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@726 -- # xtrace_disable 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@509 -- # nvmfpid=455766 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0xF 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@510 -- # waitforlisten 455766 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@835 -- # '[' -z 455766 ']' 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@840 -- # local max_retries=100 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:42:11.845 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@844 -- # xtrace_disable 00:42:11.845 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:42:11.845 [2024-12-06 17:57:53.582542] thread.c:3005:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:42:11.845 [2024-12-06 17:57:53.583607] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:42:11.845 [2024-12-06 17:57:53.583680] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:42:11.845 [2024-12-06 17:57:53.652850] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:42:12.104 [2024-12-06 17:57:53.699661] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:42:12.104 [2024-12-06 17:57:53.699730] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:42:12.104 [2024-12-06 17:57:53.699744] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:42:12.104 [2024-12-06 17:57:53.699755] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:42:12.104 [2024-12-06 17:57:53.699764] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:42:12.104 [2024-12-06 17:57:53.701302] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:42:12.104 [2024-12-06 17:57:53.701412] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:42:12.104 [2024-12-06 17:57:53.701481] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:42:12.104 [2024-12-06 17:57:53.701484] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:42:12.104 [2024-12-06 17:57:53.781708] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:42:12.104 [2024-12-06 17:57:53.781931] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:42:12.104 [2024-12-06 17:57:53.782176] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:42:12.104 [2024-12-06 17:57:53.782772] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:42:12.104 [2024-12-06 17:57:53.783009] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_003) to intr mode from intr mode. 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@868 -- # return 0 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@732 -- # xtrace_disable 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:42:12.104 [2024-12-06 17:57:53.838160] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@20 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:42:12.104 Malloc0 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@21 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@22 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@23 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:42:12.104 [2024-12-06 17:57:53.902365] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@25 -- # echo 'test case1: single bdev can'\''t be used in multiple subsystems' 00:42:12.104 test case1: single bdev can't be used in multiple subsystems 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@26 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 -a -s SPDK2 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@28 -- # nmic_status=0 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 Malloc0 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:42:12.104 [2024-12-06 17:57:53.926092] bdev.c:8515:bdev_open: *ERROR*: bdev Malloc0 already claimed: type exclusive_write by module NVMe-oF Target 00:42:12.104 [2024-12-06 17:57:53.926121] subsystem.c:2160:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode2: bdev Malloc0 cannot be opened, error=-1 00:42:12.104 [2024-12-06 17:57:53.926159] nvmf_rpc.c:1520:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:42:12.104 request: 00:42:12.104 { 00:42:12.104 "nqn": "nqn.2016-06.io.spdk:cnode2", 00:42:12.104 "namespace": { 00:42:12.104 "bdev_name": "Malloc0", 00:42:12.104 "no_auto_visible": false, 00:42:12.104 "hide_metadata": false 00:42:12.104 }, 00:42:12.104 "method": "nvmf_subsystem_add_ns", 00:42:12.104 "req_id": 1 00:42:12.104 } 00:42:12.104 Got JSON-RPC error response 00:42:12.104 response: 00:42:12.104 { 00:42:12.104 "code": -32602, 00:42:12.104 "message": "Invalid parameters" 00:42:12.104 } 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@29 -- # nmic_status=1 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@31 -- # '[' 1 -eq 0 ']' 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@36 -- # echo ' Adding namespace failed - expected result.' 00:42:12.104 Adding namespace failed - expected result. 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@39 -- # echo 'test case2: host connect to nvmf target in multiple paths' 00:42:12.104 test case2: host connect to nvmf target in multiple paths 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@40 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@563 -- # xtrace_disable 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:42:12.104 [2024-12-06 17:57:53.934187] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:42:12.104 17:57:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@41 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:42:12.363 17:57:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@42 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4421 00:42:12.620 17:57:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@44 -- # waitforserial SPDKISFASTANDAWESOME 00:42:12.620 17:57:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1202 -- # local i=0 00:42:12.620 17:57:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:42:12.620 17:57:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:42:12.620 17:57:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1209 -- # sleep 2 00:42:14.514 17:57:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:42:14.514 17:57:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:42:14.514 17:57:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:42:14.772 17:57:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:42:14.772 17:57:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:42:14.772 17:57:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1212 -- # return 0 00:42:14.772 17:57:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t write -r 1 -v 00:42:14.772 [global] 00:42:14.772 thread=1 00:42:14.772 invalidate=1 00:42:14.772 rw=write 00:42:14.772 time_based=1 00:42:14.772 runtime=1 00:42:14.772 ioengine=libaio 00:42:14.772 direct=1 00:42:14.772 bs=4096 00:42:14.772 iodepth=1 00:42:14.772 norandommap=0 00:42:14.772 numjobs=1 00:42:14.772 00:42:14.772 verify_dump=1 00:42:14.772 verify_backlog=512 00:42:14.772 verify_state_save=0 00:42:14.772 do_verify=1 00:42:14.772 verify=crc32c-intel 00:42:14.772 [job0] 00:42:14.772 filename=/dev/nvme0n1 00:42:14.772 Could not set queue depth (nvme0n1) 00:42:14.772 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:42:14.772 fio-3.35 00:42:14.772 Starting 1 thread 00:42:16.145 00:42:16.145 job0: (groupid=0, jobs=1): err= 0: pid=456263: Fri Dec 6 17:57:57 2024 00:42:16.145 read: IOPS=2434, BW=9738KiB/s (9972kB/s)(9748KiB/1001msec) 00:42:16.145 slat (nsec): min=5314, max=57769, avg=7915.34, stdev=4276.76 00:42:16.145 clat (usec): min=191, max=412, avg=229.29, stdev=17.78 00:42:16.145 lat (usec): min=197, max=430, avg=237.21, stdev=20.94 00:42:16.145 clat percentiles (usec): 00:42:16.145 | 1.00th=[ 200], 5.00th=[ 206], 10.00th=[ 212], 20.00th=[ 219], 00:42:16.145 | 30.00th=[ 221], 40.00th=[ 223], 50.00th=[ 225], 60.00th=[ 227], 00:42:16.145 | 70.00th=[ 233], 80.00th=[ 241], 90.00th=[ 260], 95.00th=[ 265], 00:42:16.145 | 99.00th=[ 277], 99.50th=[ 281], 99.90th=[ 306], 99.95th=[ 310], 00:42:16.145 | 99.99th=[ 412] 00:42:16.145 write: IOPS=2557, BW=9.99MiB/s (10.5MB/s)(10.0MiB/1001msec); 0 zone resets 00:42:16.145 slat (nsec): min=6767, max=37305, avg=8546.47, stdev=2861.16 00:42:16.145 clat (usec): min=134, max=253, avg=151.46, stdev= 8.61 00:42:16.145 lat (usec): min=147, max=281, avg=160.00, stdev=10.45 00:42:16.145 clat percentiles (usec): 00:42:16.145 | 1.00th=[ 143], 5.00th=[ 145], 10.00th=[ 145], 20.00th=[ 147], 00:42:16.145 | 30.00th=[ 147], 40.00th=[ 149], 50.00th=[ 149], 60.00th=[ 151], 00:42:16.145 | 70.00th=[ 153], 80.00th=[ 157], 90.00th=[ 161], 95.00th=[ 167], 00:42:16.145 | 99.00th=[ 184], 99.50th=[ 196], 99.90th=[ 221], 99.95th=[ 223], 00:42:16.145 | 99.99th=[ 253] 00:42:16.145 bw ( KiB/s): min=12288, max=12288, per=100.00%, avg=12288.00, stdev= 0.00, samples=1 00:42:16.145 iops : min= 3072, max= 3072, avg=3072.00, stdev= 0.00, samples=1 00:42:16.145 lat (usec) : 250=92.00%, 500=8.00% 00:42:16.145 cpu : usr=3.80%, sys=5.10%, ctx=4997, majf=0, minf=1 00:42:16.145 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:42:16.145 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:16.145 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:16.145 issued rwts: total=2437,2560,0,0 short=0,0,0,0 dropped=0,0,0,0 00:42:16.145 latency : target=0, window=0, percentile=100.00%, depth=1 00:42:16.145 00:42:16.145 Run status group 0 (all jobs): 00:42:16.145 READ: bw=9738KiB/s (9972kB/s), 9738KiB/s-9738KiB/s (9972kB/s-9972kB/s), io=9748KiB (9982kB), run=1001-1001msec 00:42:16.145 WRITE: bw=9.99MiB/s (10.5MB/s), 9.99MiB/s-9.99MiB/s (10.5MB/s-10.5MB/s), io=10.0MiB (10.5MB), run=1001-1001msec 00:42:16.145 00:42:16.145 Disk stats (read/write): 00:42:16.145 nvme0n1: ios=2128/2560, merge=0/0, ticks=465/367, in_queue=832, util=91.68% 00:42:16.145 17:57:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@48 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:42:16.404 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 2 controller(s) 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@49 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1223 -- # local i=0 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1235 -- # return 0 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@51 -- # trap - SIGINT SIGTERM EXIT 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@53 -- # nvmftestfini 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@516 -- # nvmfcleanup 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@121 -- # sync 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@124 -- # set +e 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@125 -- # for i in {1..20} 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:42:16.404 rmmod nvme_tcp 00:42:16.404 rmmod nvme_fabrics 00:42:16.404 rmmod nvme_keyring 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@128 -- # set -e 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@129 -- # return 0 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@517 -- # '[' -n 455766 ']' 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@518 -- # killprocess 455766 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@954 -- # '[' -z 455766 ']' 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@958 -- # kill -0 455766 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@959 -- # uname 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 455766 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@972 -- # echo 'killing process with pid 455766' 00:42:16.404 killing process with pid 455766 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@973 -- # kill 455766 00:42:16.404 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@978 -- # wait 455766 00:42:16.663 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:42:16.663 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:42:16.663 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:42:16.663 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@297 -- # iptr 00:42:16.663 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@791 -- # iptables-save 00:42:16.663 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@791 -- # iptables-restore 00:42:16.663 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:42:16.663 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:42:16.663 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@302 -- # remove_spdk_ns 00:42:16.663 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:42:16.663 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:42:16.663 17:57:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:42:18.585 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:42:18.585 00:42:18.585 real 0m9.247s 00:42:18.585 user 0m17.527s 00:42:18.585 sys 0m3.584s 00:42:18.585 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1130 -- # xtrace_disable 00:42:18.585 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:42:18.585 ************************************ 00:42:18.585 END TEST nvmf_nmic 00:42:18.585 ************************************ 00:42:18.844 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@34 -- # run_test nvmf_fio_target /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fio.sh --transport=tcp --interrupt-mode 00:42:18.844 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:42:18.844 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1111 -- # xtrace_disable 00:42:18.844 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:42:18.844 ************************************ 00:42:18.844 START TEST nvmf_fio_target 00:42:18.844 ************************************ 00:42:18.844 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fio.sh --transport=tcp --interrupt-mode 00:42:18.844 * Looking for test storage... 00:42:18.844 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:42:18.844 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:42:18.844 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1711 -- # lcov --version 00:42:18.844 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:42:18.844 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:42:18.844 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:42:18.844 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@333 -- # local ver1 ver1_l 00:42:18.844 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@334 -- # local ver2 ver2_l 00:42:18.844 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@336 -- # IFS=.-: 00:42:18.844 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@336 -- # read -ra ver1 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@337 -- # IFS=.-: 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@337 -- # read -ra ver2 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@338 -- # local 'op=<' 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@340 -- # ver1_l=2 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@341 -- # ver2_l=1 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@344 -- # case "$op" in 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@345 -- # : 1 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@364 -- # (( v = 0 )) 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@365 -- # decimal 1 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@353 -- # local d=1 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@355 -- # echo 1 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@365 -- # ver1[v]=1 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@366 -- # decimal 2 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@353 -- # local d=2 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@355 -- # echo 2 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@366 -- # ver2[v]=2 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@368 -- # return 0 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:42:18.845 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:42:18.845 --rc genhtml_branch_coverage=1 00:42:18.845 --rc genhtml_function_coverage=1 00:42:18.845 --rc genhtml_legend=1 00:42:18.845 --rc geninfo_all_blocks=1 00:42:18.845 --rc geninfo_unexecuted_blocks=1 00:42:18.845 00:42:18.845 ' 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:42:18.845 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:42:18.845 --rc genhtml_branch_coverage=1 00:42:18.845 --rc genhtml_function_coverage=1 00:42:18.845 --rc genhtml_legend=1 00:42:18.845 --rc geninfo_all_blocks=1 00:42:18.845 --rc geninfo_unexecuted_blocks=1 00:42:18.845 00:42:18.845 ' 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:42:18.845 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:42:18.845 --rc genhtml_branch_coverage=1 00:42:18.845 --rc genhtml_function_coverage=1 00:42:18.845 --rc genhtml_legend=1 00:42:18.845 --rc geninfo_all_blocks=1 00:42:18.845 --rc geninfo_unexecuted_blocks=1 00:42:18.845 00:42:18.845 ' 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:42:18.845 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:42:18.845 --rc genhtml_branch_coverage=1 00:42:18.845 --rc genhtml_function_coverage=1 00:42:18.845 --rc genhtml_legend=1 00:42:18.845 --rc geninfo_all_blocks=1 00:42:18.845 --rc geninfo_unexecuted_blocks=1 00:42:18.845 00:42:18.845 ' 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@7 -- # uname -s 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@15 -- # shopt -s extglob 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- paths/export.sh@5 -- # export PATH 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@51 -- # : 0 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@55 -- # have_pci_nics=0 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@11 -- # MALLOC_BDEV_SIZE=64 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@14 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:42:18.845 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@16 -- # nvmftestinit 00:42:18.846 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:42:18.846 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:42:18.846 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@476 -- # prepare_net_devs 00:42:18.846 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@438 -- # local -g is_hw=no 00:42:18.846 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@440 -- # remove_spdk_ns 00:42:18.846 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:42:18.846 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:42:18.846 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:42:18.846 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:42:18.846 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:42:18.846 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@309 -- # xtrace_disable 00:42:18.846 17:58:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:42:21.387 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:42:21.387 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@315 -- # pci_devs=() 00:42:21.387 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@315 -- # local -a pci_devs 00:42:21.387 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@316 -- # pci_net_devs=() 00:42:21.387 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:42:21.387 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@317 -- # pci_drivers=() 00:42:21.387 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@317 -- # local -A pci_drivers 00:42:21.387 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@319 -- # net_devs=() 00:42:21.387 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@319 -- # local -ga net_devs 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@320 -- # e810=() 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@320 -- # local -ga e810 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@321 -- # x722=() 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@321 -- # local -ga x722 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@322 -- # mlx=() 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@322 -- # local -ga mlx 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:42:21.388 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:42:21.388 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@418 -- # [[ up == up ]] 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:42:21.388 Found net devices under 0000:0a:00.0: cvl_0_0 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@418 -- # [[ up == up ]] 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:42:21.388 Found net devices under 0000:0a:00.1: cvl_0_1 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@442 -- # is_hw=yes 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:42:21.388 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:42:21.388 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:42:21.388 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.177 ms 00:42:21.389 00:42:21.389 --- 10.0.0.2 ping statistics --- 00:42:21.389 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:42:21.389 rtt min/avg/max/mdev = 0.177/0.177/0.177/0.000 ms 00:42:21.389 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:42:21.389 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:42:21.389 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.051 ms 00:42:21.389 00:42:21.389 --- 10.0.0.1 ping statistics --- 00:42:21.389 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:42:21.389 rtt min/avg/max/mdev = 0.051/0.051/0.051/0.000 ms 00:42:21.389 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:42:21.389 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@450 -- # return 0 00:42:21.389 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:42:21.389 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:42:21.389 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:42:21.389 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:42:21.389 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:42:21.389 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:42:21.389 17:58:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:42:21.389 17:58:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@17 -- # nvmfappstart -m 0xF 00:42:21.389 17:58:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:42:21.389 17:58:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@726 -- # xtrace_disable 00:42:21.389 17:58:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:42:21.389 17:58:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@509 -- # nvmfpid=458340 00:42:21.389 17:58:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@510 -- # waitforlisten 458340 00:42:21.389 17:58:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0xF 00:42:21.389 17:58:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@835 -- # '[' -z 458340 ']' 00:42:21.389 17:58:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:42:21.389 17:58:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@840 -- # local max_retries=100 00:42:21.389 17:58:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:42:21.389 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:42:21.389 17:58:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@844 -- # xtrace_disable 00:42:21.389 17:58:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:42:21.389 [2024-12-06 17:58:03.069816] thread.c:3005:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:42:21.389 [2024-12-06 17:58:03.070866] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:42:21.389 [2024-12-06 17:58:03.070939] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:42:21.389 [2024-12-06 17:58:03.147345] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:42:21.389 [2024-12-06 17:58:03.195372] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:42:21.389 [2024-12-06 17:58:03.195443] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:42:21.389 [2024-12-06 17:58:03.195457] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:42:21.389 [2024-12-06 17:58:03.195468] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:42:21.389 [2024-12-06 17:58:03.195477] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:42:21.389 [2024-12-06 17:58:03.197094] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:42:21.389 [2024-12-06 17:58:03.197166] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:42:21.389 [2024-12-06 17:58:03.197232] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:42:21.389 [2024-12-06 17:58:03.197235] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:42:21.646 [2024-12-06 17:58:03.286848] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:42:21.646 [2024-12-06 17:58:03.287068] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:42:21.646 [2024-12-06 17:58:03.287343] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:42:21.646 [2024-12-06 17:58:03.287961] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:42:21.646 [2024-12-06 17:58:03.288195] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_003) to intr mode from intr mode. 00:42:21.646 17:58:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:42:21.646 17:58:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@868 -- # return 0 00:42:21.646 17:58:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:42:21.646 17:58:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@732 -- # xtrace_disable 00:42:21.646 17:58:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:42:21.646 17:58:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:42:21.646 17:58:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:42:21.904 [2024-12-06 17:58:03.573927] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:42:21.904 17:58:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:42:22.161 17:58:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@21 -- # malloc_bdevs='Malloc0 ' 00:42:22.161 17:58:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@22 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:42:22.423 17:58:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@22 -- # malloc_bdevs+=Malloc1 00:42:22.423 17:58:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:42:22.989 17:58:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@24 -- # raid_malloc_bdevs='Malloc2 ' 00:42:22.989 17:58:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:42:23.247 17:58:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@25 -- # raid_malloc_bdevs+=Malloc3 00:42:23.247 17:58:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n raid0 -z 64 -r 0 -b 'Malloc2 Malloc3' 00:42:23.507 17:58:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:42:23.764 17:58:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@29 -- # concat_malloc_bdevs='Malloc4 ' 00:42:23.764 17:58:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:42:24.021 17:58:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@30 -- # concat_malloc_bdevs+='Malloc5 ' 00:42:24.021 17:58:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:42:24.279 17:58:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@31 -- # concat_malloc_bdevs+=Malloc6 00:42:24.279 17:58:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n concat0 -r concat -z 64 -b 'Malloc4 Malloc5 Malloc6' 00:42:24.536 17:58:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:42:24.794 17:58:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@35 -- # for malloc_bdev in $malloc_bdevs 00:42:24.794 17:58:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:42:25.052 17:58:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@35 -- # for malloc_bdev in $malloc_bdevs 00:42:25.052 17:58:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:42:25.309 17:58:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:42:25.568 [2024-12-06 17:58:07.406109] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:42:25.826 17:58:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 raid0 00:42:26.085 17:58:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 concat0 00:42:26.343 17:58:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@46 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:42:26.343 17:58:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@48 -- # waitforserial SPDKISFASTANDAWESOME 4 00:42:26.343 17:58:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1202 -- # local i=0 00:42:26.343 17:58:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:42:26.343 17:58:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1204 -- # [[ -n 4 ]] 00:42:26.343 17:58:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1205 -- # nvme_device_counter=4 00:42:26.343 17:58:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1209 -- # sleep 2 00:42:28.872 17:58:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:42:28.872 17:58:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:42:28.873 17:58:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:42:28.873 17:58:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1211 -- # nvme_devices=4 00:42:28.873 17:58:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:42:28.873 17:58:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1212 -- # return 0 00:42:28.873 17:58:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t write -r 1 -v 00:42:28.873 [global] 00:42:28.873 thread=1 00:42:28.873 invalidate=1 00:42:28.873 rw=write 00:42:28.873 time_based=1 00:42:28.873 runtime=1 00:42:28.873 ioengine=libaio 00:42:28.873 direct=1 00:42:28.873 bs=4096 00:42:28.873 iodepth=1 00:42:28.873 norandommap=0 00:42:28.873 numjobs=1 00:42:28.873 00:42:28.873 verify_dump=1 00:42:28.873 verify_backlog=512 00:42:28.873 verify_state_save=0 00:42:28.873 do_verify=1 00:42:28.873 verify=crc32c-intel 00:42:28.873 [job0] 00:42:28.873 filename=/dev/nvme0n1 00:42:28.873 [job1] 00:42:28.873 filename=/dev/nvme0n2 00:42:28.873 [job2] 00:42:28.873 filename=/dev/nvme0n3 00:42:28.873 [job3] 00:42:28.873 filename=/dev/nvme0n4 00:42:28.873 Could not set queue depth (nvme0n1) 00:42:28.873 Could not set queue depth (nvme0n2) 00:42:28.873 Could not set queue depth (nvme0n3) 00:42:28.873 Could not set queue depth (nvme0n4) 00:42:28.873 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:42:28.873 job1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:42:28.873 job2: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:42:28.873 job3: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:42:28.873 fio-3.35 00:42:28.873 Starting 4 threads 00:42:29.807 00:42:29.807 job0: (groupid=0, jobs=1): err= 0: pid=459406: Fri Dec 6 17:58:11 2024 00:42:29.807 read: IOPS=28, BW=114KiB/s (117kB/s)(116KiB/1015msec) 00:42:29.807 slat (nsec): min=7174, max=46103, avg=16442.86, stdev=8991.64 00:42:29.807 clat (usec): min=318, max=41250, avg=31213.07, stdev=17633.30 00:42:29.807 lat (usec): min=332, max=41257, avg=31229.51, stdev=17636.69 00:42:29.807 clat percentiles (usec): 00:42:29.807 | 1.00th=[ 318], 5.00th=[ 441], 10.00th=[ 478], 20.00th=[ 603], 00:42:29.807 | 30.00th=[41157], 40.00th=[41157], 50.00th=[41157], 60.00th=[41157], 00:42:29.807 | 70.00th=[41157], 80.00th=[41157], 90.00th=[41157], 95.00th=[41157], 00:42:29.807 | 99.00th=[41157], 99.50th=[41157], 99.90th=[41157], 99.95th=[41157], 00:42:29.807 | 99.99th=[41157] 00:42:29.807 write: IOPS=504, BW=2018KiB/s (2066kB/s)(2048KiB/1015msec); 0 zone resets 00:42:29.807 slat (nsec): min=6407, max=24850, avg=8373.02, stdev=2017.25 00:42:29.807 clat (usec): min=158, max=353, avg=202.20, stdev=28.99 00:42:29.807 lat (usec): min=167, max=363, avg=210.57, stdev=29.73 00:42:29.807 clat percentiles (usec): 00:42:29.807 | 1.00th=[ 163], 5.00th=[ 169], 10.00th=[ 174], 20.00th=[ 178], 00:42:29.807 | 30.00th=[ 182], 40.00th=[ 186], 50.00th=[ 190], 60.00th=[ 198], 00:42:29.807 | 70.00th=[ 223], 80.00th=[ 241], 90.00th=[ 245], 95.00th=[ 245], 00:42:29.807 | 99.00th=[ 253], 99.50th=[ 281], 99.90th=[ 355], 99.95th=[ 355], 00:42:29.807 | 99.99th=[ 355] 00:42:29.807 bw ( KiB/s): min= 4096, max= 4096, per=34.70%, avg=4096.00, stdev= 0.00, samples=1 00:42:29.807 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:42:29.807 lat (usec) : 250=92.98%, 500=2.40%, 750=0.55% 00:42:29.807 lat (msec) : 50=4.07% 00:42:29.807 cpu : usr=0.10%, sys=0.59%, ctx=542, majf=0, minf=1 00:42:29.807 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:42:29.807 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:29.807 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:29.807 issued rwts: total=29,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:42:29.807 latency : target=0, window=0, percentile=100.00%, depth=1 00:42:29.807 job1: (groupid=0, jobs=1): err= 0: pid=459407: Fri Dec 6 17:58:11 2024 00:42:29.807 read: IOPS=22, BW=89.8KiB/s (91.9kB/s)(92.0KiB/1025msec) 00:42:29.807 slat (nsec): min=6076, max=33331, avg=17398.17, stdev=7746.46 00:42:29.807 clat (usec): min=285, max=42050, avg=39966.76, stdev=8660.34 00:42:29.807 lat (usec): min=302, max=42062, avg=39984.16, stdev=8660.64 00:42:29.807 clat percentiles (usec): 00:42:29.807 | 1.00th=[ 285], 5.00th=[40633], 10.00th=[41157], 20.00th=[41157], 00:42:29.807 | 30.00th=[41681], 40.00th=[42206], 50.00th=[42206], 60.00th=[42206], 00:42:29.807 | 70.00th=[42206], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:42:29.807 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:42:29.807 | 99.99th=[42206] 00:42:29.807 write: IOPS=499, BW=1998KiB/s (2046kB/s)(2048KiB/1025msec); 0 zone resets 00:42:29.807 slat (nsec): min=5718, max=29092, avg=6917.41, stdev=2091.47 00:42:29.807 clat (usec): min=162, max=268, avg=196.44, stdev=20.55 00:42:29.807 lat (usec): min=169, max=297, avg=203.36, stdev=20.86 00:42:29.807 clat percentiles (usec): 00:42:29.807 | 1.00th=[ 169], 5.00th=[ 174], 10.00th=[ 176], 20.00th=[ 180], 00:42:29.807 | 30.00th=[ 184], 40.00th=[ 186], 50.00th=[ 190], 60.00th=[ 196], 00:42:29.807 | 70.00th=[ 204], 80.00th=[ 212], 90.00th=[ 231], 95.00th=[ 245], 00:42:29.807 | 99.00th=[ 249], 99.50th=[ 251], 99.90th=[ 269], 99.95th=[ 269], 00:42:29.807 | 99.99th=[ 269] 00:42:29.807 bw ( KiB/s): min= 4096, max= 4096, per=34.70%, avg=4096.00, stdev= 0.00, samples=1 00:42:29.807 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:42:29.807 lat (usec) : 250=95.14%, 500=0.75% 00:42:29.807 lat (msec) : 50=4.11% 00:42:29.807 cpu : usr=0.39%, sys=0.10%, ctx=535, majf=0, minf=1 00:42:29.807 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:42:29.807 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:29.807 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:29.807 issued rwts: total=23,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:42:29.807 latency : target=0, window=0, percentile=100.00%, depth=1 00:42:29.807 job2: (groupid=0, jobs=1): err= 0: pid=459408: Fri Dec 6 17:58:11 2024 00:42:29.807 read: IOPS=510, BW=2040KiB/s (2089kB/s)(2124KiB/1041msec) 00:42:29.807 slat (nsec): min=4437, max=32626, avg=6778.08, stdev=3809.20 00:42:29.807 clat (usec): min=195, max=41175, avg=1565.25, stdev=7176.43 00:42:29.807 lat (usec): min=201, max=41180, avg=1572.03, stdev=7178.62 00:42:29.807 clat percentiles (usec): 00:42:29.807 | 1.00th=[ 227], 5.00th=[ 233], 10.00th=[ 237], 20.00th=[ 243], 00:42:29.807 | 30.00th=[ 245], 40.00th=[ 249], 50.00th=[ 251], 60.00th=[ 253], 00:42:29.807 | 70.00th=[ 258], 80.00th=[ 265], 90.00th=[ 322], 95.00th=[ 404], 00:42:29.807 | 99.00th=[41157], 99.50th=[41157], 99.90th=[41157], 99.95th=[41157], 00:42:29.807 | 99.99th=[41157] 00:42:29.807 write: IOPS=983, BW=3935KiB/s (4029kB/s)(4096KiB/1041msec); 0 zone resets 00:42:29.807 slat (nsec): min=5778, max=27771, avg=6892.74, stdev=1806.57 00:42:29.807 clat (usec): min=147, max=413, avg=191.58, stdev=25.00 00:42:29.807 lat (usec): min=154, max=419, avg=198.47, stdev=25.20 00:42:29.807 clat percentiles (usec): 00:42:29.807 | 1.00th=[ 157], 5.00th=[ 163], 10.00th=[ 167], 20.00th=[ 172], 00:42:29.807 | 30.00th=[ 176], 40.00th=[ 180], 50.00th=[ 184], 60.00th=[ 192], 00:42:29.807 | 70.00th=[ 204], 80.00th=[ 217], 90.00th=[ 229], 95.00th=[ 235], 00:42:29.807 | 99.00th=[ 253], 99.50th=[ 265], 99.90th=[ 297], 99.95th=[ 412], 00:42:29.807 | 99.99th=[ 412] 00:42:29.807 bw ( KiB/s): min= 8192, max= 8192, per=69.40%, avg=8192.00, stdev= 0.00, samples=1 00:42:29.807 iops : min= 2048, max= 2048, avg=2048.00, stdev= 0.00, samples=1 00:42:29.807 lat (usec) : 250=80.39%, 500=18.39%, 750=0.13% 00:42:29.807 lat (msec) : 50=1.09% 00:42:29.807 cpu : usr=0.67%, sys=0.77%, ctx=1555, majf=0, minf=2 00:42:29.807 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:42:29.807 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:29.807 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:29.807 issued rwts: total=531,1024,0,0 short=0,0,0,0 dropped=0,0,0,0 00:42:29.807 latency : target=0, window=0, percentile=100.00%, depth=1 00:42:29.807 job3: (groupid=0, jobs=1): err= 0: pid=459409: Fri Dec 6 17:58:11 2024 00:42:29.807 read: IOPS=525, BW=2104KiB/s (2154kB/s)(2112KiB/1004msec) 00:42:29.807 slat (nsec): min=4321, max=33465, avg=5772.32, stdev=3266.32 00:42:29.807 clat (usec): min=198, max=41124, avg=1535.44, stdev=7197.21 00:42:29.807 lat (usec): min=203, max=41129, avg=1541.22, stdev=7199.51 00:42:29.807 clat percentiles (usec): 00:42:29.807 | 1.00th=[ 204], 5.00th=[ 208], 10.00th=[ 210], 20.00th=[ 215], 00:42:29.807 | 30.00th=[ 217], 40.00th=[ 221], 50.00th=[ 223], 60.00th=[ 227], 00:42:29.807 | 70.00th=[ 229], 80.00th=[ 233], 90.00th=[ 243], 95.00th=[ 258], 00:42:29.807 | 99.00th=[41157], 99.50th=[41157], 99.90th=[41157], 99.95th=[41157], 00:42:29.807 | 99.99th=[41157] 00:42:29.807 write: IOPS=1019, BW=4080KiB/s (4178kB/s)(4096KiB/1004msec); 0 zone resets 00:42:29.807 slat (nsec): min=5552, max=30391, avg=6731.74, stdev=1552.50 00:42:29.807 clat (usec): min=146, max=267, avg=176.29, stdev=14.79 00:42:29.807 lat (usec): min=152, max=298, avg=183.02, stdev=15.15 00:42:29.807 clat percentiles (usec): 00:42:29.807 | 1.00th=[ 151], 5.00th=[ 157], 10.00th=[ 159], 20.00th=[ 163], 00:42:29.807 | 30.00th=[ 167], 40.00th=[ 169], 50.00th=[ 176], 60.00th=[ 180], 00:42:29.807 | 70.00th=[ 184], 80.00th=[ 190], 90.00th=[ 196], 95.00th=[ 202], 00:42:29.807 | 99.00th=[ 217], 99.50th=[ 225], 99.90th=[ 235], 99.95th=[ 269], 00:42:29.807 | 99.99th=[ 269] 00:42:29.807 bw ( KiB/s): min= 8192, max= 8192, per=69.40%, avg=8192.00, stdev= 0.00, samples=1 00:42:29.807 iops : min= 2048, max= 2048, avg=2048.00, stdev= 0.00, samples=1 00:42:29.807 lat (usec) : 250=97.81%, 500=1.10% 00:42:29.807 lat (msec) : 50=1.10% 00:42:29.807 cpu : usr=0.50%, sys=1.00%, ctx=1552, majf=0, minf=1 00:42:29.807 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:42:29.807 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:29.807 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:29.807 issued rwts: total=528,1024,0,0 short=0,0,0,0 dropped=0,0,0,0 00:42:29.807 latency : target=0, window=0, percentile=100.00%, depth=1 00:42:29.807 00:42:29.807 Run status group 0 (all jobs): 00:42:29.807 READ: bw=4269KiB/s (4371kB/s), 89.8KiB/s-2104KiB/s (91.9kB/s-2154kB/s), io=4444KiB (4551kB), run=1004-1041msec 00:42:29.807 WRITE: bw=11.5MiB/s (12.1MB/s), 1998KiB/s-4080KiB/s (2046kB/s-4178kB/s), io=12.0MiB (12.6MB), run=1004-1041msec 00:42:29.807 00:42:29.807 Disk stats (read/write): 00:42:29.807 nvme0n1: ios=50/512, merge=0/0, ticks=1687/100, in_queue=1787, util=97.70% 00:42:29.807 nvme0n2: ios=39/512, merge=0/0, ticks=770/98, in_queue=868, util=90.74% 00:42:29.807 nvme0n3: ios=526/1024, merge=0/0, ticks=619/189, in_queue=808, util=88.90% 00:42:29.807 nvme0n4: ios=524/1024, merge=0/0, ticks=644/176, in_queue=820, util=89.65% 00:42:29.807 17:58:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@51 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t randwrite -r 1 -v 00:42:30.065 [global] 00:42:30.065 thread=1 00:42:30.065 invalidate=1 00:42:30.065 rw=randwrite 00:42:30.065 time_based=1 00:42:30.065 runtime=1 00:42:30.065 ioengine=libaio 00:42:30.065 direct=1 00:42:30.065 bs=4096 00:42:30.065 iodepth=1 00:42:30.065 norandommap=0 00:42:30.065 numjobs=1 00:42:30.065 00:42:30.065 verify_dump=1 00:42:30.065 verify_backlog=512 00:42:30.065 verify_state_save=0 00:42:30.065 do_verify=1 00:42:30.065 verify=crc32c-intel 00:42:30.065 [job0] 00:42:30.065 filename=/dev/nvme0n1 00:42:30.065 [job1] 00:42:30.065 filename=/dev/nvme0n2 00:42:30.065 [job2] 00:42:30.065 filename=/dev/nvme0n3 00:42:30.065 [job3] 00:42:30.065 filename=/dev/nvme0n4 00:42:30.065 Could not set queue depth (nvme0n1) 00:42:30.065 Could not set queue depth (nvme0n2) 00:42:30.065 Could not set queue depth (nvme0n3) 00:42:30.065 Could not set queue depth (nvme0n4) 00:42:30.065 job0: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:42:30.065 job1: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:42:30.065 job2: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:42:30.065 job3: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:42:30.065 fio-3.35 00:42:30.065 Starting 4 threads 00:42:31.440 00:42:31.440 job0: (groupid=0, jobs=1): err= 0: pid=459635: Fri Dec 6 17:58:13 2024 00:42:31.440 read: IOPS=1965, BW=7860KiB/s (8049kB/s)(7868KiB/1001msec) 00:42:31.440 slat (nsec): min=4316, max=61120, avg=7642.44, stdev=4717.94 00:42:31.440 clat (usec): min=163, max=40789, avg=295.62, stdev=1292.09 00:42:31.440 lat (usec): min=168, max=40794, avg=303.26, stdev=1292.10 00:42:31.440 clat percentiles (usec): 00:42:31.440 | 1.00th=[ 192], 5.00th=[ 196], 10.00th=[ 202], 20.00th=[ 208], 00:42:31.440 | 30.00th=[ 223], 40.00th=[ 235], 50.00th=[ 243], 60.00th=[ 249], 00:42:31.440 | 70.00th=[ 255], 80.00th=[ 277], 90.00th=[ 347], 95.00th=[ 383], 00:42:31.440 | 99.00th=[ 412], 99.50th=[ 424], 99.90th=[40633], 99.95th=[40633], 00:42:31.440 | 99.99th=[40633] 00:42:31.440 write: IOPS=2045, BW=8184KiB/s (8380kB/s)(8192KiB/1001msec); 0 zone resets 00:42:31.440 slat (nsec): min=5517, max=37907, avg=9159.73, stdev=5047.23 00:42:31.440 clat (usec): min=132, max=483, avg=183.47, stdev=44.94 00:42:31.440 lat (usec): min=138, max=492, avg=192.62, stdev=47.69 00:42:31.440 clat percentiles (usec): 00:42:31.440 | 1.00th=[ 135], 5.00th=[ 139], 10.00th=[ 139], 20.00th=[ 143], 00:42:31.440 | 30.00th=[ 149], 40.00th=[ 153], 50.00th=[ 163], 60.00th=[ 188], 00:42:31.440 | 70.00th=[ 212], 80.00th=[ 231], 90.00th=[ 249], 95.00th=[ 260], 00:42:31.440 | 99.00th=[ 289], 99.50th=[ 314], 99.90th=[ 383], 99.95th=[ 412], 00:42:31.440 | 99.99th=[ 486] 00:42:31.440 bw ( KiB/s): min= 8192, max= 8192, per=46.13%, avg=8192.00, stdev= 0.00, samples=1 00:42:31.440 iops : min= 2048, max= 2048, avg=2048.00, stdev= 0.00, samples=1 00:42:31.440 lat (usec) : 250=77.41%, 500=22.49% 00:42:31.440 lat (msec) : 2=0.02%, 4=0.02%, 50=0.05% 00:42:31.440 cpu : usr=2.00%, sys=3.20%, ctx=4016, majf=0, minf=1 00:42:31.440 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:42:31.440 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:31.440 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:31.440 issued rwts: total=1967,2048,0,0 short=0,0,0,0 dropped=0,0,0,0 00:42:31.440 latency : target=0, window=0, percentile=100.00%, depth=1 00:42:31.440 job1: (groupid=0, jobs=1): err= 0: pid=459637: Fri Dec 6 17:58:13 2024 00:42:31.440 read: IOPS=513, BW=2054KiB/s (2103kB/s)(2132KiB/1038msec) 00:42:31.440 slat (nsec): min=4683, max=44588, avg=12871.16, stdev=6405.06 00:42:31.440 clat (usec): min=236, max=41138, avg=1520.67, stdev=6720.98 00:42:31.440 lat (usec): min=242, max=41146, avg=1533.54, stdev=6721.26 00:42:31.440 clat percentiles (usec): 00:42:31.440 | 1.00th=[ 249], 5.00th=[ 265], 10.00th=[ 277], 20.00th=[ 310], 00:42:31.440 | 30.00th=[ 334], 40.00th=[ 363], 50.00th=[ 388], 60.00th=[ 396], 00:42:31.440 | 70.00th=[ 412], 80.00th=[ 457], 90.00th=[ 498], 95.00th=[ 519], 00:42:31.440 | 99.00th=[41157], 99.50th=[41157], 99.90th=[41157], 99.95th=[41157], 00:42:31.440 | 99.99th=[41157] 00:42:31.440 write: IOPS=986, BW=3946KiB/s (4041kB/s)(4096KiB/1038msec); 0 zone resets 00:42:31.440 slat (nsec): min=5719, max=45865, avg=10234.93, stdev=5270.72 00:42:31.440 clat (usec): min=135, max=733, avg=200.44, stdev=34.96 00:42:31.440 lat (usec): min=142, max=748, avg=210.67, stdev=35.19 00:42:31.440 clat percentiles (usec): 00:42:31.440 | 1.00th=[ 139], 5.00th=[ 147], 10.00th=[ 159], 20.00th=[ 178], 00:42:31.440 | 30.00th=[ 188], 40.00th=[ 196], 50.00th=[ 200], 60.00th=[ 206], 00:42:31.440 | 70.00th=[ 212], 80.00th=[ 221], 90.00th=[ 241], 95.00th=[ 247], 00:42:31.440 | 99.00th=[ 277], 99.50th=[ 322], 99.90th=[ 429], 99.95th=[ 734], 00:42:31.440 | 99.99th=[ 734] 00:42:31.440 bw ( KiB/s): min= 8192, max= 8192, per=46.13%, avg=8192.00, stdev= 0.00, samples=1 00:42:31.440 iops : min= 2048, max= 2048, avg=2048.00, stdev= 0.00, samples=1 00:42:31.440 lat (usec) : 250=63.71%, 500=33.14%, 750=2.18% 00:42:31.440 lat (msec) : 50=0.96% 00:42:31.440 cpu : usr=0.87%, sys=1.74%, ctx=1557, majf=0, minf=1 00:42:31.440 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:42:31.440 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:31.440 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:31.440 issued rwts: total=533,1024,0,0 short=0,0,0,0 dropped=0,0,0,0 00:42:31.440 latency : target=0, window=0, percentile=100.00%, depth=1 00:42:31.440 job2: (groupid=0, jobs=1): err= 0: pid=459638: Fri Dec 6 17:58:13 2024 00:42:31.440 read: IOPS=927, BW=3708KiB/s (3797kB/s)(3712KiB/1001msec) 00:42:31.440 slat (nsec): min=5956, max=38687, avg=8889.52, stdev=4253.52 00:42:31.440 clat (usec): min=227, max=41083, avg=797.95, stdev=4403.99 00:42:31.440 lat (usec): min=233, max=41095, avg=806.84, stdev=4404.49 00:42:31.440 clat percentiles (usec): 00:42:31.440 | 1.00th=[ 237], 5.00th=[ 243], 10.00th=[ 251], 20.00th=[ 265], 00:42:31.440 | 30.00th=[ 285], 40.00th=[ 293], 50.00th=[ 306], 60.00th=[ 314], 00:42:31.440 | 70.00th=[ 330], 80.00th=[ 343], 90.00th=[ 367], 95.00th=[ 396], 00:42:31.440 | 99.00th=[41157], 99.50th=[41157], 99.90th=[41157], 99.95th=[41157], 00:42:31.440 | 99.99th=[41157] 00:42:31.440 write: IOPS=1022, BW=4092KiB/s (4190kB/s)(4096KiB/1001msec); 0 zone resets 00:42:31.440 slat (nsec): min=7113, max=53371, avg=13907.75, stdev=7613.39 00:42:31.440 clat (usec): min=162, max=484, avg=225.42, stdev=40.74 00:42:31.440 lat (usec): min=169, max=493, avg=239.33, stdev=44.80 00:42:31.440 clat percentiles (usec): 00:42:31.440 | 1.00th=[ 167], 5.00th=[ 172], 10.00th=[ 176], 20.00th=[ 182], 00:42:31.440 | 30.00th=[ 192], 40.00th=[ 210], 50.00th=[ 225], 60.00th=[ 241], 00:42:31.440 | 70.00th=[ 255], 80.00th=[ 265], 90.00th=[ 273], 95.00th=[ 285], 00:42:31.440 | 99.00th=[ 306], 99.50th=[ 355], 99.90th=[ 404], 99.95th=[ 486], 00:42:31.440 | 99.99th=[ 486] 00:42:31.440 bw ( KiB/s): min= 8192, max= 8192, per=46.13%, avg=8192.00, stdev= 0.00, samples=1 00:42:31.440 iops : min= 2048, max= 2048, avg=2048.00, stdev= 0.00, samples=1 00:42:31.440 lat (usec) : 250=39.91%, 500=59.12%, 750=0.31% 00:42:31.440 lat (msec) : 4=0.05%, 10=0.05%, 50=0.56% 00:42:31.440 cpu : usr=1.60%, sys=3.00%, ctx=1952, majf=0, minf=1 00:42:31.440 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:42:31.440 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:31.440 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:31.440 issued rwts: total=928,1024,0,0 short=0,0,0,0 dropped=0,0,0,0 00:42:31.440 latency : target=0, window=0, percentile=100.00%, depth=1 00:42:31.440 job3: (groupid=0, jobs=1): err= 0: pid=459639: Fri Dec 6 17:58:13 2024 00:42:31.440 read: IOPS=33, BW=135KiB/s (139kB/s)(140KiB/1035msec) 00:42:31.440 slat (nsec): min=5940, max=17311, avg=12930.89, stdev=2480.92 00:42:31.440 clat (usec): min=228, max=42060, avg=25522.33, stdev=20018.98 00:42:31.440 lat (usec): min=235, max=42074, avg=25535.26, stdev=20019.15 00:42:31.440 clat percentiles (usec): 00:42:31.440 | 1.00th=[ 229], 5.00th=[ 239], 10.00th=[ 343], 20.00th=[ 429], 00:42:31.440 | 30.00th=[ 603], 40.00th=[16450], 50.00th=[41157], 60.00th=[41157], 00:42:31.440 | 70.00th=[41681], 80.00th=[41681], 90.00th=[42206], 95.00th=[42206], 00:42:31.440 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:42:31.440 | 99.99th=[42206] 00:42:31.440 write: IOPS=494, BW=1979KiB/s (2026kB/s)(2048KiB/1035msec); 0 zone resets 00:42:31.440 slat (nsec): min=6413, max=41098, avg=14986.94, stdev=6363.47 00:42:31.440 clat (usec): min=192, max=595, avg=257.05, stdev=32.58 00:42:31.440 lat (usec): min=200, max=601, avg=272.04, stdev=33.20 00:42:31.440 clat percentiles (usec): 00:42:31.440 | 1.00th=[ 206], 5.00th=[ 219], 10.00th=[ 227], 20.00th=[ 235], 00:42:31.440 | 30.00th=[ 241], 40.00th=[ 247], 50.00th=[ 255], 60.00th=[ 260], 00:42:31.440 | 70.00th=[ 265], 80.00th=[ 273], 90.00th=[ 289], 95.00th=[ 302], 00:42:31.440 | 99.00th=[ 392], 99.50th=[ 412], 99.90th=[ 594], 99.95th=[ 594], 00:42:31.440 | 99.99th=[ 594] 00:42:31.440 bw ( KiB/s): min= 4096, max= 4096, per=23.07%, avg=4096.00, stdev= 0.00, samples=1 00:42:31.440 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:42:31.440 lat (usec) : 250=41.32%, 500=53.75%, 750=0.91% 00:42:31.440 lat (msec) : 20=0.18%, 50=3.84% 00:42:31.440 cpu : usr=0.39%, sys=0.77%, ctx=547, majf=0, minf=1 00:42:31.440 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:42:31.440 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:31.440 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:31.440 issued rwts: total=35,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:42:31.440 latency : target=0, window=0, percentile=100.00%, depth=1 00:42:31.440 00:42:31.440 Run status group 0 (all jobs): 00:42:31.440 READ: bw=13.0MiB/s (13.7MB/s), 135KiB/s-7860KiB/s (139kB/s-8049kB/s), io=13.5MiB (14.2MB), run=1001-1038msec 00:42:31.440 WRITE: bw=17.3MiB/s (18.2MB/s), 1979KiB/s-8184KiB/s (2026kB/s-8380kB/s), io=18.0MiB (18.9MB), run=1001-1038msec 00:42:31.440 00:42:31.440 Disk stats (read/write): 00:42:31.440 nvme0n1: ios=1588/1834, merge=0/0, ticks=813/336, in_queue=1149, util=98.10% 00:42:31.440 nvme0n2: ios=546/1024, merge=0/0, ticks=657/202, in_queue=859, util=90.96% 00:42:31.440 nvme0n3: ios=523/1024, merge=0/0, ticks=561/227, in_queue=788, util=88.95% 00:42:31.440 nvme0n4: ios=24/512, merge=0/0, ticks=686/127, in_queue=813, util=89.60% 00:42:31.440 17:58:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 128 -t write -r 1 -v 00:42:31.441 [global] 00:42:31.441 thread=1 00:42:31.441 invalidate=1 00:42:31.441 rw=write 00:42:31.441 time_based=1 00:42:31.441 runtime=1 00:42:31.441 ioengine=libaio 00:42:31.441 direct=1 00:42:31.441 bs=4096 00:42:31.441 iodepth=128 00:42:31.441 norandommap=0 00:42:31.441 numjobs=1 00:42:31.441 00:42:31.441 verify_dump=1 00:42:31.441 verify_backlog=512 00:42:31.441 verify_state_save=0 00:42:31.441 do_verify=1 00:42:31.441 verify=crc32c-intel 00:42:31.441 [job0] 00:42:31.441 filename=/dev/nvme0n1 00:42:31.441 [job1] 00:42:31.441 filename=/dev/nvme0n2 00:42:31.441 [job2] 00:42:31.441 filename=/dev/nvme0n3 00:42:31.441 [job3] 00:42:31.441 filename=/dev/nvme0n4 00:42:31.441 Could not set queue depth (nvme0n1) 00:42:31.441 Could not set queue depth (nvme0n2) 00:42:31.441 Could not set queue depth (nvme0n3) 00:42:31.441 Could not set queue depth (nvme0n4) 00:42:31.699 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:42:31.699 job1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:42:31.699 job2: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:42:31.699 job3: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:42:31.699 fio-3.35 00:42:31.699 Starting 4 threads 00:42:33.075 00:42:33.075 job0: (groupid=0, jobs=1): err= 0: pid=459865: Fri Dec 6 17:58:14 2024 00:42:33.075 read: IOPS=1016, BW=4068KiB/s (4165kB/s)(4096KiB/1007msec) 00:42:33.075 slat (usec): min=4, max=55455, avg=538.23, stdev=3960.49 00:42:33.075 clat (msec): min=13, max=144, avg=65.63, stdev=48.24 00:42:33.075 lat (msec): min=14, max=144, avg=66.16, stdev=48.52 00:42:33.075 clat percentiles (msec): 00:42:33.075 | 1.00th=[ 16], 5.00th=[ 17], 10.00th=[ 17], 20.00th=[ 18], 00:42:33.075 | 30.00th=[ 19], 40.00th=[ 20], 50.00th=[ 55], 60.00th=[ 91], 00:42:33.075 | 70.00th=[ 113], 80.00th=[ 120], 90.00th=[ 132], 95.00th=[ 142], 00:42:33.075 | 99.00th=[ 146], 99.50th=[ 146], 99.90th=[ 146], 99.95th=[ 146], 00:42:33.075 | 99.99th=[ 146] 00:42:33.075 write: IOPS=1183, BW=4735KiB/s (4848kB/s)(4768KiB/1007msec); 0 zone resets 00:42:33.075 slat (usec): min=4, max=24964, avg=366.58, stdev=2025.84 00:42:33.075 clat (msec): min=3, max=129, avg=50.40, stdev=28.52 00:42:33.075 lat (msec): min=7, max=140, avg=50.77, stdev=28.55 00:42:33.075 clat percentiles (msec): 00:42:33.075 | 1.00th=[ 13], 5.00th=[ 17], 10.00th=[ 17], 20.00th=[ 22], 00:42:33.075 | 30.00th=[ 35], 40.00th=[ 35], 50.00th=[ 45], 60.00th=[ 57], 00:42:33.075 | 70.00th=[ 63], 80.00th=[ 74], 90.00th=[ 93], 95.00th=[ 109], 00:42:33.075 | 99.00th=[ 130], 99.50th=[ 130], 99.90th=[ 130], 99.95th=[ 130], 00:42:33.075 | 99.99th=[ 130] 00:42:33.075 bw ( KiB/s): min= 4096, max= 4416, per=7.76%, avg=4256.00, stdev=226.27, samples=2 00:42:33.075 iops : min= 1024, max= 1104, avg=1064.00, stdev=56.57, samples=2 00:42:33.075 lat (msec) : 4=0.05%, 10=0.45%, 20=28.29%, 50=21.84%, 100=29.38% 00:42:33.075 lat (msec) : 250=19.99% 00:42:33.075 cpu : usr=2.09%, sys=2.68%, ctx=95, majf=0, minf=1 00:42:33.075 IO depths : 1=0.1%, 2=0.1%, 4=0.2%, 8=0.4%, 16=0.7%, 32=1.4%, >=64=97.2% 00:42:33.075 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:33.075 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:42:33.075 issued rwts: total=1024,1192,0,0 short=0,0,0,0 dropped=0,0,0,0 00:42:33.075 latency : target=0, window=0, percentile=100.00%, depth=128 00:42:33.075 job1: (groupid=0, jobs=1): err= 0: pid=459872: Fri Dec 6 17:58:14 2024 00:42:33.075 read: IOPS=4486, BW=17.5MiB/s (18.4MB/s)(17.6MiB/1006msec) 00:42:33.075 slat (usec): min=2, max=40108, avg=105.80, stdev=1112.12 00:42:33.075 clat (usec): min=964, max=67256, avg=15024.86, stdev=9453.93 00:42:33.075 lat (usec): min=1133, max=67291, avg=15130.66, stdev=9533.15 00:42:33.075 clat percentiles (usec): 00:42:33.075 | 1.00th=[ 4490], 5.00th=[ 5800], 10.00th=[ 7242], 20.00th=[ 8586], 00:42:33.075 | 30.00th=[ 9634], 40.00th=[10421], 50.00th=[11207], 60.00th=[13566], 00:42:33.075 | 70.00th=[14877], 80.00th=[19006], 90.00th=[28967], 95.00th=[36963], 00:42:33.076 | 99.00th=[43779], 99.50th=[43779], 99.90th=[45876], 99.95th=[51119], 00:42:33.076 | 99.99th=[67634] 00:42:33.076 write: IOPS=4580, BW=17.9MiB/s (18.8MB/s)(18.0MiB/1006msec); 0 zone resets 00:42:33.076 slat (usec): min=3, max=48687, avg=94.51, stdev=1035.42 00:42:33.076 clat (usec): min=757, max=50351, avg=11065.52, stdev=4506.39 00:42:33.076 lat (usec): min=769, max=71141, avg=11160.03, stdev=4656.90 00:42:33.076 clat percentiles (usec): 00:42:33.076 | 1.00th=[ 3195], 5.00th=[ 5538], 10.00th=[ 6456], 20.00th=[ 7963], 00:42:33.076 | 30.00th=[ 9503], 40.00th=[ 9634], 50.00th=[10159], 60.00th=[10421], 00:42:33.076 | 70.00th=[11469], 80.00th=[13566], 90.00th=[17433], 95.00th=[20317], 00:42:33.076 | 99.00th=[27395], 99.50th=[28705], 99.90th=[29754], 99.95th=[31851], 00:42:33.076 | 99.99th=[50594] 00:42:33.076 bw ( KiB/s): min=18424, max=18440, per=33.60%, avg=18432.00, stdev=11.31, samples=2 00:42:33.076 iops : min= 4606, max= 4610, avg=4608.00, stdev= 2.83, samples=2 00:42:33.076 lat (usec) : 1000=0.08% 00:42:33.076 lat (msec) : 2=0.10%, 4=0.96%, 10=39.51%, 20=46.74%, 50=12.56% 00:42:33.076 lat (msec) : 100=0.04% 00:42:33.076 cpu : usr=3.98%, sys=8.26%, ctx=267, majf=0, minf=1 00:42:33.076 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.4%, >=64=99.3% 00:42:33.076 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:33.076 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:42:33.076 issued rwts: total=4513,4608,0,0 short=0,0,0,0 dropped=0,0,0,0 00:42:33.076 latency : target=0, window=0, percentile=100.00%, depth=128 00:42:33.076 job2: (groupid=0, jobs=1): err= 0: pid=459893: Fri Dec 6 17:58:14 2024 00:42:33.076 read: IOPS=4571, BW=17.9MiB/s (18.7MB/s)(18.0MiB/1008msec) 00:42:33.076 slat (usec): min=2, max=8412, avg=98.35, stdev=558.87 00:42:33.076 clat (usec): min=7459, max=25448, avg=12841.69, stdev=2771.44 00:42:33.076 lat (usec): min=7464, max=28743, avg=12940.04, stdev=2816.38 00:42:33.076 clat percentiles (usec): 00:42:33.076 | 1.00th=[ 8586], 5.00th=[10028], 10.00th=[10159], 20.00th=[10290], 00:42:33.076 | 30.00th=[10945], 40.00th=[11469], 50.00th=[12256], 60.00th=[13042], 00:42:33.076 | 70.00th=[13829], 80.00th=[14353], 90.00th=[16909], 95.00th=[17695], 00:42:33.076 | 99.00th=[22152], 99.50th=[22414], 99.90th=[24773], 99.95th=[25035], 00:42:33.076 | 99.99th=[25560] 00:42:33.076 write: IOPS=4935, BW=19.3MiB/s (20.2MB/s)(19.4MiB/1008msec); 0 zone resets 00:42:33.076 slat (usec): min=3, max=19179, avg=100.19, stdev=607.01 00:42:33.076 clat (usec): min=5519, max=45686, avg=13745.55, stdev=6762.64 00:42:33.076 lat (usec): min=7014, max=45720, avg=13845.74, stdev=6807.52 00:42:33.076 clat percentiles (usec): 00:42:33.076 | 1.00th=[ 7504], 5.00th=[ 7898], 10.00th=[ 8356], 20.00th=[10421], 00:42:33.076 | 30.00th=[10814], 40.00th=[11338], 50.00th=[12125], 60.00th=[12780], 00:42:33.076 | 70.00th=[13173], 80.00th=[13566], 90.00th=[20579], 95.00th=[33817], 00:42:33.076 | 99.00th=[37487], 99.50th=[37487], 99.90th=[42206], 99.95th=[42206], 00:42:33.076 | 99.99th=[45876] 00:42:33.076 bw ( KiB/s): min=18264, max=20512, per=35.34%, avg=19388.00, stdev=1589.58, samples=2 00:42:33.076 iops : min= 4566, max= 5128, avg=4847.00, stdev=397.39, samples=2 00:42:33.076 lat (msec) : 10=9.61%, 20=83.85%, 50=6.54% 00:42:33.076 cpu : usr=6.26%, sys=10.33%, ctx=383, majf=0, minf=1 00:42:33.076 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.3%, >=64=99.3% 00:42:33.076 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:33.076 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:42:33.076 issued rwts: total=4608,4975,0,0 short=0,0,0,0 dropped=0,0,0,0 00:42:33.076 latency : target=0, window=0, percentile=100.00%, depth=128 00:42:33.076 job3: (groupid=0, jobs=1): err= 0: pid=459903: Fri Dec 6 17:58:14 2024 00:42:33.076 read: IOPS=3117, BW=12.2MiB/s (12.8MB/s)(12.8MiB/1047msec) 00:42:33.076 slat (usec): min=3, max=26956, avg=140.44, stdev=1166.98 00:42:33.076 clat (usec): min=7455, max=55894, avg=20278.28, stdev=9908.02 00:42:33.076 lat (usec): min=7466, max=64171, avg=20418.73, stdev=9978.06 00:42:33.076 clat percentiles (usec): 00:42:33.076 | 1.00th=[ 9110], 5.00th=[10290], 10.00th=[11207], 20.00th=[13435], 00:42:33.076 | 30.00th=[13960], 40.00th=[14353], 50.00th=[17695], 60.00th=[19792], 00:42:33.076 | 70.00th=[22152], 80.00th=[27395], 90.00th=[32637], 95.00th=[40633], 00:42:33.076 | 99.00th=[55837], 99.50th=[55837], 99.90th=[55837], 99.95th=[55837], 00:42:33.076 | 99.99th=[55837] 00:42:33.076 write: IOPS=3423, BW=13.4MiB/s (14.0MB/s)(14.0MiB/1047msec); 0 zone resets 00:42:33.076 slat (usec): min=4, max=20580, avg=135.90, stdev=1013.22 00:42:33.076 clat (usec): min=3272, max=67204, avg=18458.61, stdev=10801.90 00:42:33.076 lat (usec): min=3281, max=67230, avg=18594.52, stdev=10899.57 00:42:33.076 clat percentiles (usec): 00:42:33.076 | 1.00th=[ 7046], 5.00th=[ 8979], 10.00th=[ 9372], 20.00th=[10421], 00:42:33.076 | 30.00th=[12256], 40.00th=[13173], 50.00th=[14877], 60.00th=[17171], 00:42:33.076 | 70.00th=[20317], 80.00th=[23200], 90.00th=[33817], 95.00th=[37487], 00:42:33.076 | 99.00th=[62129], 99.50th=[65274], 99.90th=[67634], 99.95th=[67634], 00:42:33.076 | 99.99th=[67634] 00:42:33.076 bw ( KiB/s): min=12288, max=16384, per=26.13%, avg=14336.00, stdev=2896.31, samples=2 00:42:33.076 iops : min= 3072, max= 4096, avg=3584.00, stdev=724.08, samples=2 00:42:33.076 lat (msec) : 4=0.12%, 10=9.11%, 20=56.75%, 50=30.99%, 100=3.04% 00:42:33.076 cpu : usr=5.54%, sys=8.03%, ctx=175, majf=0, minf=1 00:42:33.076 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.5%, >=64=99.1% 00:42:33.076 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:33.076 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:42:33.076 issued rwts: total=3264,3584,0,0 short=0,0,0,0 dropped=0,0,0,0 00:42:33.076 latency : target=0, window=0, percentile=100.00%, depth=128 00:42:33.076 00:42:33.076 Run status group 0 (all jobs): 00:42:33.076 READ: bw=50.0MiB/s (52.5MB/s), 4068KiB/s-17.9MiB/s (4165kB/s-18.7MB/s), io=52.4MiB (54.9MB), run=1006-1047msec 00:42:33.076 WRITE: bw=53.6MiB/s (56.2MB/s), 4735KiB/s-19.3MiB/s (4848kB/s-20.2MB/s), io=56.1MiB (58.8MB), run=1006-1047msec 00:42:33.076 00:42:33.076 Disk stats (read/write): 00:42:33.076 nvme0n1: ios=716/1024, merge=0/0, ticks=16484/10421, in_queue=26905, util=99.80% 00:42:33.076 nvme0n2: ios=3635/4011, merge=0/0, ticks=54884/40045, in_queue=94929, util=96.54% 00:42:33.076 nvme0n3: ios=4156/4543, merge=0/0, ticks=20672/20664, in_queue=41336, util=97.81% 00:42:33.076 nvme0n4: ios=2593/2879, merge=0/0, ticks=49829/51899, in_queue=101728, util=100.00% 00:42:33.076 17:58:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 128 -t randwrite -r 1 -v 00:42:33.076 [global] 00:42:33.076 thread=1 00:42:33.076 invalidate=1 00:42:33.076 rw=randwrite 00:42:33.076 time_based=1 00:42:33.076 runtime=1 00:42:33.076 ioengine=libaio 00:42:33.076 direct=1 00:42:33.076 bs=4096 00:42:33.076 iodepth=128 00:42:33.076 norandommap=0 00:42:33.076 numjobs=1 00:42:33.076 00:42:33.076 verify_dump=1 00:42:33.076 verify_backlog=512 00:42:33.076 verify_state_save=0 00:42:33.076 do_verify=1 00:42:33.076 verify=crc32c-intel 00:42:33.076 [job0] 00:42:33.076 filename=/dev/nvme0n1 00:42:33.076 [job1] 00:42:33.076 filename=/dev/nvme0n2 00:42:33.076 [job2] 00:42:33.076 filename=/dev/nvme0n3 00:42:33.076 [job3] 00:42:33.076 filename=/dev/nvme0n4 00:42:33.076 Could not set queue depth (nvme0n1) 00:42:33.076 Could not set queue depth (nvme0n2) 00:42:33.076 Could not set queue depth (nvme0n3) 00:42:33.076 Could not set queue depth (nvme0n4) 00:42:33.076 job0: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:42:33.076 job1: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:42:33.076 job2: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:42:33.076 job3: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:42:33.076 fio-3.35 00:42:33.076 Starting 4 threads 00:42:34.450 00:42:34.450 job0: (groupid=0, jobs=1): err= 0: pid=460215: Fri Dec 6 17:58:16 2024 00:42:34.450 read: IOPS=4095, BW=16.0MiB/s (16.8MB/s)(16.1MiB/1006msec) 00:42:34.450 slat (usec): min=2, max=11194, avg=106.43, stdev=706.27 00:42:34.450 clat (usec): min=1046, max=40706, avg=13630.79, stdev=4496.34 00:42:34.450 lat (usec): min=3995, max=40720, avg=13737.22, stdev=4560.39 00:42:34.450 clat percentiles (usec): 00:42:34.450 | 1.00th=[ 5342], 5.00th=[ 7701], 10.00th=[ 9503], 20.00th=[10683], 00:42:34.450 | 30.00th=[11207], 40.00th=[11600], 50.00th=[12649], 60.00th=[13173], 00:42:34.450 | 70.00th=[14353], 80.00th=[17171], 90.00th=[19792], 95.00th=[21627], 00:42:34.450 | 99.00th=[27132], 99.50th=[32113], 99.90th=[36963], 99.95th=[36963], 00:42:34.450 | 99.99th=[40633] 00:42:34.450 write: IOPS=4580, BW=17.9MiB/s (18.8MB/s)(18.0MiB/1006msec); 0 zone resets 00:42:34.450 slat (usec): min=3, max=22699, avg=113.89, stdev=754.50 00:42:34.450 clat (usec): min=5707, max=49716, avg=15421.82, stdev=7329.24 00:42:34.450 lat (usec): min=6293, max=49734, avg=15535.71, stdev=7381.69 00:42:34.450 clat percentiles (usec): 00:42:34.450 | 1.00th=[ 7046], 5.00th=[ 7767], 10.00th=[ 9503], 20.00th=[11338], 00:42:34.450 | 30.00th=[12125], 40.00th=[12518], 50.00th=[13173], 60.00th=[14746], 00:42:34.450 | 70.00th=[17171], 80.00th=[17957], 90.00th=[20055], 95.00th=[34341], 00:42:34.450 | 99.00th=[46924], 99.50th=[47973], 99.90th=[48497], 99.95th=[48497], 00:42:34.450 | 99.99th=[49546] 00:42:34.450 bw ( KiB/s): min=17176, max=18893, per=29.18%, avg=18034.50, stdev=1214.10, samples=2 00:42:34.450 iops : min= 4294, max= 4723, avg=4508.50, stdev=303.35, samples=2 00:42:34.450 lat (msec) : 2=0.01%, 4=0.19%, 10=12.60%, 20=77.11%, 50=10.08% 00:42:34.450 cpu : usr=3.18%, sys=5.47%, ctx=339, majf=0, minf=1 00:42:34.450 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.4%, >=64=99.3% 00:42:34.450 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:34.450 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:42:34.450 issued rwts: total=4120,4608,0,0 short=0,0,0,0 dropped=0,0,0,0 00:42:34.450 latency : target=0, window=0, percentile=100.00%, depth=128 00:42:34.450 job1: (groupid=0, jobs=1): err= 0: pid=460216: Fri Dec 6 17:58:16 2024 00:42:34.450 read: IOPS=4087, BW=16.0MiB/s (16.7MB/s)(16.0MiB/1002msec) 00:42:34.450 slat (usec): min=2, max=19341, avg=112.81, stdev=699.15 00:42:34.450 clat (usec): min=6280, max=54443, avg=14439.13, stdev=6632.19 00:42:34.450 lat (usec): min=6284, max=54448, avg=14551.94, stdev=6673.06 00:42:34.450 clat percentiles (usec): 00:42:34.450 | 1.00th=[ 6718], 5.00th=[ 9372], 10.00th=[10421], 20.00th=[11207], 00:42:34.450 | 30.00th=[11731], 40.00th=[11994], 50.00th=[12256], 60.00th=[12518], 00:42:34.450 | 70.00th=[13304], 80.00th=[16581], 90.00th=[20317], 95.00th=[26346], 00:42:34.450 | 99.00th=[48497], 99.50th=[48497], 99.90th=[51119], 99.95th=[54264], 00:42:34.450 | 99.99th=[54264] 00:42:34.450 write: IOPS=4328, BW=16.9MiB/s (17.7MB/s)(16.9MiB/1002msec); 0 zone resets 00:42:34.450 slat (usec): min=3, max=5752, avg=117.61, stdev=587.44 00:42:34.450 clat (usec): min=1171, max=43451, avg=15558.35, stdev=6862.44 00:42:34.450 lat (usec): min=1182, max=43459, avg=15675.95, stdev=6913.00 00:42:34.450 clat percentiles (usec): 00:42:34.450 | 1.00th=[ 4555], 5.00th=[ 9372], 10.00th=[10290], 20.00th=[11469], 00:42:34.450 | 30.00th=[11731], 40.00th=[11863], 50.00th=[12387], 60.00th=[14484], 00:42:34.450 | 70.00th=[17433], 80.00th=[17957], 90.00th=[29754], 95.00th=[33424], 00:42:34.450 | 99.00th=[33817], 99.50th=[34341], 99.90th=[43254], 99.95th=[43254], 00:42:34.450 | 99.99th=[43254] 00:42:34.450 bw ( KiB/s): min=16512, max=17168, per=27.25%, avg=16840.00, stdev=463.86, samples=2 00:42:34.450 iops : min= 4128, max= 4292, avg=4210.00, stdev=115.97, samples=2 00:42:34.450 lat (msec) : 2=0.19%, 4=0.02%, 10=7.80%, 20=79.82%, 50=11.94% 00:42:34.450 lat (msec) : 100=0.23% 00:42:34.450 cpu : usr=3.00%, sys=5.79%, ctx=453, majf=0, minf=2 00:42:34.450 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.4%, >=64=99.3% 00:42:34.450 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:34.450 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:42:34.450 issued rwts: total=4096,4337,0,0 short=0,0,0,0 dropped=0,0,0,0 00:42:34.450 latency : target=0, window=0, percentile=100.00%, depth=128 00:42:34.450 job2: (groupid=0, jobs=1): err= 0: pid=460217: Fri Dec 6 17:58:16 2024 00:42:34.450 read: IOPS=2959, BW=11.6MiB/s (12.1MB/s)(11.6MiB/1006msec) 00:42:34.450 slat (usec): min=2, max=11921, avg=134.17, stdev=802.31 00:42:34.450 clat (usec): min=1455, max=41149, avg=16339.36, stdev=5596.62 00:42:34.450 lat (usec): min=7544, max=41187, avg=16473.53, stdev=5662.26 00:42:34.450 clat percentiles (usec): 00:42:34.450 | 1.00th=[ 8717], 5.00th=[11076], 10.00th=[12256], 20.00th=[13173], 00:42:34.450 | 30.00th=[13566], 40.00th=[13960], 50.00th=[14484], 60.00th=[15401], 00:42:34.450 | 70.00th=[15795], 80.00th=[17433], 90.00th=[25822], 95.00th=[31589], 00:42:34.450 | 99.00th=[31851], 99.50th=[31851], 99.90th=[40633], 99.95th=[41157], 00:42:34.450 | 99.99th=[41157] 00:42:34.450 write: IOPS=3053, BW=11.9MiB/s (12.5MB/s)(12.0MiB/1006msec); 0 zone resets 00:42:34.450 slat (usec): min=3, max=17473, avg=191.75, stdev=1046.90 00:42:34.450 clat (usec): min=7852, max=69682, avg=25552.38, stdev=13427.85 00:42:34.450 lat (usec): min=7870, max=69688, avg=25744.12, stdev=13504.75 00:42:34.450 clat percentiles (usec): 00:42:34.450 | 1.00th=[10290], 5.00th=[12649], 10.00th=[13566], 20.00th=[15270], 00:42:34.450 | 30.00th=[17695], 40.00th=[20841], 50.00th=[22152], 60.00th=[23462], 00:42:34.450 | 70.00th=[25560], 80.00th=[30016], 90.00th=[49021], 95.00th=[58459], 00:42:34.450 | 99.00th=[67634], 99.50th=[69731], 99.90th=[69731], 99.95th=[69731], 00:42:34.450 | 99.99th=[69731] 00:42:34.450 bw ( KiB/s): min=11208, max=13368, per=19.88%, avg=12288.00, stdev=1527.35, samples=2 00:42:34.450 iops : min= 2802, max= 3342, avg=3072.00, stdev=381.84, samples=2 00:42:34.450 lat (msec) : 2=0.02%, 10=1.42%, 20=59.17%, 50=34.60%, 100=4.79% 00:42:34.450 cpu : usr=2.09%, sys=2.79%, ctx=315, majf=0, minf=1 00:42:34.450 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.3%, 32=0.5%, >=64=99.0% 00:42:34.450 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:34.450 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:42:34.450 issued rwts: total=2977,3072,0,0 short=0,0,0,0 dropped=0,0,0,0 00:42:34.450 latency : target=0, window=0, percentile=100.00%, depth=128 00:42:34.450 job3: (groupid=0, jobs=1): err= 0: pid=460218: Fri Dec 6 17:58:16 2024 00:42:34.450 read: IOPS=3062, BW=12.0MiB/s (12.5MB/s)(12.0MiB/1003msec) 00:42:34.450 slat (usec): min=3, max=15005, avg=136.90, stdev=835.24 00:42:34.450 clat (usec): min=5333, max=60687, avg=15630.97, stdev=7142.11 00:42:34.450 lat (usec): min=5341, max=65702, avg=15767.87, stdev=7228.88 00:42:34.450 clat percentiles (usec): 00:42:34.450 | 1.00th=[ 6980], 5.00th=[11469], 10.00th=[11731], 20.00th=[11863], 00:42:34.450 | 30.00th=[12256], 40.00th=[12518], 50.00th=[12780], 60.00th=[13435], 00:42:34.450 | 70.00th=[14222], 80.00th=[17171], 90.00th=[23462], 95.00th=[29492], 00:42:34.450 | 99.00th=[51643], 99.50th=[58459], 99.90th=[60556], 99.95th=[60556], 00:42:34.450 | 99.99th=[60556] 00:42:34.450 write: IOPS=3516, BW=13.7MiB/s (14.4MB/s)(13.8MiB/1003msec); 0 zone resets 00:42:34.450 slat (usec): min=4, max=7510, avg=154.81, stdev=664.14 00:42:34.450 clat (usec): min=326, max=65689, avg=22448.12, stdev=14908.81 00:42:34.450 lat (usec): min=347, max=65697, avg=22602.93, stdev=14993.66 00:42:34.450 clat percentiles (usec): 00:42:34.450 | 1.00th=[ 2507], 5.00th=[ 5145], 10.00th=[ 7963], 20.00th=[11863], 00:42:34.450 | 30.00th=[12387], 40.00th=[13304], 50.00th=[16450], 60.00th=[22152], 00:42:34.450 | 70.00th=[25822], 80.00th=[35390], 90.00th=[47973], 95.00th=[54264], 00:42:34.450 | 99.00th=[60031], 99.50th=[61604], 99.90th=[63701], 99.95th=[63701], 00:42:34.450 | 99.99th=[65799] 00:42:34.450 bw ( KiB/s): min=11376, max=15816, per=22.00%, avg=13596.00, stdev=3139.55, samples=2 00:42:34.450 iops : min= 2844, max= 3954, avg=3399.00, stdev=784.89, samples=2 00:42:34.450 lat (usec) : 500=0.05% 00:42:34.450 lat (msec) : 2=0.27%, 4=1.27%, 10=8.06%, 20=57.72%, 50=27.90% 00:42:34.450 lat (msec) : 100=4.73% 00:42:34.450 cpu : usr=3.39%, sys=6.49%, ctx=384, majf=0, minf=2 00:42:34.450 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.5%, >=64=99.0% 00:42:34.450 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:34.450 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:42:34.450 issued rwts: total=3072,3527,0,0 short=0,0,0,0 dropped=0,0,0,0 00:42:34.450 latency : target=0, window=0, percentile=100.00%, depth=128 00:42:34.450 00:42:34.450 Run status group 0 (all jobs): 00:42:34.450 READ: bw=55.4MiB/s (58.1MB/s), 11.6MiB/s-16.0MiB/s (12.1MB/s-16.8MB/s), io=55.7MiB (58.4MB), run=1002-1006msec 00:42:34.451 WRITE: bw=60.4MiB/s (63.3MB/s), 11.9MiB/s-17.9MiB/s (12.5MB/s-18.8MB/s), io=60.7MiB (63.7MB), run=1002-1006msec 00:42:34.451 00:42:34.451 Disk stats (read/write): 00:42:34.451 nvme0n1: ios=3633/4067, merge=0/0, ticks=23223/27374, in_queue=50597, util=97.39% 00:42:34.451 nvme0n2: ios=3231/3584, merge=0/0, ticks=15558/17785, in_queue=33343, util=99.59% 00:42:34.451 nvme0n3: ios=2583/2774, merge=0/0, ticks=18292/32830, in_queue=51122, util=98.02% 00:42:34.451 nvme0n4: ios=2542/2560, merge=0/0, ticks=36639/65506, in_queue=102145, util=98.32% 00:42:34.451 17:58:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@55 -- # sync 00:42:34.451 17:58:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@59 -- # fio_pid=460350 00:42:34.451 17:58:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t read -r 10 00:42:34.451 17:58:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@61 -- # sleep 3 00:42:34.451 [global] 00:42:34.451 thread=1 00:42:34.451 invalidate=1 00:42:34.451 rw=read 00:42:34.451 time_based=1 00:42:34.451 runtime=10 00:42:34.451 ioengine=libaio 00:42:34.451 direct=1 00:42:34.451 bs=4096 00:42:34.451 iodepth=1 00:42:34.451 norandommap=1 00:42:34.451 numjobs=1 00:42:34.451 00:42:34.451 [job0] 00:42:34.451 filename=/dev/nvme0n1 00:42:34.451 [job1] 00:42:34.451 filename=/dev/nvme0n2 00:42:34.451 [job2] 00:42:34.451 filename=/dev/nvme0n3 00:42:34.451 [job3] 00:42:34.451 filename=/dev/nvme0n4 00:42:34.451 Could not set queue depth (nvme0n1) 00:42:34.451 Could not set queue depth (nvme0n2) 00:42:34.451 Could not set queue depth (nvme0n3) 00:42:34.451 Could not set queue depth (nvme0n4) 00:42:34.451 job0: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:42:34.451 job1: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:42:34.451 job2: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:42:34.451 job3: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:42:34.451 fio-3.35 00:42:34.451 Starting 4 threads 00:42:37.734 17:58:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_delete concat0 00:42:37.734 17:58:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_delete raid0 00:42:37.734 fio: io_u error on file /dev/nvme0n4: Operation not supported: read offset=37761024, buflen=4096 00:42:37.734 fio: pid=460442, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:42:37.992 17:58:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:42:37.992 17:58:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc0 00:42:37.992 fio: io_u error on file /dev/nvme0n3: Operation not supported: read offset=2560000, buflen=4096 00:42:37.992 fio: pid=460441, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:42:38.250 fio: io_u error on file /dev/nvme0n1: Operation not supported: read offset=49688576, buflen=4096 00:42:38.250 fio: pid=460438, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:42:38.250 17:58:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:42:38.250 17:58:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc1 00:42:38.509 17:58:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:42:38.509 17:58:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc2 00:42:38.509 fio: io_u error on file /dev/nvme0n2: Operation not supported: read offset=30404608, buflen=4096 00:42:38.509 fio: pid=460439, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:42:38.509 00:42:38.509 job0: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=460438: Fri Dec 6 17:58:20 2024 00:42:38.509 read: IOPS=3491, BW=13.6MiB/s (14.3MB/s)(47.4MiB/3475msec) 00:42:38.509 slat (usec): min=3, max=16444, avg=14.01, stdev=252.10 00:42:38.509 clat (usec): min=196, max=41090, avg=268.38, stdev=643.47 00:42:38.509 lat (usec): min=203, max=41101, avg=282.39, stdev=691.70 00:42:38.509 clat percentiles (usec): 00:42:38.509 | 1.00th=[ 208], 5.00th=[ 215], 10.00th=[ 219], 20.00th=[ 227], 00:42:38.509 | 30.00th=[ 233], 40.00th=[ 239], 50.00th=[ 243], 60.00th=[ 249], 00:42:38.509 | 70.00th=[ 258], 80.00th=[ 273], 90.00th=[ 314], 95.00th=[ 367], 00:42:38.509 | 99.00th=[ 494], 99.50th=[ 519], 99.90th=[ 553], 99.95th=[ 832], 00:42:38.509 | 99.99th=[41157] 00:42:38.509 bw ( KiB/s): min=10456, max=15816, per=44.83%, avg=13905.33, stdev=1907.08, samples=6 00:42:38.509 iops : min= 2614, max= 3954, avg=3476.33, stdev=476.77, samples=6 00:42:38.509 lat (usec) : 250=62.45%, 500=36.66%, 750=0.82%, 1000=0.01% 00:42:38.509 lat (msec) : 2=0.01%, 4=0.01%, 10=0.01%, 50=0.02% 00:42:38.509 cpu : usr=1.32%, sys=3.89%, ctx=12136, majf=0, minf=1 00:42:38.509 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:42:38.509 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:38.509 complete : 0=0.1%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:38.509 issued rwts: total=12132,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:42:38.509 latency : target=0, window=0, percentile=100.00%, depth=1 00:42:38.509 job1: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=460439: Fri Dec 6 17:58:20 2024 00:42:38.509 read: IOPS=1958, BW=7832KiB/s (8020kB/s)(29.0MiB/3791msec) 00:42:38.509 slat (usec): min=3, max=9868, avg=12.47, stdev=115.13 00:42:38.509 clat (usec): min=184, max=41245, avg=492.48, stdev=2911.11 00:42:38.509 lat (usec): min=189, max=50952, avg=504.96, stdev=2933.85 00:42:38.509 clat percentiles (usec): 00:42:38.509 | 1.00th=[ 206], 5.00th=[ 212], 10.00th=[ 219], 20.00th=[ 229], 00:42:38.509 | 30.00th=[ 237], 40.00th=[ 245], 50.00th=[ 260], 60.00th=[ 273], 00:42:38.509 | 70.00th=[ 297], 80.00th=[ 347], 90.00th=[ 383], 95.00th=[ 412], 00:42:38.509 | 99.00th=[ 494], 99.50th=[40633], 99.90th=[41157], 99.95th=[41157], 00:42:38.509 | 99.99th=[41157] 00:42:38.509 bw ( KiB/s): min= 93, max=11304, per=27.32%, avg=8475.00, stdev=3907.71, samples=7 00:42:38.509 iops : min= 23, max= 2826, avg=2118.71, stdev=977.02, samples=7 00:42:38.509 lat (usec) : 250=44.25%, 500=54.85%, 750=0.32%, 1000=0.01% 00:42:38.509 lat (msec) : 2=0.03%, 20=0.01%, 50=0.51% 00:42:38.509 cpu : usr=0.74%, sys=2.66%, ctx=7428, majf=0, minf=2 00:42:38.509 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:42:38.509 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:38.509 complete : 0=0.1%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:38.509 issued rwts: total=7424,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:42:38.509 latency : target=0, window=0, percentile=100.00%, depth=1 00:42:38.509 job2: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=460441: Fri Dec 6 17:58:20 2024 00:42:38.509 read: IOPS=193, BW=774KiB/s (793kB/s)(2500KiB/3229msec) 00:42:38.509 slat (nsec): min=5257, max=64532, avg=20254.36, stdev=9184.19 00:42:38.509 clat (usec): min=226, max=41466, avg=5092.01, stdev=13030.79 00:42:38.509 lat (usec): min=232, max=41479, avg=5112.27, stdev=13030.13 00:42:38.509 clat percentiles (usec): 00:42:38.509 | 1.00th=[ 235], 5.00th=[ 245], 10.00th=[ 251], 20.00th=[ 269], 00:42:38.509 | 30.00th=[ 277], 40.00th=[ 306], 50.00th=[ 330], 60.00th=[ 445], 00:42:38.509 | 70.00th=[ 474], 80.00th=[ 502], 90.00th=[40633], 95.00th=[41157], 00:42:38.509 | 99.00th=[41157], 99.50th=[41157], 99.90th=[41681], 99.95th=[41681], 00:42:38.509 | 99.99th=[41681] 00:42:38.509 bw ( KiB/s): min= 96, max= 4328, per=2.66%, avg=824.00, stdev=1716.73, samples=6 00:42:38.509 iops : min= 24, max= 1082, avg=206.00, stdev=429.18, samples=6 00:42:38.509 lat (usec) : 250=9.58%, 500=69.65%, 750=8.95% 00:42:38.509 lat (msec) : 50=11.66% 00:42:38.509 cpu : usr=0.19%, sys=0.46%, ctx=626, majf=0, minf=2 00:42:38.509 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:42:38.509 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:38.509 complete : 0=0.2%, 4=99.8%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:38.509 issued rwts: total=626,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:42:38.509 latency : target=0, window=0, percentile=100.00%, depth=1 00:42:38.509 job3: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=460442: Fri Dec 6 17:58:20 2024 00:42:38.509 read: IOPS=3158, BW=12.3MiB/s (12.9MB/s)(36.0MiB/2919msec) 00:42:38.509 slat (nsec): min=4345, max=63314, avg=12282.12, stdev=7214.65 00:42:38.509 clat (usec): min=188, max=41238, avg=298.98, stdev=1138.34 00:42:38.509 lat (usec): min=193, max=41248, avg=311.26, stdev=1138.51 00:42:38.509 clat percentiles (usec): 00:42:38.509 | 1.00th=[ 212], 5.00th=[ 223], 10.00th=[ 233], 20.00th=[ 241], 00:42:38.509 | 30.00th=[ 247], 40.00th=[ 251], 50.00th=[ 258], 60.00th=[ 265], 00:42:38.509 | 70.00th=[ 273], 80.00th=[ 281], 90.00th=[ 326], 95.00th=[ 351], 00:42:38.509 | 99.00th=[ 383], 99.50th=[ 408], 99.90th=[ 494], 99.95th=[41157], 00:42:38.510 | 99.99th=[41157] 00:42:38.510 bw ( KiB/s): min=11224, max=14192, per=41.12%, avg=12755.20, stdev=1088.01, samples=5 00:42:38.510 iops : min= 2806, max= 3548, avg=3188.80, stdev=272.00, samples=5 00:42:38.510 lat (usec) : 250=38.20%, 500=61.69%, 750=0.01% 00:42:38.510 lat (msec) : 20=0.01%, 50=0.08% 00:42:38.510 cpu : usr=1.85%, sys=4.28%, ctx=9220, majf=0, minf=1 00:42:38.510 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:42:38.510 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:38.510 complete : 0=0.1%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:42:38.510 issued rwts: total=9220,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:42:38.510 latency : target=0, window=0, percentile=100.00%, depth=1 00:42:38.510 00:42:38.510 Run status group 0 (all jobs): 00:42:38.510 READ: bw=30.3MiB/s (31.8MB/s), 774KiB/s-13.6MiB/s (793kB/s-14.3MB/s), io=115MiB (120MB), run=2919-3791msec 00:42:38.510 00:42:38.510 Disk stats (read/write): 00:42:38.510 nvme0n1: ios=11684/0, merge=0/0, ticks=3066/0, in_queue=3066, util=94.39% 00:42:38.510 nvme0n2: ios=7418/0, merge=0/0, ticks=3414/0, in_queue=3414, util=96.27% 00:42:38.510 nvme0n3: ios=620/0, merge=0/0, ticks=3059/0, in_queue=3059, util=96.79% 00:42:38.510 nvme0n4: ios=9202/0, merge=0/0, ticks=2579/0, in_queue=2579, util=96.71% 00:42:38.769 17:58:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:42:38.769 17:58:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc3 00:42:39.027 17:58:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:42:39.027 17:58:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc4 00:42:39.285 17:58:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:42:39.285 17:58:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc5 00:42:39.543 17:58:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:42:39.543 17:58:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc6 00:42:39.801 17:58:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@69 -- # fio_status=0 00:42:39.801 17:58:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@70 -- # wait 460350 00:42:39.801 17:58:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@70 -- # fio_status=4 00:42:39.801 17:58:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@72 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:42:40.060 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:42:40.060 17:58:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@73 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:42:40.060 17:58:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1223 -- # local i=0 00:42:40.060 17:58:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:42:40.060 17:58:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:42:40.060 17:58:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:42:40.060 17:58:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:42:40.060 17:58:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1235 -- # return 0 00:42:40.060 17:58:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@75 -- # '[' 4 -eq 0 ']' 00:42:40.060 17:58:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@80 -- # echo 'nvmf hotplug test: fio failed as expected' 00:42:40.060 nvmf hotplug test: fio failed as expected 00:42:40.060 17:58:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:42:40.318 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@85 -- # rm -f ./local-job0-0-verify.state 00:42:40.318 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@86 -- # rm -f ./local-job1-1-verify.state 00:42:40.318 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@87 -- # rm -f ./local-job2-2-verify.state 00:42:40.318 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@89 -- # trap - SIGINT SIGTERM EXIT 00:42:40.318 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@91 -- # nvmftestfini 00:42:40.318 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@516 -- # nvmfcleanup 00:42:40.318 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@121 -- # sync 00:42:40.318 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:42:40.318 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@124 -- # set +e 00:42:40.318 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@125 -- # for i in {1..20} 00:42:40.318 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:42:40.318 rmmod nvme_tcp 00:42:40.318 rmmod nvme_fabrics 00:42:40.318 rmmod nvme_keyring 00:42:40.318 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:42:40.318 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@128 -- # set -e 00:42:40.318 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@129 -- # return 0 00:42:40.318 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@517 -- # '[' -n 458340 ']' 00:42:40.318 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@518 -- # killprocess 458340 00:42:40.318 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@954 -- # '[' -z 458340 ']' 00:42:40.318 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@958 -- # kill -0 458340 00:42:40.318 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@959 -- # uname 00:42:40.318 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:42:40.318 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 458340 00:42:40.318 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:42:40.318 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:42:40.318 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@972 -- # echo 'killing process with pid 458340' 00:42:40.318 killing process with pid 458340 00:42:40.318 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@973 -- # kill 458340 00:42:40.318 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@978 -- # wait 458340 00:42:40.577 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:42:40.577 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:42:40.577 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:42:40.577 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@297 -- # iptr 00:42:40.577 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@791 -- # iptables-save 00:42:40.577 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@791 -- # iptables-restore 00:42:40.577 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:42:40.577 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:42:40.577 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@302 -- # remove_spdk_ns 00:42:40.577 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:42:40.577 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:42:40.577 17:58:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:42:43.110 00:42:43.110 real 0m23.897s 00:42:43.110 user 1m7.161s 00:42:43.110 sys 0m10.244s 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1130 -- # xtrace_disable 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:42:43.110 ************************************ 00:42:43.110 END TEST nvmf_fio_target 00:42:43.110 ************************************ 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@35 -- # run_test nvmf_bdevio /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp --interrupt-mode 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1111 -- # xtrace_disable 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:42:43.110 ************************************ 00:42:43.110 START TEST nvmf_bdevio 00:42:43.110 ************************************ 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp --interrupt-mode 00:42:43.110 * Looking for test storage... 00:42:43.110 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1711 -- # lcov --version 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@333 -- # local ver1 ver1_l 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@334 -- # local ver2 ver2_l 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@336 -- # IFS=.-: 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@336 -- # read -ra ver1 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@337 -- # IFS=.-: 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@337 -- # read -ra ver2 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@338 -- # local 'op=<' 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@340 -- # ver1_l=2 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@341 -- # ver2_l=1 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@344 -- # case "$op" in 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@345 -- # : 1 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@364 -- # (( v = 0 )) 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@365 -- # decimal 1 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@353 -- # local d=1 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@355 -- # echo 1 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@365 -- # ver1[v]=1 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@366 -- # decimal 2 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@353 -- # local d=2 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@355 -- # echo 2 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@366 -- # ver2[v]=2 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@368 -- # return 0 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:42:43.110 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:42:43.110 --rc genhtml_branch_coverage=1 00:42:43.110 --rc genhtml_function_coverage=1 00:42:43.110 --rc genhtml_legend=1 00:42:43.110 --rc geninfo_all_blocks=1 00:42:43.110 --rc geninfo_unexecuted_blocks=1 00:42:43.110 00:42:43.110 ' 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:42:43.110 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:42:43.110 --rc genhtml_branch_coverage=1 00:42:43.110 --rc genhtml_function_coverage=1 00:42:43.110 --rc genhtml_legend=1 00:42:43.110 --rc geninfo_all_blocks=1 00:42:43.110 --rc geninfo_unexecuted_blocks=1 00:42:43.110 00:42:43.110 ' 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:42:43.110 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:42:43.110 --rc genhtml_branch_coverage=1 00:42:43.110 --rc genhtml_function_coverage=1 00:42:43.110 --rc genhtml_legend=1 00:42:43.110 --rc geninfo_all_blocks=1 00:42:43.110 --rc geninfo_unexecuted_blocks=1 00:42:43.110 00:42:43.110 ' 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:42:43.110 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:42:43.110 --rc genhtml_branch_coverage=1 00:42:43.110 --rc genhtml_function_coverage=1 00:42:43.110 --rc genhtml_legend=1 00:42:43.110 --rc geninfo_all_blocks=1 00:42:43.110 --rc geninfo_unexecuted_blocks=1 00:42:43.110 00:42:43.110 ' 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@7 -- # uname -s 00:42:43.110 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@15 -- # shopt -s extglob 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- paths/export.sh@5 -- # export PATH 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@51 -- # : 0 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@55 -- # have_pci_nics=0 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@11 -- # MALLOC_BDEV_SIZE=64 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@14 -- # nvmftestinit 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@476 -- # prepare_net_devs 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@438 -- # local -g is_hw=no 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@440 -- # remove_spdk_ns 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@309 -- # xtrace_disable 00:42:43.111 17:58:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:42:45.095 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:42:45.095 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@315 -- # pci_devs=() 00:42:45.095 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@315 -- # local -a pci_devs 00:42:45.095 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@316 -- # pci_net_devs=() 00:42:45.095 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:42:45.095 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@317 -- # pci_drivers=() 00:42:45.095 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@317 -- # local -A pci_drivers 00:42:45.095 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@319 -- # net_devs=() 00:42:45.095 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@319 -- # local -ga net_devs 00:42:45.095 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@320 -- # e810=() 00:42:45.095 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@320 -- # local -ga e810 00:42:45.095 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@321 -- # x722=() 00:42:45.095 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@321 -- # local -ga x722 00:42:45.095 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@322 -- # mlx=() 00:42:45.095 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@322 -- # local -ga mlx 00:42:45.095 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:42:45.095 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:42:45.095 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:42:45.095 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:42:45.095 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:42:45.095 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:42:45.095 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:42:45.095 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:42:45.095 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:42:45.095 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:42:45.095 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:42:45.095 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:42:45.096 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:42:45.096 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@418 -- # [[ up == up ]] 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:42:45.096 Found net devices under 0000:0a:00.0: cvl_0_0 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@418 -- # [[ up == up ]] 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:42:45.096 Found net devices under 0000:0a:00.1: cvl_0_1 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@442 -- # is_hw=yes 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:42:45.096 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:42:45.096 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.184 ms 00:42:45.096 00:42:45.096 --- 10.0.0.2 ping statistics --- 00:42:45.096 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:42:45.096 rtt min/avg/max/mdev = 0.184/0.184/0.184/0.000 ms 00:42:45.096 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:42:45.096 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:42:45.096 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.122 ms 00:42:45.096 00:42:45.096 --- 10.0.0.1 ping statistics --- 00:42:45.097 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:42:45.097 rtt min/avg/max/mdev = 0.122/0.122/0.122/0.000 ms 00:42:45.097 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:42:45.097 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@450 -- # return 0 00:42:45.097 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:42:45.097 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:42:45.097 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:42:45.097 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:42:45.097 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:42:45.097 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:42:45.097 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:42:45.097 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@16 -- # nvmfappstart -m 0x78 00:42:45.097 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:42:45.097 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@726 -- # xtrace_disable 00:42:45.097 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:42:45.097 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@509 -- # nvmfpid=463081 00:42:45.097 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x78 00:42:45.097 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@510 -- # waitforlisten 463081 00:42:45.097 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@835 -- # '[' -z 463081 ']' 00:42:45.097 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:42:45.097 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@840 -- # local max_retries=100 00:42:45.097 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:42:45.097 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:42:45.097 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@844 -- # xtrace_disable 00:42:45.097 17:58:26 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:42:45.097 [2024-12-06 17:58:26.911882] thread.c:3005:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:42:45.097 [2024-12-06 17:58:26.912921] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:42:45.097 [2024-12-06 17:58:26.912974] [ DPDK EAL parameters: nvmf -c 0x78 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:42:45.355 [2024-12-06 17:58:26.984766] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:42:45.355 [2024-12-06 17:58:27.031722] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:42:45.355 [2024-12-06 17:58:27.031795] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:42:45.355 [2024-12-06 17:58:27.031825] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:42:45.356 [2024-12-06 17:58:27.031836] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:42:45.356 [2024-12-06 17:58:27.031846] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:42:45.356 [2024-12-06 17:58:27.033431] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:42:45.356 [2024-12-06 17:58:27.033492] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 5 00:42:45.356 [2024-12-06 17:58:27.033560] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 6 00:42:45.356 [2024-12-06 17:58:27.033563] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:42:45.356 [2024-12-06 17:58:27.117937] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:42:45.356 [2024-12-06 17:58:27.118164] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:42:45.356 [2024-12-06 17:58:27.118543] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:42:45.356 [2024-12-06 17:58:27.119071] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:42:45.356 [2024-12-06 17:58:27.119283] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_003) to intr mode from intr mode. 00:42:45.356 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:42:45.356 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@868 -- # return 0 00:42:45.356 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:42:45.356 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@732 -- # xtrace_disable 00:42:45.356 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:42:45.356 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:42:45.356 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:42:45.356 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@563 -- # xtrace_disable 00:42:45.356 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:42:45.356 [2024-12-06 17:58:27.178236] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:42:45.615 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:42:45.615 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:42:45.615 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@563 -- # xtrace_disable 00:42:45.615 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:42:45.615 Malloc0 00:42:45.615 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:42:45.615 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@20 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:42:45.615 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@563 -- # xtrace_disable 00:42:45.615 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:42:45.615 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:42:45.615 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@21 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:42:45.615 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@563 -- # xtrace_disable 00:42:45.615 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:42:45.615 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:42:45.615 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@22 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:42:45.615 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@563 -- # xtrace_disable 00:42:45.615 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:42:45.615 [2024-12-06 17:58:27.254425] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:42:45.615 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:42:45.615 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/bdev/bdevio/bdevio --json /dev/fd/62 00:42:45.615 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@24 -- # gen_nvmf_target_json 00:42:45.615 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@560 -- # config=() 00:42:45.615 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@560 -- # local subsystem config 00:42:45.615 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:42:45.615 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:42:45.615 { 00:42:45.615 "params": { 00:42:45.615 "name": "Nvme$subsystem", 00:42:45.615 "trtype": "$TEST_TRANSPORT", 00:42:45.615 "traddr": "$NVMF_FIRST_TARGET_IP", 00:42:45.615 "adrfam": "ipv4", 00:42:45.615 "trsvcid": "$NVMF_PORT", 00:42:45.615 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:42:45.615 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:42:45.615 "hdgst": ${hdgst:-false}, 00:42:45.615 "ddgst": ${ddgst:-false} 00:42:45.615 }, 00:42:45.615 "method": "bdev_nvme_attach_controller" 00:42:45.615 } 00:42:45.615 EOF 00:42:45.615 )") 00:42:45.615 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@582 -- # cat 00:42:45.615 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@584 -- # jq . 00:42:45.615 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@585 -- # IFS=, 00:42:45.615 17:58:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:42:45.615 "params": { 00:42:45.615 "name": "Nvme1", 00:42:45.615 "trtype": "tcp", 00:42:45.615 "traddr": "10.0.0.2", 00:42:45.615 "adrfam": "ipv4", 00:42:45.615 "trsvcid": "4420", 00:42:45.615 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:42:45.615 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:42:45.615 "hdgst": false, 00:42:45.615 "ddgst": false 00:42:45.615 }, 00:42:45.615 "method": "bdev_nvme_attach_controller" 00:42:45.615 }' 00:42:45.615 [2024-12-06 17:58:27.301790] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:42:45.615 [2024-12-06 17:58:27.301869] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid463199 ] 00:42:45.615 [2024-12-06 17:58:27.373348] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:42:45.615 [2024-12-06 17:58:27.425528] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:42:45.615 [2024-12-06 17:58:27.425582] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:42:45.615 [2024-12-06 17:58:27.425585] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:42:45.874 I/O targets: 00:42:45.874 Nvme1n1: 131072 blocks of 512 bytes (64 MiB) 00:42:45.874 00:42:45.874 00:42:45.874 CUnit - A unit testing framework for C - Version 2.1-3 00:42:45.874 http://cunit.sourceforge.net/ 00:42:45.874 00:42:45.874 00:42:45.874 Suite: bdevio tests on: Nvme1n1 00:42:45.874 Test: blockdev write read block ...passed 00:42:45.874 Test: blockdev write zeroes read block ...passed 00:42:45.874 Test: blockdev write zeroes read no split ...passed 00:42:46.132 Test: blockdev write zeroes read split ...passed 00:42:46.133 Test: blockdev write zeroes read split partial ...passed 00:42:46.133 Test: blockdev reset ...[2024-12-06 17:58:27.795578] nvme_ctrlr.c:1728:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 1] resetting controller 00:42:46.133 [2024-12-06 17:58:27.795690] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1aec700 (9): Bad file descriptor 00:42:46.133 [2024-12-06 17:58:27.799903] bdev_nvme.c:2286:bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller successful. 00:42:46.133 passed 00:42:46.133 Test: blockdev write read 8 blocks ...passed 00:42:46.133 Test: blockdev write read size > 128k ...passed 00:42:46.133 Test: blockdev write read invalid size ...passed 00:42:46.133 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:42:46.133 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:42:46.133 Test: blockdev write read max offset ...passed 00:42:46.133 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:42:46.391 Test: blockdev writev readv 8 blocks ...passed 00:42:46.391 Test: blockdev writev readv 30 x 1block ...passed 00:42:46.391 Test: blockdev writev readv block ...passed 00:42:46.391 Test: blockdev writev readv size > 128k ...passed 00:42:46.391 Test: blockdev writev readv size > 128k in two iovs ...passed 00:42:46.391 Test: blockdev comparev and writev ...[2024-12-06 17:58:28.055489] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:42:46.391 [2024-12-06 17:58:28.055526] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:42:46.391 [2024-12-06 17:58:28.055550] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:42:46.391 [2024-12-06 17:58:28.055567] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:42:46.391 [2024-12-06 17:58:28.055967] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:42:46.391 [2024-12-06 17:58:28.055993] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:42:46.391 [2024-12-06 17:58:28.056015] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:42:46.391 [2024-12-06 17:58:28.056032] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:42:46.391 [2024-12-06 17:58:28.056421] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:42:46.391 [2024-12-06 17:58:28.056445] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:42:46.391 [2024-12-06 17:58:28.056466] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:42:46.391 [2024-12-06 17:58:28.056482] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:42:46.391 [2024-12-06 17:58:28.056867] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:42:46.391 [2024-12-06 17:58:28.056892] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:42:46.391 [2024-12-06 17:58:28.056913] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:42:46.391 [2024-12-06 17:58:28.056929] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:42:46.391 passed 00:42:46.391 Test: blockdev nvme passthru rw ...passed 00:42:46.391 Test: blockdev nvme passthru vendor specific ...[2024-12-06 17:58:28.140944] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:42:46.391 [2024-12-06 17:58:28.140973] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:42:46.391 [2024-12-06 17:58:28.141146] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:42:46.391 [2024-12-06 17:58:28.141171] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:42:46.391 [2024-12-06 17:58:28.141337] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:42:46.391 [2024-12-06 17:58:28.141361] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:42:46.391 [2024-12-06 17:58:28.141519] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:42:46.391 [2024-12-06 17:58:28.141543] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:42:46.391 passed 00:42:46.391 Test: blockdev nvme admin passthru ...passed 00:42:46.391 Test: blockdev copy ...passed 00:42:46.391 00:42:46.391 Run Summary: Type Total Ran Passed Failed Inactive 00:42:46.391 suites 1 1 n/a 0 0 00:42:46.391 tests 23 23 23 0 0 00:42:46.391 asserts 152 152 152 0 n/a 00:42:46.391 00:42:46.391 Elapsed time = 1.184 seconds 00:42:46.649 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@26 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:42:46.649 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@563 -- # xtrace_disable 00:42:46.649 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:42:46.649 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:42:46.649 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@28 -- # trap - SIGINT SIGTERM EXIT 00:42:46.649 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@30 -- # nvmftestfini 00:42:46.649 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@516 -- # nvmfcleanup 00:42:46.649 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@121 -- # sync 00:42:46.649 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:42:46.649 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@124 -- # set +e 00:42:46.649 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@125 -- # for i in {1..20} 00:42:46.649 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:42:46.649 rmmod nvme_tcp 00:42:46.649 rmmod nvme_fabrics 00:42:46.649 rmmod nvme_keyring 00:42:46.649 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:42:46.649 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@128 -- # set -e 00:42:46.649 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@129 -- # return 0 00:42:46.649 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@517 -- # '[' -n 463081 ']' 00:42:46.649 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@518 -- # killprocess 463081 00:42:46.649 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@954 -- # '[' -z 463081 ']' 00:42:46.649 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@958 -- # kill -0 463081 00:42:46.649 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@959 -- # uname 00:42:46.649 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:42:46.649 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 463081 00:42:46.649 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@960 -- # process_name=reactor_3 00:42:46.649 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@964 -- # '[' reactor_3 = sudo ']' 00:42:46.649 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@972 -- # echo 'killing process with pid 463081' 00:42:46.649 killing process with pid 463081 00:42:46.649 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@973 -- # kill 463081 00:42:46.649 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@978 -- # wait 463081 00:42:46.908 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:42:46.908 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:42:46.908 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:42:46.908 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@297 -- # iptr 00:42:46.908 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@791 -- # iptables-save 00:42:46.908 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@791 -- # iptables-restore 00:42:46.908 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:42:46.908 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:42:46.908 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@302 -- # remove_spdk_ns 00:42:46.908 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:42:46.908 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:42:46.908 17:58:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:42:49.446 17:58:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:42:49.446 00:42:49.446 real 0m6.287s 00:42:49.446 user 0m7.938s 00:42:49.446 sys 0m2.408s 00:42:49.446 17:58:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1130 -- # xtrace_disable 00:42:49.446 17:58:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:42:49.446 ************************************ 00:42:49.446 END TEST nvmf_bdevio 00:42:49.446 ************************************ 00:42:49.446 17:58:30 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:42:49.446 00:42:49.446 real 3m53.697s 00:42:49.446 user 8m49.910s 00:42:49.446 sys 1m24.289s 00:42:49.446 17:58:30 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1130 -- # xtrace_disable 00:42:49.446 17:58:30 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:42:49.446 ************************************ 00:42:49.446 END TEST nvmf_target_core_interrupt_mode 00:42:49.446 ************************************ 00:42:49.446 17:58:30 nvmf_tcp -- nvmf/nvmf.sh@21 -- # run_test nvmf_interrupt /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/interrupt.sh --transport=tcp --interrupt-mode 00:42:49.446 17:58:30 nvmf_tcp -- common/autotest_common.sh@1105 -- # '[' 4 -le 1 ']' 00:42:49.446 17:58:30 nvmf_tcp -- common/autotest_common.sh@1111 -- # xtrace_disable 00:42:49.446 17:58:30 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:42:49.446 ************************************ 00:42:49.446 START TEST nvmf_interrupt 00:42:49.446 ************************************ 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/interrupt.sh --transport=tcp --interrupt-mode 00:42:49.446 * Looking for test storage... 00:42:49.446 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1711 -- # lcov --version 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@333 -- # local ver1 ver1_l 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@334 -- # local ver2 ver2_l 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@336 -- # IFS=.-: 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@336 -- # read -ra ver1 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@337 -- # IFS=.-: 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@337 -- # read -ra ver2 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@338 -- # local 'op=<' 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@340 -- # ver1_l=2 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@341 -- # ver2_l=1 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@344 -- # case "$op" in 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@345 -- # : 1 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@364 -- # (( v = 0 )) 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@365 -- # decimal 1 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@353 -- # local d=1 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@355 -- # echo 1 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@365 -- # ver1[v]=1 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@366 -- # decimal 2 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@353 -- # local d=2 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@355 -- # echo 2 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@366 -- # ver2[v]=2 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@368 -- # return 0 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:42:49.446 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:42:49.446 --rc genhtml_branch_coverage=1 00:42:49.446 --rc genhtml_function_coverage=1 00:42:49.446 --rc genhtml_legend=1 00:42:49.446 --rc geninfo_all_blocks=1 00:42:49.446 --rc geninfo_unexecuted_blocks=1 00:42:49.446 00:42:49.446 ' 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:42:49.446 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:42:49.446 --rc genhtml_branch_coverage=1 00:42:49.446 --rc genhtml_function_coverage=1 00:42:49.446 --rc genhtml_legend=1 00:42:49.446 --rc geninfo_all_blocks=1 00:42:49.446 --rc geninfo_unexecuted_blocks=1 00:42:49.446 00:42:49.446 ' 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:42:49.446 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:42:49.446 --rc genhtml_branch_coverage=1 00:42:49.446 --rc genhtml_function_coverage=1 00:42:49.446 --rc genhtml_legend=1 00:42:49.446 --rc geninfo_all_blocks=1 00:42:49.446 --rc geninfo_unexecuted_blocks=1 00:42:49.446 00:42:49.446 ' 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:42:49.446 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:42:49.446 --rc genhtml_branch_coverage=1 00:42:49.446 --rc genhtml_function_coverage=1 00:42:49.446 --rc genhtml_legend=1 00:42:49.446 --rc geninfo_all_blocks=1 00:42:49.446 --rc geninfo_unexecuted_blocks=1 00:42:49.446 00:42:49.446 ' 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@7 -- # uname -s 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@15 -- # shopt -s extglob 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- paths/export.sh@5 -- # export PATH 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:42:49.446 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@51 -- # : 0 00:42:49.447 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:42:49.447 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:42:49.447 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:42:49.447 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:42:49.447 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:42:49.447 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:42:49.447 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:42:49.447 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:42:49.447 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:42:49.447 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@55 -- # have_pci_nics=0 00:42:49.447 17:58:30 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/interrupt/common.sh 00:42:49.447 17:58:30 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@12 -- # NQN=nqn.2016-06.io.spdk:cnode1 00:42:49.447 17:58:30 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@14 -- # nvmftestinit 00:42:49.447 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:42:49.447 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:42:49.447 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@476 -- # prepare_net_devs 00:42:49.447 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@438 -- # local -g is_hw=no 00:42:49.447 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@440 -- # remove_spdk_ns 00:42:49.447 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:42:49.447 17:58:30 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:42:49.447 17:58:30 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:42:49.447 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:42:49.447 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:42:49.447 17:58:30 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@309 -- # xtrace_disable 00:42:49.447 17:58:30 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@315 -- # pci_devs=() 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@315 -- # local -a pci_devs 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@316 -- # pci_net_devs=() 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@317 -- # pci_drivers=() 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@317 -- # local -A pci_drivers 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@319 -- # net_devs=() 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@319 -- # local -ga net_devs 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@320 -- # e810=() 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@320 -- # local -ga e810 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@321 -- # x722=() 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@321 -- # local -ga x722 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@322 -- # mlx=() 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@322 -- # local -ga mlx 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:42:51.348 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:42:51.348 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@418 -- # [[ up == up ]] 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:42:51.348 Found net devices under 0000:0a:00.0: cvl_0_0 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:42:51.348 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:42:51.349 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@418 -- # [[ up == up ]] 00:42:51.349 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:42:51.349 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:42:51.349 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:42:51.349 Found net devices under 0000:0a:00.1: cvl_0_1 00:42:51.349 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:42:51.349 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:42:51.349 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@442 -- # is_hw=yes 00:42:51.349 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:42:51.349 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:42:51.349 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:42:51.349 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:42:51.349 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:42:51.349 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:42:51.349 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:42:51.349 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:42:51.349 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:42:51.349 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:42:51.349 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:42:51.349 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:42:51.349 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:42:51.349 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:42:51.349 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:42:51.349 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:42:51.349 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:42:51.349 17:58:32 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:42:51.349 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:42:51.349 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.218 ms 00:42:51.349 00:42:51.349 --- 10.0.0.2 ping statistics --- 00:42:51.349 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:42:51.349 rtt min/avg/max/mdev = 0.218/0.218/0.218/0.000 ms 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:42:51.349 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:42:51.349 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.141 ms 00:42:51.349 00:42:51.349 --- 10.0.0.1 ping statistics --- 00:42:51.349 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:42:51.349 rtt min/avg/max/mdev = 0.141/0.141/0.141/0.000 ms 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@450 -- # return 0 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@15 -- # nvmfappstart -m 0x3 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@726 -- # xtrace_disable 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@509 -- # nvmfpid=465308 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x3 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@510 -- # waitforlisten 465308 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@835 -- # '[' -z 465308 ']' 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@840 -- # local max_retries=100 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:42:51.349 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@844 -- # xtrace_disable 00:42:51.349 17:58:33 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:42:51.607 [2024-12-06 17:58:33.192909] thread.c:3005:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:42:51.607 [2024-12-06 17:58:33.194029] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:42:51.607 [2024-12-06 17:58:33.194093] [ DPDK EAL parameters: nvmf -c 0x3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:42:51.607 [2024-12-06 17:58:33.264661] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:42:51.607 [2024-12-06 17:58:33.309482] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:42:51.607 [2024-12-06 17:58:33.309532] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:42:51.607 [2024-12-06 17:58:33.309561] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:42:51.607 [2024-12-06 17:58:33.309572] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:42:51.607 [2024-12-06 17:58:33.309582] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:42:51.607 [2024-12-06 17:58:33.310855] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:42:51.607 [2024-12-06 17:58:33.310860] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:42:51.607 [2024-12-06 17:58:33.394394] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:42:51.607 [2024-12-06 17:58:33.394404] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:42:51.607 [2024-12-06 17:58:33.394654] thread.c:2143:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:42:51.607 17:58:33 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:42:51.607 17:58:33 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@868 -- # return 0 00:42:51.607 17:58:33 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:42:51.607 17:58:33 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@732 -- # xtrace_disable 00:42:51.607 17:58:33 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@16 -- # setup_bdev_aio 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@77 -- # uname -s 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@77 -- # [[ Linux != \F\r\e\e\B\S\D ]] 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@78 -- # dd if=/dev/zero of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aiofile bs=2048 count=5000 00:42:51.866 5000+0 records in 00:42:51.866 5000+0 records out 00:42:51.866 10240000 bytes (10 MB, 9.8 MiB) copied, 0.013749 s, 745 MB/s 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@79 -- # rpc_cmd bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aiofile AIO0 2048 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@563 -- # xtrace_disable 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:42:51.866 AIO0 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -q 256 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@563 -- # xtrace_disable 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:42:51.866 [2024-12-06 17:58:33.527509] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@19 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@563 -- # xtrace_disable 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@20 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 AIO0 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@563 -- # xtrace_disable 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@21 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@563 -- # xtrace_disable 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:42:51.866 [2024-12-06 17:58:33.555773] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@24 -- # for i in {0..1} 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@25 -- # reactor_is_idle 465308 0 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@53 -- # reactor_is_busy_or_idle 465308 0 idle 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=465308 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=0 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=65 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \b\u\s\y ]] 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \i\d\l\e ]] 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 465308 -w 256 00:42:51.866 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_0 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor=' 465308 root 20 0 128.2g 47232 34176 S 0.0 0.1 0:00.25 reactor_0' 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 465308 root 20 0 128.2g 47232 34176 S 0.0 0.1 0:00.25 reactor_0 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=0.0 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=0 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ idle = \b\u\s\y ]] 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ idle = \i\d\l\e ]] 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # (( cpu_rate > idle_threshold )) 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@24 -- # for i in {0..1} 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@25 -- # reactor_is_idle 465308 1 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@53 -- # reactor_is_busy_or_idle 465308 1 idle 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=465308 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=1 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=65 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \b\u\s\y ]] 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \i\d\l\e ]] 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 465308 -w 256 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_1 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor=' 465312 root 20 0 128.2g 47232 34176 S 0.0 0.1 0:00.00 reactor_1' 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 465312 root 20 0 128.2g 47232 34176 S 0.0 0.1 0:00.00 reactor_1 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=0.0 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=0 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ idle = \b\u\s\y ]] 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ idle = \i\d\l\e ]] 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # (( cpu_rate > idle_threshold )) 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@28 -- # perf=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@35 -- # perf_pid=465356 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 256 -o 4096 -w randrw -M 30 -t 10 -c 0xC -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@38 -- # for i in {0..1} 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@39 -- # BUSY_THRESHOLD=30 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@39 -- # reactor_is_busy 465308 0 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@49 -- # reactor_is_busy_or_idle 465308 0 busy 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=465308 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=0 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=busy 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=30 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ busy != \b\u\s\y ]] 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 465308 -w 256 00:42:52.124 17:58:33 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_0 00:42:52.381 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor=' 465308 root 20 0 128.2g 48384 34560 R 86.7 0.1 0:00.38 reactor_0' 00:42:52.381 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 465308 root 20 0 128.2g 48384 34560 R 86.7 0.1 0:00.38 reactor_0 00:42:52.381 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:42:52.381 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:42:52.381 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=86.7 00:42:52.381 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=86 00:42:52.381 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ busy = \b\u\s\y ]] 00:42:52.381 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # (( cpu_rate < busy_threshold )) 00:42:52.381 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ busy = \i\d\l\e ]] 00:42:52.381 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:42:52.381 17:58:34 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@38 -- # for i in {0..1} 00:42:52.381 17:58:34 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@39 -- # BUSY_THRESHOLD=30 00:42:52.381 17:58:34 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@39 -- # reactor_is_busy 465308 1 00:42:52.381 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@49 -- # reactor_is_busy_or_idle 465308 1 busy 00:42:52.381 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=465308 00:42:52.381 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=1 00:42:52.381 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=busy 00:42:52.381 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=30 00:42:52.381 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:42:52.381 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ busy != \b\u\s\y ]] 00:42:52.381 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:42:52.381 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:42:52.381 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:42:52.381 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 465308 -w 256 00:42:52.381 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_1 00:42:52.639 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor=' 465312 root 20 0 128.2g 48384 34560 R 99.9 0.1 0:00.23 reactor_1' 00:42:52.639 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 465312 root 20 0 128.2g 48384 34560 R 99.9 0.1 0:00.23 reactor_1 00:42:52.639 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:42:52.639 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:42:52.639 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=99.9 00:42:52.639 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=99 00:42:52.639 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ busy = \b\u\s\y ]] 00:42:52.639 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # (( cpu_rate < busy_threshold )) 00:42:52.639 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ busy = \i\d\l\e ]] 00:42:52.639 17:58:34 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:42:52.639 17:58:34 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@42 -- # wait 465356 00:43:02.610 Initializing NVMe Controllers 00:43:02.610 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:43:02.610 Controller IO queue size 256, less than required. 00:43:02.610 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:43:02.610 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 2 00:43:02.610 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 3 00:43:02.610 Initialization complete. Launching workers. 00:43:02.610 ======================================================== 00:43:02.610 Latency(us) 00:43:02.610 Device Information : IOPS MiB/s Average min max 00:43:02.610 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 2: 13877.69 54.21 18458.66 3946.99 55981.43 00:43:02.610 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 3: 13804.59 53.92 18556.49 4136.26 58671.99 00:43:02.610 ======================================================== 00:43:02.610 Total : 27682.29 108.13 18507.45 3946.99 58671.99 00:43:02.610 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@45 -- # for i in {0..1} 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@46 -- # reactor_is_idle 465308 0 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@53 -- # reactor_is_busy_or_idle 465308 0 idle 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=465308 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=0 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=65 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \b\u\s\y ]] 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \i\d\l\e ]] 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 465308 -w 256 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_0 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor=' 465308 root 20 0 128.2g 48384 34560 S 0.0 0.1 0:20.18 reactor_0' 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 465308 root 20 0 128.2g 48384 34560 S 0.0 0.1 0:20.18 reactor_0 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=0.0 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=0 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ idle = \b\u\s\y ]] 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ idle = \i\d\l\e ]] 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # (( cpu_rate > idle_threshold )) 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@45 -- # for i in {0..1} 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@46 -- # reactor_is_idle 465308 1 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@53 -- # reactor_is_busy_or_idle 465308 1 idle 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=465308 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=1 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=65 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \b\u\s\y ]] 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \i\d\l\e ]] 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:43:02.610 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:43:02.611 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:43:02.611 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 465308 -w 256 00:43:02.611 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_1 00:43:02.611 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor=' 465312 root 20 0 128.2g 48384 34560 S 0.0 0.1 0:09.96 reactor_1' 00:43:02.611 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 465312 root 20 0 128.2g 48384 34560 S 0.0 0.1 0:09.96 reactor_1 00:43:02.611 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:43:02.611 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:43:02.611 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=0.0 00:43:02.611 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=0 00:43:02.611 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ idle = \b\u\s\y ]] 00:43:02.611 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ idle = \i\d\l\e ]] 00:43:02.611 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # (( cpu_rate > idle_threshold )) 00:43:02.611 17:58:44 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:43:02.611 17:58:44 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@50 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:43:03.177 17:58:44 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@51 -- # waitforserial SPDKISFASTANDAWESOME 00:43:03.177 17:58:44 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1202 -- # local i=0 00:43:03.177 17:58:44 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1203 -- # local nvme_device_counter=1 nvme_devices=0 00:43:03.177 17:58:44 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1204 -- # [[ -n '' ]] 00:43:03.177 17:58:44 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1209 -- # sleep 2 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1210 -- # (( i++ <= 15 )) 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1211 -- # lsblk -l -o NAME,SERIAL 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1211 -- # grep -c SPDKISFASTANDAWESOME 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1211 -- # nvme_devices=1 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1212 -- # (( nvme_devices == nvme_device_counter )) 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1212 -- # return 0 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@52 -- # for i in {0..1} 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@53 -- # reactor_is_idle 465308 0 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@53 -- # reactor_is_busy_or_idle 465308 0 idle 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=465308 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=0 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=65 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \b\u\s\y ]] 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \i\d\l\e ]] 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 465308 -w 256 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_0 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor=' 465308 root 20 0 128.2g 60672 34560 S 0.0 0.1 0:20.27 reactor_0' 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 465308 root 20 0 128.2g 60672 34560 S 0.0 0.1 0:20.27 reactor_0 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=0.0 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=0 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ idle = \b\u\s\y ]] 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ idle = \i\d\l\e ]] 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # (( cpu_rate > idle_threshold )) 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@52 -- # for i in {0..1} 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@53 -- # reactor_is_idle 465308 1 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@53 -- # reactor_is_busy_or_idle 465308 1 idle 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=465308 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=1 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=65 00:43:05.079 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:43:05.337 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \b\u\s\y ]] 00:43:05.337 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \i\d\l\e ]] 00:43:05.337 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:43:05.337 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:43:05.337 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:43:05.337 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 465308 -w 256 00:43:05.337 17:58:46 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_1 00:43:05.337 17:58:47 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor=' 465312 root 20 0 128.2g 60672 34560 S 0.0 0.1 0:09.99 reactor_1' 00:43:05.337 17:58:47 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 465312 root 20 0 128.2g 60672 34560 S 0.0 0.1 0:09.99 reactor_1 00:43:05.337 17:58:47 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:43:05.337 17:58:47 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:43:05.337 17:58:47 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=0.0 00:43:05.337 17:58:47 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=0 00:43:05.337 17:58:47 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ idle = \b\u\s\y ]] 00:43:05.337 17:58:47 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ idle = \i\d\l\e ]] 00:43:05.337 17:58:47 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # (( cpu_rate > idle_threshold )) 00:43:05.337 17:58:47 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:43:05.337 17:58:47 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@55 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:43:05.594 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@56 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1223 -- # local i=0 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1224 -- # lsblk -o NAME,SERIAL 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1224 -- # grep -q -w SPDKISFASTANDAWESOME 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1231 -- # lsblk -l -o NAME,SERIAL 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1231 -- # grep -q -w SPDKISFASTANDAWESOME 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1235 -- # return 0 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@58 -- # trap - SIGINT SIGTERM EXIT 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@59 -- # nvmftestfini 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@516 -- # nvmfcleanup 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@121 -- # sync 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@124 -- # set +e 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@125 -- # for i in {1..20} 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:43:05.594 rmmod nvme_tcp 00:43:05.594 rmmod nvme_fabrics 00:43:05.594 rmmod nvme_keyring 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@128 -- # set -e 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@129 -- # return 0 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@517 -- # '[' -n 465308 ']' 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@518 -- # killprocess 465308 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@954 -- # '[' -z 465308 ']' 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@958 -- # kill -0 465308 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@959 -- # uname 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 465308 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@972 -- # echo 'killing process with pid 465308' 00:43:05.594 killing process with pid 465308 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@973 -- # kill 465308 00:43:05.594 17:58:47 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@978 -- # wait 465308 00:43:05.852 17:58:47 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:43:05.852 17:58:47 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:43:05.852 17:58:47 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:43:05.852 17:58:47 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@297 -- # iptr 00:43:05.852 17:58:47 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@791 -- # iptables-save 00:43:05.852 17:58:47 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:43:05.852 17:58:47 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@791 -- # iptables-restore 00:43:05.852 17:58:47 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:43:05.852 17:58:47 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@302 -- # remove_spdk_ns 00:43:05.852 17:58:47 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:43:05.852 17:58:47 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:43:05.852 17:58:47 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:43:08.390 17:58:49 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:43:08.390 00:43:08.390 real 0m18.873s 00:43:08.390 user 0m37.509s 00:43:08.390 sys 0m6.311s 00:43:08.390 17:58:49 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1130 -- # xtrace_disable 00:43:08.390 17:58:49 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:43:08.390 ************************************ 00:43:08.390 END TEST nvmf_interrupt 00:43:08.390 ************************************ 00:43:08.390 00:43:08.390 real 32m57.914s 00:43:08.390 user 87m12.967s 00:43:08.390 sys 7m58.940s 00:43:08.390 17:58:49 nvmf_tcp -- common/autotest_common.sh@1130 -- # xtrace_disable 00:43:08.390 17:58:49 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:43:08.390 ************************************ 00:43:08.390 END TEST nvmf_tcp 00:43:08.390 ************************************ 00:43:08.390 17:58:49 -- spdk/autotest.sh@285 -- # [[ 0 -eq 0 ]] 00:43:08.390 17:58:49 -- spdk/autotest.sh@286 -- # run_test spdkcli_nvmf_tcp /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/nvmf.sh --transport=tcp 00:43:08.390 17:58:49 -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:43:08.390 17:58:49 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:43:08.390 17:58:49 -- common/autotest_common.sh@10 -- # set +x 00:43:08.390 ************************************ 00:43:08.390 START TEST spdkcli_nvmf_tcp 00:43:08.390 ************************************ 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/nvmf.sh --transport=tcp 00:43:08.391 * Looking for test storage... 00:43:08.391 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- common/autotest_common.sh@1711 -- # lcov --version 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@333 -- # local ver1 ver1_l 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@334 -- # local ver2 ver2_l 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@336 -- # IFS=.-: 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@336 -- # read -ra ver1 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@337 -- # IFS=.-: 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@337 -- # read -ra ver2 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@338 -- # local 'op=<' 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@340 -- # ver1_l=2 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@341 -- # ver2_l=1 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@344 -- # case "$op" in 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@345 -- # : 1 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@364 -- # (( v = 0 )) 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@365 -- # decimal 1 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@353 -- # local d=1 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@355 -- # echo 1 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@365 -- # ver1[v]=1 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@366 -- # decimal 2 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@353 -- # local d=2 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@355 -- # echo 2 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@366 -- # ver2[v]=2 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@368 -- # return 0 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:43:08.391 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:43:08.391 --rc genhtml_branch_coverage=1 00:43:08.391 --rc genhtml_function_coverage=1 00:43:08.391 --rc genhtml_legend=1 00:43:08.391 --rc geninfo_all_blocks=1 00:43:08.391 --rc geninfo_unexecuted_blocks=1 00:43:08.391 00:43:08.391 ' 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:43:08.391 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:43:08.391 --rc genhtml_branch_coverage=1 00:43:08.391 --rc genhtml_function_coverage=1 00:43:08.391 --rc genhtml_legend=1 00:43:08.391 --rc geninfo_all_blocks=1 00:43:08.391 --rc geninfo_unexecuted_blocks=1 00:43:08.391 00:43:08.391 ' 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:43:08.391 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:43:08.391 --rc genhtml_branch_coverage=1 00:43:08.391 --rc genhtml_function_coverage=1 00:43:08.391 --rc genhtml_legend=1 00:43:08.391 --rc geninfo_all_blocks=1 00:43:08.391 --rc geninfo_unexecuted_blocks=1 00:43:08.391 00:43:08.391 ' 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:43:08.391 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:43:08.391 --rc genhtml_branch_coverage=1 00:43:08.391 --rc genhtml_function_coverage=1 00:43:08.391 --rc genhtml_legend=1 00:43:08.391 --rc geninfo_all_blocks=1 00:43:08.391 --rc geninfo_unexecuted_blocks=1 00:43:08.391 00:43:08.391 ' 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/common.sh 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- spdkcli/common.sh@6 -- # spdkcli_job=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_job.py 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/clear_config.py 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- nvmf/common.sh@7 -- # uname -s 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@15 -- # shopt -s extglob 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- paths/export.sh@5 -- # export PATH 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- nvmf/common.sh@51 -- # : 0 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:43:08.391 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- nvmf/common.sh@55 -- # have_pci_nics=0 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@12 -- # MATCH_FILE=spdkcli_nvmf.test 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@13 -- # SPDKCLI_BRANCH=/nvmf 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@15 -- # trap cleanup EXIT 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@17 -- # timing_enter run_nvmf_tgt 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- common/autotest_common.sh@726 -- # xtrace_disable 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@18 -- # run_nvmf_tgt 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- spdkcli/common.sh@33 -- # nvmf_tgt_pid=467350 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- spdkcli/common.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -m 0x3 -p 0 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- spdkcli/common.sh@34 -- # waitforlisten 467350 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- common/autotest_common.sh@835 -- # '[' -z 467350 ']' 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- common/autotest_common.sh@840 -- # local max_retries=100 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:43:08.391 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- common/autotest_common.sh@844 -- # xtrace_disable 00:43:08.391 17:58:49 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:43:08.392 [2024-12-06 17:58:49.892590] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:43:08.392 [2024-12-06 17:58:49.892705] [ DPDK EAL parameters: nvmf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid467350 ] 00:43:08.392 [2024-12-06 17:58:49.965583] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:43:08.392 [2024-12-06 17:58:50.019417] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:43:08.392 [2024-12-06 17:58:50.019430] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:43:08.392 17:58:50 spdkcli_nvmf_tcp -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:43:08.392 17:58:50 spdkcli_nvmf_tcp -- common/autotest_common.sh@868 -- # return 0 00:43:08.392 17:58:50 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@19 -- # timing_exit run_nvmf_tgt 00:43:08.392 17:58:50 spdkcli_nvmf_tcp -- common/autotest_common.sh@732 -- # xtrace_disable 00:43:08.392 17:58:50 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:43:08.392 17:58:50 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@21 -- # NVMF_TARGET_IP=127.0.0.1 00:43:08.392 17:58:50 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@22 -- # [[ tcp == \r\d\m\a ]] 00:43:08.392 17:58:50 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@27 -- # timing_enter spdkcli_create_nvmf_config 00:43:08.392 17:58:50 spdkcli_nvmf_tcp -- common/autotest_common.sh@726 -- # xtrace_disable 00:43:08.392 17:58:50 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:43:08.392 17:58:50 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@65 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc create 32 512 Malloc1'\'' '\''Malloc1'\'' True 00:43:08.392 '\''/bdevs/malloc create 32 512 Malloc2'\'' '\''Malloc2'\'' True 00:43:08.392 '\''/bdevs/malloc create 32 512 Malloc3'\'' '\''Malloc3'\'' True 00:43:08.392 '\''/bdevs/malloc create 32 512 Malloc4'\'' '\''Malloc4'\'' True 00:43:08.392 '\''/bdevs/malloc create 32 512 Malloc5'\'' '\''Malloc5'\'' True 00:43:08.392 '\''/bdevs/malloc create 32 512 Malloc6'\'' '\''Malloc6'\'' True 00:43:08.392 '\''nvmf/transport create tcp max_io_qpairs_per_ctrlr=4 io_unit_size=8192'\'' '\'''\'' True 00:43:08.392 '\''/nvmf/subsystem create nqn.2014-08.org.spdk:cnode1 N37SXV509SRW max_namespaces=4 allow_any_host=True'\'' '\''nqn.2014-08.org.spdk:cnode1'\'' True 00:43:08.392 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc3 1'\'' '\''Malloc3'\'' True 00:43:08.392 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc4 2'\'' '\''Malloc4'\'' True 00:43:08.392 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4260 IPv4'\'' '\''127.0.0.1:4260'\'' True 00:43:08.392 '\''/nvmf/subsystem create nqn.2014-08.org.spdk:cnode2 N37SXV509SRD max_namespaces=2 allow_any_host=True'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' True 00:43:08.392 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode2/namespaces create Malloc2'\'' '\''Malloc2'\'' True 00:43:08.392 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode2/listen_addresses create tcp 127.0.0.1 4260 IPv4'\'' '\''127.0.0.1:4260'\'' True 00:43:08.392 '\''/nvmf/subsystem create nqn.2014-08.org.spdk:cnode3 N37SXV509SRR max_namespaces=2 allow_any_host=True'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' True 00:43:08.392 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/namespaces create Malloc1'\'' '\''Malloc1'\'' True 00:43:08.392 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/listen_addresses create tcp 127.0.0.1 4260 IPv4'\'' '\''127.0.0.1:4260'\'' True 00:43:08.392 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/listen_addresses create tcp 127.0.0.1 4261 IPv4'\'' '\''127.0.0.1:4261'\'' True 00:43:08.392 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts create nqn.2014-08.org.spdk:cnode1'\'' '\''nqn.2014-08.org.spdk:cnode1'\'' True 00:43:08.392 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts create nqn.2014-08.org.spdk:cnode2'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' True 00:43:08.392 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1 allow_any_host True'\'' '\''Allow any host'\'' 00:43:08.392 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1 allow_any_host False'\'' '\''Allow any host'\'' True 00:43:08.392 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4261 IPv4'\'' '\''127.0.0.1:4261'\'' True 00:43:08.392 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4262 IPv4'\'' '\''127.0.0.1:4262'\'' True 00:43:08.392 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/hosts create nqn.2014-08.org.spdk:cnode2'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' True 00:43:08.392 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc5'\'' '\''Malloc5'\'' True 00:43:08.392 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc6'\'' '\''Malloc6'\'' True 00:43:08.392 '\''/nvmf/referral create tcp 127.0.0.2 4030 IPv4'\'' 00:43:08.392 ' 00:43:11.675 [2024-12-06 17:58:52.867367] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:43:12.609 [2024-12-06 17:58:54.135773] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4260 *** 00:43:15.136 [2024-12-06 17:58:56.498997] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4261 *** 00:43:17.036 [2024-12-06 17:58:58.509130] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4262 *** 00:43:18.409 Executing command: ['/bdevs/malloc create 32 512 Malloc1', 'Malloc1', True] 00:43:18.409 Executing command: ['/bdevs/malloc create 32 512 Malloc2', 'Malloc2', True] 00:43:18.409 Executing command: ['/bdevs/malloc create 32 512 Malloc3', 'Malloc3', True] 00:43:18.409 Executing command: ['/bdevs/malloc create 32 512 Malloc4', 'Malloc4', True] 00:43:18.410 Executing command: ['/bdevs/malloc create 32 512 Malloc5', 'Malloc5', True] 00:43:18.410 Executing command: ['/bdevs/malloc create 32 512 Malloc6', 'Malloc6', True] 00:43:18.410 Executing command: ['nvmf/transport create tcp max_io_qpairs_per_ctrlr=4 io_unit_size=8192', '', True] 00:43:18.410 Executing command: ['/nvmf/subsystem create nqn.2014-08.org.spdk:cnode1 N37SXV509SRW max_namespaces=4 allow_any_host=True', 'nqn.2014-08.org.spdk:cnode1', True] 00:43:18.410 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc3 1', 'Malloc3', True] 00:43:18.410 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc4 2', 'Malloc4', True] 00:43:18.410 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4260 IPv4', '127.0.0.1:4260', True] 00:43:18.410 Executing command: ['/nvmf/subsystem create nqn.2014-08.org.spdk:cnode2 N37SXV509SRD max_namespaces=2 allow_any_host=True', 'nqn.2014-08.org.spdk:cnode2', True] 00:43:18.410 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode2/namespaces create Malloc2', 'Malloc2', True] 00:43:18.410 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode2/listen_addresses create tcp 127.0.0.1 4260 IPv4', '127.0.0.1:4260', True] 00:43:18.410 Executing command: ['/nvmf/subsystem create nqn.2014-08.org.spdk:cnode3 N37SXV509SRR max_namespaces=2 allow_any_host=True', 'nqn.2014-08.org.spdk:cnode2', True] 00:43:18.410 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/namespaces create Malloc1', 'Malloc1', True] 00:43:18.410 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/listen_addresses create tcp 127.0.0.1 4260 IPv4', '127.0.0.1:4260', True] 00:43:18.410 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/listen_addresses create tcp 127.0.0.1 4261 IPv4', '127.0.0.1:4261', True] 00:43:18.410 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts create nqn.2014-08.org.spdk:cnode1', 'nqn.2014-08.org.spdk:cnode1', True] 00:43:18.410 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts create nqn.2014-08.org.spdk:cnode2', 'nqn.2014-08.org.spdk:cnode2', True] 00:43:18.410 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1 allow_any_host True', 'Allow any host', False] 00:43:18.410 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1 allow_any_host False', 'Allow any host', True] 00:43:18.410 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4261 IPv4', '127.0.0.1:4261', True] 00:43:18.410 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4262 IPv4', '127.0.0.1:4262', True] 00:43:18.410 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/hosts create nqn.2014-08.org.spdk:cnode2', 'nqn.2014-08.org.spdk:cnode2', True] 00:43:18.410 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc5', 'Malloc5', True] 00:43:18.410 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc6', 'Malloc6', True] 00:43:18.410 Executing command: ['/nvmf/referral create tcp 127.0.0.2 4030 IPv4', False] 00:43:18.410 17:59:00 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@66 -- # timing_exit spdkcli_create_nvmf_config 00:43:18.410 17:59:00 spdkcli_nvmf_tcp -- common/autotest_common.sh@732 -- # xtrace_disable 00:43:18.410 17:59:00 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:43:18.410 17:59:00 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@68 -- # timing_enter spdkcli_check_match 00:43:18.410 17:59:00 spdkcli_nvmf_tcp -- common/autotest_common.sh@726 -- # xtrace_disable 00:43:18.410 17:59:00 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:43:18.410 17:59:00 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@69 -- # check_match 00:43:18.410 17:59:00 spdkcli_nvmf_tcp -- spdkcli/common.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdkcli.py ll /nvmf 00:43:18.976 17:59:00 spdkcli_nvmf_tcp -- spdkcli/common.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/match/match /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/match_files/spdkcli_nvmf.test.match 00:43:18.976 17:59:00 spdkcli_nvmf_tcp -- spdkcli/common.sh@46 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/match_files/spdkcli_nvmf.test 00:43:18.976 17:59:00 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@70 -- # timing_exit spdkcli_check_match 00:43:18.976 17:59:00 spdkcli_nvmf_tcp -- common/autotest_common.sh@732 -- # xtrace_disable 00:43:18.976 17:59:00 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:43:18.976 17:59:00 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@72 -- # timing_enter spdkcli_clear_nvmf_config 00:43:18.976 17:59:00 spdkcli_nvmf_tcp -- common/autotest_common.sh@726 -- # xtrace_disable 00:43:18.976 17:59:00 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:43:18.976 17:59:00 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@87 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_job.py ''\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces delete nsid=1'\'' '\''Malloc3'\'' 00:43:18.976 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces delete_all'\'' '\''Malloc4'\'' 00:43:18.976 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/hosts delete nqn.2014-08.org.spdk:cnode2'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' 00:43:18.976 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts delete_all'\'' '\''nqn.2014-08.org.spdk:cnode1'\'' 00:43:18.976 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses delete tcp 127.0.0.1 4262'\'' '\''127.0.0.1:4262'\'' 00:43:18.976 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses delete_all'\'' '\''127.0.0.1:4261'\'' 00:43:18.976 '\''/nvmf/subsystem delete nqn.2014-08.org.spdk:cnode3'\'' '\''nqn.2014-08.org.spdk:cnode3'\'' 00:43:18.976 '\''/nvmf/subsystem delete_all'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' 00:43:18.976 '\''/bdevs/malloc delete Malloc6'\'' '\''Malloc6'\'' 00:43:18.976 '\''/bdevs/malloc delete Malloc5'\'' '\''Malloc5'\'' 00:43:18.976 '\''/bdevs/malloc delete Malloc4'\'' '\''Malloc4'\'' 00:43:18.976 '\''/bdevs/malloc delete Malloc3'\'' '\''Malloc3'\'' 00:43:18.976 '\''/bdevs/malloc delete Malloc2'\'' '\''Malloc2'\'' 00:43:18.976 '\''/bdevs/malloc delete Malloc1'\'' '\''Malloc1'\'' 00:43:18.976 ' 00:43:24.290 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces delete nsid=1', 'Malloc3', False] 00:43:24.290 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces delete_all', 'Malloc4', False] 00:43:24.290 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/hosts delete nqn.2014-08.org.spdk:cnode2', 'nqn.2014-08.org.spdk:cnode2', False] 00:43:24.290 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts delete_all', 'nqn.2014-08.org.spdk:cnode1', False] 00:43:24.290 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses delete tcp 127.0.0.1 4262', '127.0.0.1:4262', False] 00:43:24.290 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses delete_all', '127.0.0.1:4261', False] 00:43:24.290 Executing command: ['/nvmf/subsystem delete nqn.2014-08.org.spdk:cnode3', 'nqn.2014-08.org.spdk:cnode3', False] 00:43:24.290 Executing command: ['/nvmf/subsystem delete_all', 'nqn.2014-08.org.spdk:cnode2', False] 00:43:24.290 Executing command: ['/bdevs/malloc delete Malloc6', 'Malloc6', False] 00:43:24.290 Executing command: ['/bdevs/malloc delete Malloc5', 'Malloc5', False] 00:43:24.290 Executing command: ['/bdevs/malloc delete Malloc4', 'Malloc4', False] 00:43:24.290 Executing command: ['/bdevs/malloc delete Malloc3', 'Malloc3', False] 00:43:24.290 Executing command: ['/bdevs/malloc delete Malloc2', 'Malloc2', False] 00:43:24.290 Executing command: ['/bdevs/malloc delete Malloc1', 'Malloc1', False] 00:43:24.547 17:59:06 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@88 -- # timing_exit spdkcli_clear_nvmf_config 00:43:24.547 17:59:06 spdkcli_nvmf_tcp -- common/autotest_common.sh@732 -- # xtrace_disable 00:43:24.547 17:59:06 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:43:24.547 17:59:06 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@90 -- # killprocess 467350 00:43:24.547 17:59:06 spdkcli_nvmf_tcp -- common/autotest_common.sh@954 -- # '[' -z 467350 ']' 00:43:24.547 17:59:06 spdkcli_nvmf_tcp -- common/autotest_common.sh@958 -- # kill -0 467350 00:43:24.547 17:59:06 spdkcli_nvmf_tcp -- common/autotest_common.sh@959 -- # uname 00:43:24.547 17:59:06 spdkcli_nvmf_tcp -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:43:24.547 17:59:06 spdkcli_nvmf_tcp -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 467350 00:43:24.547 17:59:06 spdkcli_nvmf_tcp -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:43:24.547 17:59:06 spdkcli_nvmf_tcp -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:43:24.547 17:59:06 spdkcli_nvmf_tcp -- common/autotest_common.sh@972 -- # echo 'killing process with pid 467350' 00:43:24.547 killing process with pid 467350 00:43:24.547 17:59:06 spdkcli_nvmf_tcp -- common/autotest_common.sh@973 -- # kill 467350 00:43:24.547 17:59:06 spdkcli_nvmf_tcp -- common/autotest_common.sh@978 -- # wait 467350 00:43:24.804 17:59:06 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@1 -- # cleanup 00:43:24.804 17:59:06 spdkcli_nvmf_tcp -- spdkcli/common.sh@10 -- # '[' -n '' ']' 00:43:24.804 17:59:06 spdkcli_nvmf_tcp -- spdkcli/common.sh@13 -- # '[' -n 467350 ']' 00:43:24.804 17:59:06 spdkcli_nvmf_tcp -- spdkcli/common.sh@14 -- # killprocess 467350 00:43:24.804 17:59:06 spdkcli_nvmf_tcp -- common/autotest_common.sh@954 -- # '[' -z 467350 ']' 00:43:24.804 17:59:06 spdkcli_nvmf_tcp -- common/autotest_common.sh@958 -- # kill -0 467350 00:43:24.804 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 958: kill: (467350) - No such process 00:43:24.804 17:59:06 spdkcli_nvmf_tcp -- common/autotest_common.sh@981 -- # echo 'Process with pid 467350 is not found' 00:43:24.804 Process with pid 467350 is not found 00:43:24.804 17:59:06 spdkcli_nvmf_tcp -- spdkcli/common.sh@16 -- # '[' -n '' ']' 00:43:24.805 17:59:06 spdkcli_nvmf_tcp -- spdkcli/common.sh@19 -- # '[' -n '' ']' 00:43:24.805 17:59:06 spdkcli_nvmf_tcp -- spdkcli/common.sh@22 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_nvmf.test /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/match_files/spdkcli_details_vhost.test /tmp/sample_aio 00:43:24.805 00:43:24.805 real 0m16.704s 00:43:24.805 user 0m35.639s 00:43:24.805 sys 0m0.878s 00:43:24.805 17:59:06 spdkcli_nvmf_tcp -- common/autotest_common.sh@1130 -- # xtrace_disable 00:43:24.805 17:59:06 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:43:24.805 ************************************ 00:43:24.805 END TEST spdkcli_nvmf_tcp 00:43:24.805 ************************************ 00:43:24.805 17:59:06 -- spdk/autotest.sh@287 -- # run_test nvmf_identify_passthru /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/identify_passthru.sh --transport=tcp 00:43:24.805 17:59:06 -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:43:24.805 17:59:06 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:43:24.805 17:59:06 -- common/autotest_common.sh@10 -- # set +x 00:43:24.805 ************************************ 00:43:24.805 START TEST nvmf_identify_passthru 00:43:24.805 ************************************ 00:43:24.805 17:59:06 nvmf_identify_passthru -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/identify_passthru.sh --transport=tcp 00:43:24.805 * Looking for test storage... 00:43:24.805 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:43:24.805 17:59:06 nvmf_identify_passthru -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:43:24.805 17:59:06 nvmf_identify_passthru -- common/autotest_common.sh@1711 -- # lcov --version 00:43:24.805 17:59:06 nvmf_identify_passthru -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:43:24.805 17:59:06 nvmf_identify_passthru -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@333 -- # local ver1 ver1_l 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@334 -- # local ver2 ver2_l 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@336 -- # IFS=.-: 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@336 -- # read -ra ver1 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@337 -- # IFS=.-: 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@337 -- # read -ra ver2 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@338 -- # local 'op=<' 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@340 -- # ver1_l=2 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@341 -- # ver2_l=1 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@344 -- # case "$op" in 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@345 -- # : 1 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@364 -- # (( v = 0 )) 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@365 -- # decimal 1 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@353 -- # local d=1 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@355 -- # echo 1 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@365 -- # ver1[v]=1 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@366 -- # decimal 2 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@353 -- # local d=2 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@355 -- # echo 2 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@366 -- # ver2[v]=2 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@368 -- # return 0 00:43:24.805 17:59:06 nvmf_identify_passthru -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:43:24.805 17:59:06 nvmf_identify_passthru -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:43:24.805 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:43:24.805 --rc genhtml_branch_coverage=1 00:43:24.805 --rc genhtml_function_coverage=1 00:43:24.805 --rc genhtml_legend=1 00:43:24.805 --rc geninfo_all_blocks=1 00:43:24.805 --rc geninfo_unexecuted_blocks=1 00:43:24.805 00:43:24.805 ' 00:43:24.805 17:59:06 nvmf_identify_passthru -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:43:24.805 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:43:24.805 --rc genhtml_branch_coverage=1 00:43:24.805 --rc genhtml_function_coverage=1 00:43:24.805 --rc genhtml_legend=1 00:43:24.805 --rc geninfo_all_blocks=1 00:43:24.805 --rc geninfo_unexecuted_blocks=1 00:43:24.805 00:43:24.805 ' 00:43:24.805 17:59:06 nvmf_identify_passthru -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:43:24.805 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:43:24.805 --rc genhtml_branch_coverage=1 00:43:24.805 --rc genhtml_function_coverage=1 00:43:24.805 --rc genhtml_legend=1 00:43:24.805 --rc geninfo_all_blocks=1 00:43:24.805 --rc geninfo_unexecuted_blocks=1 00:43:24.805 00:43:24.805 ' 00:43:24.805 17:59:06 nvmf_identify_passthru -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:43:24.805 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:43:24.805 --rc genhtml_branch_coverage=1 00:43:24.805 --rc genhtml_function_coverage=1 00:43:24.805 --rc genhtml_legend=1 00:43:24.805 --rc geninfo_all_blocks=1 00:43:24.805 --rc geninfo_unexecuted_blocks=1 00:43:24.805 00:43:24.805 ' 00:43:24.805 17:59:06 nvmf_identify_passthru -- target/identify_passthru.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@7 -- # uname -s 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@15 -- # shopt -s extglob 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:43:24.805 17:59:06 nvmf_identify_passthru -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:43:24.805 17:59:06 nvmf_identify_passthru -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:43:24.805 17:59:06 nvmf_identify_passthru -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:43:24.805 17:59:06 nvmf_identify_passthru -- paths/export.sh@5 -- # export PATH 00:43:24.805 17:59:06 nvmf_identify_passthru -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@51 -- # : 0 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:43:24.805 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@55 -- # have_pci_nics=0 00:43:24.805 17:59:06 nvmf_identify_passthru -- target/identify_passthru.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@15 -- # shopt -s extglob 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:43:24.805 17:59:06 nvmf_identify_passthru -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:43:24.805 17:59:06 nvmf_identify_passthru -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:43:24.805 17:59:06 nvmf_identify_passthru -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:43:24.805 17:59:06 nvmf_identify_passthru -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:43:24.805 17:59:06 nvmf_identify_passthru -- paths/export.sh@5 -- # export PATH 00:43:24.805 17:59:06 nvmf_identify_passthru -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:43:24.805 17:59:06 nvmf_identify_passthru -- target/identify_passthru.sh@12 -- # nvmftestinit 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@476 -- # prepare_net_devs 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@438 -- # local -g is_hw=no 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@440 -- # remove_spdk_ns 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:43:24.805 17:59:06 nvmf_identify_passthru -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:43:24.805 17:59:06 nvmf_identify_passthru -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:43:24.805 17:59:06 nvmf_identify_passthru -- nvmf/common.sh@309 -- # xtrace_disable 00:43:24.805 17:59:06 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@315 -- # pci_devs=() 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@315 -- # local -a pci_devs 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@316 -- # pci_net_devs=() 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@317 -- # pci_drivers=() 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@317 -- # local -A pci_drivers 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@319 -- # net_devs=() 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@319 -- # local -ga net_devs 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@320 -- # e810=() 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@320 -- # local -ga e810 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@321 -- # x722=() 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@321 -- # local -ga x722 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@322 -- # mlx=() 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@322 -- # local -ga mlx 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:43:27.340 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:43:27.340 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@418 -- # [[ up == up ]] 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:43:27.340 Found net devices under 0000:0a:00.0: cvl_0_0 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@418 -- # [[ up == up ]] 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:43:27.340 Found net devices under 0000:0a:00.1: cvl_0_1 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@442 -- # is_hw=yes 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:43:27.340 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:43:27.341 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:43:27.341 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:43:27.341 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:43:27.341 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:43:27.341 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:43:27.341 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:43:27.341 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:43:27.341 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:43:27.341 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:43:27.341 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:43:27.341 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:43:27.341 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.171 ms 00:43:27.341 00:43:27.341 --- 10.0.0.2 ping statistics --- 00:43:27.341 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:43:27.341 rtt min/avg/max/mdev = 0.171/0.171/0.171/0.000 ms 00:43:27.341 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:43:27.341 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:43:27.341 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.100 ms 00:43:27.341 00:43:27.341 --- 10.0.0.1 ping statistics --- 00:43:27.341 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:43:27.341 rtt min/avg/max/mdev = 0.100/0.100/0.100/0.000 ms 00:43:27.341 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:43:27.341 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@450 -- # return 0 00:43:27.341 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@478 -- # '[' '' == iso ']' 00:43:27.341 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:43:27.341 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:43:27.341 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:43:27.341 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:43:27.341 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:43:27.341 17:59:08 nvmf_identify_passthru -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:43:27.341 17:59:08 nvmf_identify_passthru -- target/identify_passthru.sh@14 -- # timing_enter nvme_identify 00:43:27.341 17:59:08 nvmf_identify_passthru -- common/autotest_common.sh@726 -- # xtrace_disable 00:43:27.341 17:59:08 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:43:27.341 17:59:08 nvmf_identify_passthru -- target/identify_passthru.sh@16 -- # get_first_nvme_bdf 00:43:27.341 17:59:08 nvmf_identify_passthru -- common/autotest_common.sh@1509 -- # bdfs=() 00:43:27.341 17:59:08 nvmf_identify_passthru -- common/autotest_common.sh@1509 -- # local bdfs 00:43:27.341 17:59:08 nvmf_identify_passthru -- common/autotest_common.sh@1510 -- # bdfs=($(get_nvme_bdfs)) 00:43:27.341 17:59:08 nvmf_identify_passthru -- common/autotest_common.sh@1510 -- # get_nvme_bdfs 00:43:27.341 17:59:08 nvmf_identify_passthru -- common/autotest_common.sh@1498 -- # bdfs=() 00:43:27.341 17:59:08 nvmf_identify_passthru -- common/autotest_common.sh@1498 -- # local bdfs 00:43:27.341 17:59:08 nvmf_identify_passthru -- common/autotest_common.sh@1499 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:43:27.341 17:59:08 nvmf_identify_passthru -- common/autotest_common.sh@1499 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh 00:43:27.341 17:59:08 nvmf_identify_passthru -- common/autotest_common.sh@1499 -- # jq -r '.config[].params.traddr' 00:43:27.341 17:59:08 nvmf_identify_passthru -- common/autotest_common.sh@1500 -- # (( 1 == 0 )) 00:43:27.341 17:59:08 nvmf_identify_passthru -- common/autotest_common.sh@1504 -- # printf '%s\n' 0000:88:00.0 00:43:27.341 17:59:08 nvmf_identify_passthru -- common/autotest_common.sh@1512 -- # echo 0000:88:00.0 00:43:27.341 17:59:08 nvmf_identify_passthru -- target/identify_passthru.sh@16 -- # bdf=0000:88:00.0 00:43:27.341 17:59:08 nvmf_identify_passthru -- target/identify_passthru.sh@17 -- # '[' -z 0000:88:00.0 ']' 00:43:27.341 17:59:08 nvmf_identify_passthru -- target/identify_passthru.sh@23 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r 'trtype:PCIe traddr:0000:88:00.0' -i 0 00:43:27.341 17:59:08 nvmf_identify_passthru -- target/identify_passthru.sh@23 -- # grep 'Serial Number:' 00:43:27.341 17:59:08 nvmf_identify_passthru -- target/identify_passthru.sh@23 -- # awk '{print $3}' 00:43:31.630 17:59:13 nvmf_identify_passthru -- target/identify_passthru.sh@23 -- # nvme_serial_number=PHLJ916004901P0FGN 00:43:31.630 17:59:13 nvmf_identify_passthru -- target/identify_passthru.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r 'trtype:PCIe traddr:0000:88:00.0' -i 0 00:43:31.630 17:59:13 nvmf_identify_passthru -- target/identify_passthru.sh@24 -- # awk '{print $3}' 00:43:31.630 17:59:13 nvmf_identify_passthru -- target/identify_passthru.sh@24 -- # grep 'Model Number:' 00:43:35.816 17:59:17 nvmf_identify_passthru -- target/identify_passthru.sh@24 -- # nvme_model_number=INTEL 00:43:35.816 17:59:17 nvmf_identify_passthru -- target/identify_passthru.sh@26 -- # timing_exit nvme_identify 00:43:35.816 17:59:17 nvmf_identify_passthru -- common/autotest_common.sh@732 -- # xtrace_disable 00:43:35.816 17:59:17 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:43:35.816 17:59:17 nvmf_identify_passthru -- target/identify_passthru.sh@28 -- # timing_enter start_nvmf_tgt 00:43:35.816 17:59:17 nvmf_identify_passthru -- common/autotest_common.sh@726 -- # xtrace_disable 00:43:35.816 17:59:17 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:43:35.816 17:59:17 nvmf_identify_passthru -- target/identify_passthru.sh@31 -- # nvmfpid=471990 00:43:35.817 17:59:17 nvmf_identify_passthru -- target/identify_passthru.sh@30 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF --wait-for-rpc 00:43:35.817 17:59:17 nvmf_identify_passthru -- target/identify_passthru.sh@33 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:43:35.817 17:59:17 nvmf_identify_passthru -- target/identify_passthru.sh@35 -- # waitforlisten 471990 00:43:35.817 17:59:17 nvmf_identify_passthru -- common/autotest_common.sh@835 -- # '[' -z 471990 ']' 00:43:35.817 17:59:17 nvmf_identify_passthru -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:43:35.817 17:59:17 nvmf_identify_passthru -- common/autotest_common.sh@840 -- # local max_retries=100 00:43:35.817 17:59:17 nvmf_identify_passthru -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:43:35.817 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:43:35.817 17:59:17 nvmf_identify_passthru -- common/autotest_common.sh@844 -- # xtrace_disable 00:43:35.817 17:59:17 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:43:35.817 [2024-12-06 17:59:17.479503] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:43:35.817 [2024-12-06 17:59:17.479593] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:43:35.817 [2024-12-06 17:59:17.553641] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:43:35.817 [2024-12-06 17:59:17.601087] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:43:35.817 [2024-12-06 17:59:17.601142] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:43:35.817 [2024-12-06 17:59:17.601167] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:43:35.817 [2024-12-06 17:59:17.601193] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:43:35.817 [2024-12-06 17:59:17.601203] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:43:35.817 [2024-12-06 17:59:17.602621] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:43:35.817 [2024-12-06 17:59:17.602687] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:43:35.817 [2024-12-06 17:59:17.602752] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:43:35.817 [2024-12-06 17:59:17.602755] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:43:36.075 17:59:17 nvmf_identify_passthru -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:43:36.075 17:59:17 nvmf_identify_passthru -- common/autotest_common.sh@868 -- # return 0 00:43:36.075 17:59:17 nvmf_identify_passthru -- target/identify_passthru.sh@36 -- # rpc_cmd -v nvmf_set_config --passthru-identify-ctrlr 00:43:36.075 17:59:17 nvmf_identify_passthru -- common/autotest_common.sh@563 -- # xtrace_disable 00:43:36.075 17:59:17 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:43:36.075 INFO: Log level set to 20 00:43:36.075 INFO: Requests: 00:43:36.075 { 00:43:36.075 "jsonrpc": "2.0", 00:43:36.075 "method": "nvmf_set_config", 00:43:36.075 "id": 1, 00:43:36.075 "params": { 00:43:36.075 "admin_cmd_passthru": { 00:43:36.075 "identify_ctrlr": true 00:43:36.075 } 00:43:36.075 } 00:43:36.075 } 00:43:36.075 00:43:36.075 INFO: response: 00:43:36.075 { 00:43:36.075 "jsonrpc": "2.0", 00:43:36.075 "id": 1, 00:43:36.075 "result": true 00:43:36.075 } 00:43:36.075 00:43:36.075 17:59:17 nvmf_identify_passthru -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:43:36.075 17:59:17 nvmf_identify_passthru -- target/identify_passthru.sh@37 -- # rpc_cmd -v framework_start_init 00:43:36.075 17:59:17 nvmf_identify_passthru -- common/autotest_common.sh@563 -- # xtrace_disable 00:43:36.075 17:59:17 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:43:36.075 INFO: Setting log level to 20 00:43:36.075 INFO: Setting log level to 20 00:43:36.075 INFO: Log level set to 20 00:43:36.075 INFO: Log level set to 20 00:43:36.075 INFO: Requests: 00:43:36.075 { 00:43:36.075 "jsonrpc": "2.0", 00:43:36.075 "method": "framework_start_init", 00:43:36.075 "id": 1 00:43:36.075 } 00:43:36.075 00:43:36.075 INFO: Requests: 00:43:36.075 { 00:43:36.075 "jsonrpc": "2.0", 00:43:36.075 "method": "framework_start_init", 00:43:36.075 "id": 1 00:43:36.075 } 00:43:36.075 00:43:36.075 [2024-12-06 17:59:17.813474] nvmf_tgt.c: 462:nvmf_tgt_advance_state: *NOTICE*: Custom identify ctrlr handler enabled 00:43:36.075 INFO: response: 00:43:36.075 { 00:43:36.075 "jsonrpc": "2.0", 00:43:36.075 "id": 1, 00:43:36.075 "result": true 00:43:36.075 } 00:43:36.075 00:43:36.075 INFO: response: 00:43:36.075 { 00:43:36.075 "jsonrpc": "2.0", 00:43:36.075 "id": 1, 00:43:36.075 "result": true 00:43:36.075 } 00:43:36.075 00:43:36.075 17:59:17 nvmf_identify_passthru -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:43:36.075 17:59:17 nvmf_identify_passthru -- target/identify_passthru.sh@38 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:43:36.075 17:59:17 nvmf_identify_passthru -- common/autotest_common.sh@563 -- # xtrace_disable 00:43:36.075 17:59:17 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:43:36.075 INFO: Setting log level to 40 00:43:36.075 INFO: Setting log level to 40 00:43:36.075 INFO: Setting log level to 40 00:43:36.075 [2024-12-06 17:59:17.823497] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:43:36.075 17:59:17 nvmf_identify_passthru -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:43:36.075 17:59:17 nvmf_identify_passthru -- target/identify_passthru.sh@39 -- # timing_exit start_nvmf_tgt 00:43:36.075 17:59:17 nvmf_identify_passthru -- common/autotest_common.sh@732 -- # xtrace_disable 00:43:36.075 17:59:17 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:43:36.075 17:59:17 nvmf_identify_passthru -- target/identify_passthru.sh@41 -- # rpc_cmd bdev_nvme_attach_controller -b Nvme0 -t PCIe -a 0000:88:00.0 00:43:36.075 17:59:17 nvmf_identify_passthru -- common/autotest_common.sh@563 -- # xtrace_disable 00:43:36.075 17:59:17 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:43:39.354 Nvme0n1 00:43:39.354 17:59:20 nvmf_identify_passthru -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:43:39.354 17:59:20 nvmf_identify_passthru -- target/identify_passthru.sh@42 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 1 00:43:39.354 17:59:20 nvmf_identify_passthru -- common/autotest_common.sh@563 -- # xtrace_disable 00:43:39.354 17:59:20 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:43:39.354 17:59:20 nvmf_identify_passthru -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:43:39.354 17:59:20 nvmf_identify_passthru -- target/identify_passthru.sh@43 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Nvme0n1 00:43:39.354 17:59:20 nvmf_identify_passthru -- common/autotest_common.sh@563 -- # xtrace_disable 00:43:39.354 17:59:20 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:43:39.354 17:59:20 nvmf_identify_passthru -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:43:39.354 17:59:20 nvmf_identify_passthru -- target/identify_passthru.sh@44 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:43:39.354 17:59:20 nvmf_identify_passthru -- common/autotest_common.sh@563 -- # xtrace_disable 00:43:39.354 17:59:20 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:43:39.354 [2024-12-06 17:59:20.716466] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:43:39.354 17:59:20 nvmf_identify_passthru -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:43:39.354 17:59:20 nvmf_identify_passthru -- target/identify_passthru.sh@46 -- # rpc_cmd nvmf_get_subsystems 00:43:39.354 17:59:20 nvmf_identify_passthru -- common/autotest_common.sh@563 -- # xtrace_disable 00:43:39.354 17:59:20 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:43:39.354 [ 00:43:39.354 { 00:43:39.354 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:43:39.354 "subtype": "Discovery", 00:43:39.354 "listen_addresses": [], 00:43:39.354 "allow_any_host": true, 00:43:39.354 "hosts": [] 00:43:39.354 }, 00:43:39.354 { 00:43:39.354 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:43:39.354 "subtype": "NVMe", 00:43:39.354 "listen_addresses": [ 00:43:39.354 { 00:43:39.354 "trtype": "TCP", 00:43:39.354 "adrfam": "IPv4", 00:43:39.354 "traddr": "10.0.0.2", 00:43:39.354 "trsvcid": "4420" 00:43:39.354 } 00:43:39.354 ], 00:43:39.354 "allow_any_host": true, 00:43:39.354 "hosts": [], 00:43:39.354 "serial_number": "SPDK00000000000001", 00:43:39.354 "model_number": "SPDK bdev Controller", 00:43:39.354 "max_namespaces": 1, 00:43:39.354 "min_cntlid": 1, 00:43:39.354 "max_cntlid": 65519, 00:43:39.354 "namespaces": [ 00:43:39.354 { 00:43:39.354 "nsid": 1, 00:43:39.354 "bdev_name": "Nvme0n1", 00:43:39.354 "name": "Nvme0n1", 00:43:39.354 "nguid": "7461B1993B874BD490DFF1A563A10E06", 00:43:39.354 "uuid": "7461b199-3b87-4bd4-90df-f1a563a10e06" 00:43:39.354 } 00:43:39.354 ] 00:43:39.354 } 00:43:39.354 ] 00:43:39.354 17:59:20 nvmf_identify_passthru -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:43:39.354 17:59:20 nvmf_identify_passthru -- target/identify_passthru.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:43:39.354 17:59:20 nvmf_identify_passthru -- target/identify_passthru.sh@54 -- # grep 'Serial Number:' 00:43:39.354 17:59:20 nvmf_identify_passthru -- target/identify_passthru.sh@54 -- # awk '{print $3}' 00:43:39.354 17:59:20 nvmf_identify_passthru -- target/identify_passthru.sh@54 -- # nvmf_serial_number=PHLJ916004901P0FGN 00:43:39.354 17:59:20 nvmf_identify_passthru -- target/identify_passthru.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:43:39.354 17:59:20 nvmf_identify_passthru -- target/identify_passthru.sh@61 -- # grep 'Model Number:' 00:43:39.354 17:59:20 nvmf_identify_passthru -- target/identify_passthru.sh@61 -- # awk '{print $3}' 00:43:39.611 17:59:21 nvmf_identify_passthru -- target/identify_passthru.sh@61 -- # nvmf_model_number=INTEL 00:43:39.611 17:59:21 nvmf_identify_passthru -- target/identify_passthru.sh@63 -- # '[' PHLJ916004901P0FGN '!=' PHLJ916004901P0FGN ']' 00:43:39.611 17:59:21 nvmf_identify_passthru -- target/identify_passthru.sh@68 -- # '[' INTEL '!=' INTEL ']' 00:43:39.611 17:59:21 nvmf_identify_passthru -- target/identify_passthru.sh@73 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:43:39.611 17:59:21 nvmf_identify_passthru -- common/autotest_common.sh@563 -- # xtrace_disable 00:43:39.611 17:59:21 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:43:39.611 17:59:21 nvmf_identify_passthru -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:43:39.611 17:59:21 nvmf_identify_passthru -- target/identify_passthru.sh@75 -- # trap - SIGINT SIGTERM EXIT 00:43:39.611 17:59:21 nvmf_identify_passthru -- target/identify_passthru.sh@77 -- # nvmftestfini 00:43:39.611 17:59:21 nvmf_identify_passthru -- nvmf/common.sh@516 -- # nvmfcleanup 00:43:39.611 17:59:21 nvmf_identify_passthru -- nvmf/common.sh@121 -- # sync 00:43:39.611 17:59:21 nvmf_identify_passthru -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:43:39.611 17:59:21 nvmf_identify_passthru -- nvmf/common.sh@124 -- # set +e 00:43:39.611 17:59:21 nvmf_identify_passthru -- nvmf/common.sh@125 -- # for i in {1..20} 00:43:39.611 17:59:21 nvmf_identify_passthru -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:43:39.611 rmmod nvme_tcp 00:43:39.611 rmmod nvme_fabrics 00:43:39.611 rmmod nvme_keyring 00:43:39.611 17:59:21 nvmf_identify_passthru -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:43:39.611 17:59:21 nvmf_identify_passthru -- nvmf/common.sh@128 -- # set -e 00:43:39.611 17:59:21 nvmf_identify_passthru -- nvmf/common.sh@129 -- # return 0 00:43:39.611 17:59:21 nvmf_identify_passthru -- nvmf/common.sh@517 -- # '[' -n 471990 ']' 00:43:39.611 17:59:21 nvmf_identify_passthru -- nvmf/common.sh@518 -- # killprocess 471990 00:43:39.611 17:59:21 nvmf_identify_passthru -- common/autotest_common.sh@954 -- # '[' -z 471990 ']' 00:43:39.611 17:59:21 nvmf_identify_passthru -- common/autotest_common.sh@958 -- # kill -0 471990 00:43:39.611 17:59:21 nvmf_identify_passthru -- common/autotest_common.sh@959 -- # uname 00:43:39.611 17:59:21 nvmf_identify_passthru -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:43:39.611 17:59:21 nvmf_identify_passthru -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 471990 00:43:39.611 17:59:21 nvmf_identify_passthru -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:43:39.611 17:59:21 nvmf_identify_passthru -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:43:39.611 17:59:21 nvmf_identify_passthru -- common/autotest_common.sh@972 -- # echo 'killing process with pid 471990' 00:43:39.611 killing process with pid 471990 00:43:39.611 17:59:21 nvmf_identify_passthru -- common/autotest_common.sh@973 -- # kill 471990 00:43:39.611 17:59:21 nvmf_identify_passthru -- common/autotest_common.sh@978 -- # wait 471990 00:43:41.507 17:59:22 nvmf_identify_passthru -- nvmf/common.sh@520 -- # '[' '' == iso ']' 00:43:41.507 17:59:22 nvmf_identify_passthru -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:43:41.507 17:59:22 nvmf_identify_passthru -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:43:41.507 17:59:22 nvmf_identify_passthru -- nvmf/common.sh@297 -- # iptr 00:43:41.507 17:59:22 nvmf_identify_passthru -- nvmf/common.sh@791 -- # iptables-save 00:43:41.507 17:59:22 nvmf_identify_passthru -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:43:41.507 17:59:22 nvmf_identify_passthru -- nvmf/common.sh@791 -- # iptables-restore 00:43:41.507 17:59:22 nvmf_identify_passthru -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:43:41.507 17:59:22 nvmf_identify_passthru -- nvmf/common.sh@302 -- # remove_spdk_ns 00:43:41.507 17:59:22 nvmf_identify_passthru -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:43:41.507 17:59:22 nvmf_identify_passthru -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:43:41.507 17:59:22 nvmf_identify_passthru -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:43:43.414 17:59:24 nvmf_identify_passthru -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:43:43.414 00:43:43.414 real 0m18.471s 00:43:43.414 user 0m27.792s 00:43:43.414 sys 0m2.491s 00:43:43.414 17:59:24 nvmf_identify_passthru -- common/autotest_common.sh@1130 -- # xtrace_disable 00:43:43.414 17:59:24 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:43:43.414 ************************************ 00:43:43.414 END TEST nvmf_identify_passthru 00:43:43.414 ************************************ 00:43:43.414 17:59:24 -- spdk/autotest.sh@289 -- # run_test nvmf_dif /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/dif.sh 00:43:43.414 17:59:24 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:43:43.414 17:59:24 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:43:43.414 17:59:24 -- common/autotest_common.sh@10 -- # set +x 00:43:43.414 ************************************ 00:43:43.414 START TEST nvmf_dif 00:43:43.414 ************************************ 00:43:43.414 17:59:24 nvmf_dif -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/dif.sh 00:43:43.414 * Looking for test storage... 00:43:43.414 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:43:43.414 17:59:25 nvmf_dif -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:43:43.414 17:59:25 nvmf_dif -- common/autotest_common.sh@1711 -- # lcov --version 00:43:43.414 17:59:25 nvmf_dif -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:43:43.414 17:59:25 nvmf_dif -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:43:43.414 17:59:25 nvmf_dif -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:43:43.414 17:59:25 nvmf_dif -- scripts/common.sh@333 -- # local ver1 ver1_l 00:43:43.414 17:59:25 nvmf_dif -- scripts/common.sh@334 -- # local ver2 ver2_l 00:43:43.414 17:59:25 nvmf_dif -- scripts/common.sh@336 -- # IFS=.-: 00:43:43.414 17:59:25 nvmf_dif -- scripts/common.sh@336 -- # read -ra ver1 00:43:43.414 17:59:25 nvmf_dif -- scripts/common.sh@337 -- # IFS=.-: 00:43:43.414 17:59:25 nvmf_dif -- scripts/common.sh@337 -- # read -ra ver2 00:43:43.414 17:59:25 nvmf_dif -- scripts/common.sh@338 -- # local 'op=<' 00:43:43.414 17:59:25 nvmf_dif -- scripts/common.sh@340 -- # ver1_l=2 00:43:43.414 17:59:25 nvmf_dif -- scripts/common.sh@341 -- # ver2_l=1 00:43:43.414 17:59:25 nvmf_dif -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:43:43.414 17:59:25 nvmf_dif -- scripts/common.sh@344 -- # case "$op" in 00:43:43.414 17:59:25 nvmf_dif -- scripts/common.sh@345 -- # : 1 00:43:43.414 17:59:25 nvmf_dif -- scripts/common.sh@364 -- # (( v = 0 )) 00:43:43.414 17:59:25 nvmf_dif -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:43:43.414 17:59:25 nvmf_dif -- scripts/common.sh@365 -- # decimal 1 00:43:43.414 17:59:25 nvmf_dif -- scripts/common.sh@353 -- # local d=1 00:43:43.414 17:59:25 nvmf_dif -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:43:43.414 17:59:25 nvmf_dif -- scripts/common.sh@355 -- # echo 1 00:43:43.414 17:59:25 nvmf_dif -- scripts/common.sh@365 -- # ver1[v]=1 00:43:43.414 17:59:25 nvmf_dif -- scripts/common.sh@366 -- # decimal 2 00:43:43.414 17:59:25 nvmf_dif -- scripts/common.sh@353 -- # local d=2 00:43:43.414 17:59:25 nvmf_dif -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:43:43.414 17:59:25 nvmf_dif -- scripts/common.sh@355 -- # echo 2 00:43:43.414 17:59:25 nvmf_dif -- scripts/common.sh@366 -- # ver2[v]=2 00:43:43.414 17:59:25 nvmf_dif -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:43:43.414 17:59:25 nvmf_dif -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:43:43.414 17:59:25 nvmf_dif -- scripts/common.sh@368 -- # return 0 00:43:43.414 17:59:25 nvmf_dif -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:43:43.414 17:59:25 nvmf_dif -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:43:43.414 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:43:43.414 --rc genhtml_branch_coverage=1 00:43:43.414 --rc genhtml_function_coverage=1 00:43:43.414 --rc genhtml_legend=1 00:43:43.414 --rc geninfo_all_blocks=1 00:43:43.414 --rc geninfo_unexecuted_blocks=1 00:43:43.414 00:43:43.414 ' 00:43:43.414 17:59:25 nvmf_dif -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:43:43.414 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:43:43.414 --rc genhtml_branch_coverage=1 00:43:43.414 --rc genhtml_function_coverage=1 00:43:43.414 --rc genhtml_legend=1 00:43:43.414 --rc geninfo_all_blocks=1 00:43:43.414 --rc geninfo_unexecuted_blocks=1 00:43:43.414 00:43:43.414 ' 00:43:43.414 17:59:25 nvmf_dif -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:43:43.414 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:43:43.414 --rc genhtml_branch_coverage=1 00:43:43.414 --rc genhtml_function_coverage=1 00:43:43.414 --rc genhtml_legend=1 00:43:43.414 --rc geninfo_all_blocks=1 00:43:43.414 --rc geninfo_unexecuted_blocks=1 00:43:43.414 00:43:43.414 ' 00:43:43.414 17:59:25 nvmf_dif -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:43:43.414 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:43:43.414 --rc genhtml_branch_coverage=1 00:43:43.414 --rc genhtml_function_coverage=1 00:43:43.415 --rc genhtml_legend=1 00:43:43.415 --rc geninfo_all_blocks=1 00:43:43.415 --rc geninfo_unexecuted_blocks=1 00:43:43.415 00:43:43.415 ' 00:43:43.415 17:59:25 nvmf_dif -- target/dif.sh@13 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@7 -- # uname -s 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:43:43.415 17:59:25 nvmf_dif -- scripts/common.sh@15 -- # shopt -s extglob 00:43:43.415 17:59:25 nvmf_dif -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:43:43.415 17:59:25 nvmf_dif -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:43:43.415 17:59:25 nvmf_dif -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:43:43.415 17:59:25 nvmf_dif -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:43:43.415 17:59:25 nvmf_dif -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:43:43.415 17:59:25 nvmf_dif -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:43:43.415 17:59:25 nvmf_dif -- paths/export.sh@5 -- # export PATH 00:43:43.415 17:59:25 nvmf_dif -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@51 -- # : 0 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:43:43.415 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@55 -- # have_pci_nics=0 00:43:43.415 17:59:25 nvmf_dif -- target/dif.sh@15 -- # NULL_META=16 00:43:43.415 17:59:25 nvmf_dif -- target/dif.sh@15 -- # NULL_BLOCK_SIZE=512 00:43:43.415 17:59:25 nvmf_dif -- target/dif.sh@15 -- # NULL_SIZE=64 00:43:43.415 17:59:25 nvmf_dif -- target/dif.sh@15 -- # NULL_DIF=1 00:43:43.415 17:59:25 nvmf_dif -- target/dif.sh@135 -- # nvmftestinit 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@476 -- # prepare_net_devs 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@438 -- # local -g is_hw=no 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@440 -- # remove_spdk_ns 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:43:43.415 17:59:25 nvmf_dif -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:43:43.415 17:59:25 nvmf_dif -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:43:43.415 17:59:25 nvmf_dif -- nvmf/common.sh@309 -- # xtrace_disable 00:43:43.415 17:59:25 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@315 -- # pci_devs=() 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@315 -- # local -a pci_devs 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@316 -- # pci_net_devs=() 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@317 -- # pci_drivers=() 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@317 -- # local -A pci_drivers 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@319 -- # net_devs=() 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@319 -- # local -ga net_devs 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@320 -- # e810=() 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@320 -- # local -ga e810 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@321 -- # x722=() 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@321 -- # local -ga x722 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@322 -- # mlx=() 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@322 -- # local -ga mlx 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:43:45.950 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:43:45.950 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:43:45.950 17:59:27 nvmf_dif -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@418 -- # [[ up == up ]] 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:43:45.951 Found net devices under 0000:0a:00.0: cvl_0_0 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@418 -- # [[ up == up ]] 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:43:45.951 Found net devices under 0000:0a:00.1: cvl_0_1 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@442 -- # is_hw=yes 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:43:45.951 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:43:45.951 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.167 ms 00:43:45.951 00:43:45.951 --- 10.0.0.2 ping statistics --- 00:43:45.951 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:43:45.951 rtt min/avg/max/mdev = 0.167/0.167/0.167/0.000 ms 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:43:45.951 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:43:45.951 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.159 ms 00:43:45.951 00:43:45.951 --- 10.0.0.1 ping statistics --- 00:43:45.951 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:43:45.951 rtt min/avg/max/mdev = 0.159/0.159/0.159/0.000 ms 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@450 -- # return 0 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@478 -- # '[' iso == iso ']' 00:43:45.951 17:59:27 nvmf_dif -- nvmf/common.sh@479 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:43:46.886 0000:00:04.7 (8086 0e27): Already using the vfio-pci driver 00:43:46.886 0000:88:00.0 (8086 0a54): Already using the vfio-pci driver 00:43:46.886 0000:00:04.6 (8086 0e26): Already using the vfio-pci driver 00:43:46.886 0000:00:04.5 (8086 0e25): Already using the vfio-pci driver 00:43:46.886 0000:00:04.4 (8086 0e24): Already using the vfio-pci driver 00:43:46.886 0000:00:04.3 (8086 0e23): Already using the vfio-pci driver 00:43:46.886 0000:00:04.2 (8086 0e22): Already using the vfio-pci driver 00:43:46.886 0000:00:04.1 (8086 0e21): Already using the vfio-pci driver 00:43:46.886 0000:00:04.0 (8086 0e20): Already using the vfio-pci driver 00:43:46.886 0000:80:04.7 (8086 0e27): Already using the vfio-pci driver 00:43:46.886 0000:80:04.6 (8086 0e26): Already using the vfio-pci driver 00:43:46.886 0000:80:04.5 (8086 0e25): Already using the vfio-pci driver 00:43:46.886 0000:80:04.4 (8086 0e24): Already using the vfio-pci driver 00:43:46.886 0000:80:04.3 (8086 0e23): Already using the vfio-pci driver 00:43:46.886 0000:80:04.2 (8086 0e22): Already using the vfio-pci driver 00:43:46.886 0000:80:04.1 (8086 0e21): Already using the vfio-pci driver 00:43:46.886 0000:80:04.0 (8086 0e20): Already using the vfio-pci driver 00:43:47.144 17:59:28 nvmf_dif -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:43:47.144 17:59:28 nvmf_dif -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:43:47.144 17:59:28 nvmf_dif -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:43:47.144 17:59:28 nvmf_dif -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:43:47.144 17:59:28 nvmf_dif -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:43:47.144 17:59:28 nvmf_dif -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:43:47.144 17:59:28 nvmf_dif -- target/dif.sh@136 -- # NVMF_TRANSPORT_OPTS+=' --dif-insert-or-strip' 00:43:47.144 17:59:28 nvmf_dif -- target/dif.sh@137 -- # nvmfappstart 00:43:47.144 17:59:28 nvmf_dif -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:43:47.144 17:59:28 nvmf_dif -- common/autotest_common.sh@726 -- # xtrace_disable 00:43:47.144 17:59:28 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:43:47.144 17:59:28 nvmf_dif -- nvmf/common.sh@509 -- # nvmfpid=475257 00:43:47.145 17:59:28 nvmf_dif -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF 00:43:47.145 17:59:28 nvmf_dif -- nvmf/common.sh@510 -- # waitforlisten 475257 00:43:47.145 17:59:28 nvmf_dif -- common/autotest_common.sh@835 -- # '[' -z 475257 ']' 00:43:47.145 17:59:28 nvmf_dif -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:43:47.145 17:59:28 nvmf_dif -- common/autotest_common.sh@840 -- # local max_retries=100 00:43:47.145 17:59:28 nvmf_dif -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:43:47.145 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:43:47.145 17:59:28 nvmf_dif -- common/autotest_common.sh@844 -- # xtrace_disable 00:43:47.145 17:59:28 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:43:47.145 [2024-12-06 17:59:28.847580] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:43:47.145 [2024-12-06 17:59:28.847686] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:43:47.145 [2024-12-06 17:59:28.930259] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:43:47.145 [2024-12-06 17:59:28.975936] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:43:47.145 [2024-12-06 17:59:28.975999] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:43:47.145 [2024-12-06 17:59:28.976013] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:43:47.145 [2024-12-06 17:59:28.976025] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:43:47.145 [2024-12-06 17:59:28.976034] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:43:47.145 [2024-12-06 17:59:28.976597] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:43:47.403 17:59:29 nvmf_dif -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:43:47.403 17:59:29 nvmf_dif -- common/autotest_common.sh@868 -- # return 0 00:43:47.403 17:59:29 nvmf_dif -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:43:47.403 17:59:29 nvmf_dif -- common/autotest_common.sh@732 -- # xtrace_disable 00:43:47.403 17:59:29 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:43:47.403 17:59:29 nvmf_dif -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:43:47.403 17:59:29 nvmf_dif -- target/dif.sh@139 -- # create_transport 00:43:47.403 17:59:29 nvmf_dif -- target/dif.sh@50 -- # rpc_cmd nvmf_create_transport -t tcp -o --dif-insert-or-strip 00:43:47.403 17:59:29 nvmf_dif -- common/autotest_common.sh@563 -- # xtrace_disable 00:43:47.403 17:59:29 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:43:47.403 [2024-12-06 17:59:29.124370] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:43:47.403 17:59:29 nvmf_dif -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:43:47.403 17:59:29 nvmf_dif -- target/dif.sh@141 -- # run_test fio_dif_1_default fio_dif_1 00:43:47.403 17:59:29 nvmf_dif -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:43:47.403 17:59:29 nvmf_dif -- common/autotest_common.sh@1111 -- # xtrace_disable 00:43:47.403 17:59:29 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:43:47.403 ************************************ 00:43:47.403 START TEST fio_dif_1_default 00:43:47.403 ************************************ 00:43:47.403 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1129 -- # fio_dif_1 00:43:47.403 17:59:29 nvmf_dif.fio_dif_1_default -- target/dif.sh@86 -- # create_subsystems 0 00:43:47.403 17:59:29 nvmf_dif.fio_dif_1_default -- target/dif.sh@28 -- # local sub 00:43:47.403 17:59:29 nvmf_dif.fio_dif_1_default -- target/dif.sh@30 -- # for sub in "$@" 00:43:47.403 17:59:29 nvmf_dif.fio_dif_1_default -- target/dif.sh@31 -- # create_subsystem 0 00:43:47.403 17:59:29 nvmf_dif.fio_dif_1_default -- target/dif.sh@18 -- # local sub_id=0 00:43:47.403 17:59:29 nvmf_dif.fio_dif_1_default -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 1 00:43:47.403 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@563 -- # xtrace_disable 00:43:47.403 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:43:47.403 bdev_null0 00:43:47.403 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@563 -- # xtrace_disable 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@563 -- # xtrace_disable 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@563 -- # xtrace_disable 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:43:47.404 [2024-12-06 17:59:29.188739] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- target/dif.sh@87 -- # fio /dev/fd/62 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- target/dif.sh@87 -- # create_json_sub_conf 0 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- target/dif.sh@51 -- # gen_nvmf_target_json 0 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1360 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@560 -- # config=() 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- target/dif.sh@82 -- # gen_fio_conf 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1341 -- # local fio_dir=/usr/src/fio 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@560 -- # local subsystem config 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1343 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- target/dif.sh@54 -- # local file 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1343 -- # local sanitizers 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1344 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:43:47.404 { 00:43:47.404 "params": { 00:43:47.404 "name": "Nvme$subsystem", 00:43:47.404 "trtype": "$TEST_TRANSPORT", 00:43:47.404 "traddr": "$NVMF_FIRST_TARGET_IP", 00:43:47.404 "adrfam": "ipv4", 00:43:47.404 "trsvcid": "$NVMF_PORT", 00:43:47.404 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:43:47.404 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:43:47.404 "hdgst": ${hdgst:-false}, 00:43:47.404 "ddgst": ${ddgst:-false} 00:43:47.404 }, 00:43:47.404 "method": "bdev_nvme_attach_controller" 00:43:47.404 } 00:43:47.404 EOF 00:43:47.404 )") 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- target/dif.sh@56 -- # cat 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1345 -- # shift 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1347 -- # local asan_lib= 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@582 -- # cat 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1349 -- # grep libasan 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- target/dif.sh@72 -- # (( file = 1 )) 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- target/dif.sh@72 -- # (( file <= files )) 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@584 -- # jq . 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@585 -- # IFS=, 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:43:47.404 "params": { 00:43:47.404 "name": "Nvme0", 00:43:47.404 "trtype": "tcp", 00:43:47.404 "traddr": "10.0.0.2", 00:43:47.404 "adrfam": "ipv4", 00:43:47.404 "trsvcid": "4420", 00:43:47.404 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:43:47.404 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:43:47.404 "hdgst": false, 00:43:47.404 "ddgst": false 00:43:47.404 }, 00:43:47.404 "method": "bdev_nvme_attach_controller" 00:43:47.404 }' 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1349 -- # asan_lib= 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1349 -- # grep libclang_rt.asan 00:43:47.404 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:43:47.662 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1349 -- # asan_lib= 00:43:47.662 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:43:47.662 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1356 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:43:47.662 17:59:29 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1356 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:43:47.662 filename0: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=4 00:43:47.662 fio-3.35 00:43:47.662 Starting 1 thread 00:43:59.859 00:43:59.859 filename0: (groupid=0, jobs=1): err= 0: pid=475481: Fri Dec 6 17:59:40 2024 00:43:59.859 read: IOPS=207, BW=828KiB/s (848kB/s)(8304KiB/10026msec) 00:43:59.859 slat (nsec): min=6660, max=70782, avg=8499.13, stdev=3099.48 00:43:59.859 clat (usec): min=532, max=44599, avg=19290.66, stdev=20361.72 00:43:59.859 lat (usec): min=539, max=44646, avg=19299.16, stdev=20361.59 00:43:59.859 clat percentiles (usec): 00:43:59.859 | 1.00th=[ 562], 5.00th=[ 578], 10.00th=[ 594], 20.00th=[ 611], 00:43:59.859 | 30.00th=[ 627], 40.00th=[ 652], 50.00th=[ 701], 60.00th=[41157], 00:43:59.859 | 70.00th=[41157], 80.00th=[41157], 90.00th=[41157], 95.00th=[42206], 00:43:59.859 | 99.00th=[42206], 99.50th=[42206], 99.90th=[44827], 99.95th=[44827], 00:43:59.859 | 99.99th=[44827] 00:43:59.859 bw ( KiB/s): min= 704, max= 1024, per=99.97%, avg=828.80, stdev=74.79, samples=20 00:43:59.859 iops : min= 176, max= 256, avg=207.20, stdev=18.70, samples=20 00:43:59.859 lat (usec) : 750=52.89%, 1000=1.45% 00:43:59.859 lat (msec) : 50=45.66% 00:43:59.859 cpu : usr=90.96%, sys=8.76%, ctx=17, majf=0, minf=272 00:43:59.859 IO depths : 1=25.0%, 2=50.0%, 4=25.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:43:59.859 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:43:59.859 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:43:59.859 issued rwts: total=2076,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:43:59.859 latency : target=0, window=0, percentile=100.00%, depth=4 00:43:59.859 00:43:59.859 Run status group 0 (all jobs): 00:43:59.859 READ: bw=828KiB/s (848kB/s), 828KiB/s-828KiB/s (848kB/s-848kB/s), io=8304KiB (8503kB), run=10026-10026msec 00:43:59.859 17:59:40 nvmf_dif.fio_dif_1_default -- target/dif.sh@88 -- # destroy_subsystems 0 00:43:59.859 17:59:40 nvmf_dif.fio_dif_1_default -- target/dif.sh@43 -- # local sub 00:43:59.859 17:59:40 nvmf_dif.fio_dif_1_default -- target/dif.sh@45 -- # for sub in "$@" 00:43:59.859 17:59:40 nvmf_dif.fio_dif_1_default -- target/dif.sh@46 -- # destroy_subsystem 0 00:43:59.859 17:59:40 nvmf_dif.fio_dif_1_default -- target/dif.sh@36 -- # local sub_id=0 00:43:59.859 17:59:40 nvmf_dif.fio_dif_1_default -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:43:59.859 17:59:40 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@563 -- # xtrace_disable 00:43:59.859 17:59:40 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:43:59.859 17:59:40 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:43:59.859 17:59:40 nvmf_dif.fio_dif_1_default -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:43:59.859 17:59:40 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@563 -- # xtrace_disable 00:43:59.859 17:59:40 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:43:59.859 17:59:40 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:43:59.859 00:43:59.859 real 0m11.176s 00:43:59.859 user 0m10.190s 00:43:59.859 sys 0m1.157s 00:43:59.859 17:59:40 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1130 -- # xtrace_disable 00:43:59.859 17:59:40 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:43:59.859 ************************************ 00:43:59.860 END TEST fio_dif_1_default 00:43:59.860 ************************************ 00:43:59.860 17:59:40 nvmf_dif -- target/dif.sh@142 -- # run_test fio_dif_1_multi_subsystems fio_dif_1_multi_subsystems 00:43:59.860 17:59:40 nvmf_dif -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:43:59.860 17:59:40 nvmf_dif -- common/autotest_common.sh@1111 -- # xtrace_disable 00:43:59.860 17:59:40 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:43:59.860 ************************************ 00:43:59.860 START TEST fio_dif_1_multi_subsystems 00:43:59.860 ************************************ 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1129 -- # fio_dif_1_multi_subsystems 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@92 -- # local files=1 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@94 -- # create_subsystems 0 1 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@28 -- # local sub 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@30 -- # for sub in "$@" 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@31 -- # create_subsystem 0 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@18 -- # local sub_id=0 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 1 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@563 -- # xtrace_disable 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:43:59.860 bdev_null0 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@563 -- # xtrace_disable 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@563 -- # xtrace_disable 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@563 -- # xtrace_disable 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:43:59.860 [2024-12-06 17:59:40.414881] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@30 -- # for sub in "$@" 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@31 -- # create_subsystem 1 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@18 -- # local sub_id=1 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null1 64 512 --md-size 16 --dif-type 1 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@563 -- # xtrace_disable 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:43:59.860 bdev_null1 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 --serial-number 53313233-1 --allow-any-host 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@563 -- # xtrace_disable 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 bdev_null1 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@563 -- # xtrace_disable 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@563 -- # xtrace_disable 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@95 -- # fio /dev/fd/62 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@95 -- # create_json_sub_conf 0 1 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@51 -- # gen_nvmf_target_json 0 1 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@560 -- # config=() 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@560 -- # local subsystem config 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:43:59.860 { 00:43:59.860 "params": { 00:43:59.860 "name": "Nvme$subsystem", 00:43:59.860 "trtype": "$TEST_TRANSPORT", 00:43:59.860 "traddr": "$NVMF_FIRST_TARGET_IP", 00:43:59.860 "adrfam": "ipv4", 00:43:59.860 "trsvcid": "$NVMF_PORT", 00:43:59.860 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:43:59.860 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:43:59.860 "hdgst": ${hdgst:-false}, 00:43:59.860 "ddgst": ${ddgst:-false} 00:43:59.860 }, 00:43:59.860 "method": "bdev_nvme_attach_controller" 00:43:59.860 } 00:43:59.860 EOF 00:43:59.860 )") 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@82 -- # gen_fio_conf 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1360 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@54 -- # local file 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1341 -- # local fio_dir=/usr/src/fio 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@56 -- # cat 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1343 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1343 -- # local sanitizers 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1344 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1345 -- # shift 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1347 -- # local asan_lib= 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@582 -- # cat 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@72 -- # (( file = 1 )) 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@72 -- # (( file <= files )) 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@73 -- # cat 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1349 -- # grep libasan 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:43:59.860 { 00:43:59.860 "params": { 00:43:59.860 "name": "Nvme$subsystem", 00:43:59.860 "trtype": "$TEST_TRANSPORT", 00:43:59.860 "traddr": "$NVMF_FIRST_TARGET_IP", 00:43:59.860 "adrfam": "ipv4", 00:43:59.860 "trsvcid": "$NVMF_PORT", 00:43:59.860 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:43:59.860 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:43:59.860 "hdgst": ${hdgst:-false}, 00:43:59.860 "ddgst": ${ddgst:-false} 00:43:59.860 }, 00:43:59.860 "method": "bdev_nvme_attach_controller" 00:43:59.860 } 00:43:59.860 EOF 00:43:59.860 )") 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@72 -- # (( file++ )) 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@72 -- # (( file <= files )) 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@582 -- # cat 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@584 -- # jq . 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@585 -- # IFS=, 00:43:59.860 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:43:59.860 "params": { 00:43:59.860 "name": "Nvme0", 00:43:59.860 "trtype": "tcp", 00:43:59.860 "traddr": "10.0.0.2", 00:43:59.860 "adrfam": "ipv4", 00:43:59.860 "trsvcid": "4420", 00:43:59.860 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:43:59.860 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:43:59.860 "hdgst": false, 00:43:59.860 "ddgst": false 00:43:59.860 }, 00:43:59.860 "method": "bdev_nvme_attach_controller" 00:43:59.860 },{ 00:43:59.860 "params": { 00:43:59.860 "name": "Nvme1", 00:43:59.860 "trtype": "tcp", 00:43:59.860 "traddr": "10.0.0.2", 00:43:59.860 "adrfam": "ipv4", 00:43:59.861 "trsvcid": "4420", 00:43:59.861 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:43:59.861 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:43:59.861 "hdgst": false, 00:43:59.861 "ddgst": false 00:43:59.861 }, 00:43:59.861 "method": "bdev_nvme_attach_controller" 00:43:59.861 }' 00:43:59.861 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1349 -- # asan_lib= 00:43:59.861 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:43:59.861 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:43:59.861 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:43:59.861 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1349 -- # grep libclang_rt.asan 00:43:59.861 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:43:59.861 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1349 -- # asan_lib= 00:43:59.861 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:43:59.861 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1356 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:43:59.861 17:59:40 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1356 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:43:59.861 filename0: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=4 00:43:59.861 filename1: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=4 00:43:59.861 fio-3.35 00:43:59.861 Starting 2 threads 00:44:09.820 00:44:09.820 filename0: (groupid=0, jobs=1): err= 0: pid=476994: Fri Dec 6 17:59:51 2024 00:44:09.820 read: IOPS=144, BW=577KiB/s (591kB/s)(5776KiB/10011msec) 00:44:09.820 slat (nsec): min=7417, max=31106, avg=9888.43, stdev=2714.04 00:44:09.820 clat (usec): min=551, max=44479, avg=27699.11, stdev=19057.22 00:44:09.820 lat (usec): min=559, max=44498, avg=27709.00, stdev=19057.14 00:44:09.820 clat percentiles (usec): 00:44:09.820 | 1.00th=[ 578], 5.00th=[ 611], 10.00th=[ 627], 20.00th=[ 668], 00:44:09.820 | 30.00th=[ 914], 40.00th=[41157], 50.00th=[41157], 60.00th=[41157], 00:44:09.820 | 70.00th=[41157], 80.00th=[41157], 90.00th=[41157], 95.00th=[41157], 00:44:09.820 | 99.00th=[42206], 99.50th=[42206], 99.90th=[44303], 99.95th=[44303], 00:44:09.820 | 99.99th=[44303] 00:44:09.820 bw ( KiB/s): min= 384, max= 832, per=59.72%, avg=576.05, stdev=187.80, samples=20 00:44:09.820 iops : min= 96, max= 208, avg=144.00, stdev=46.94, samples=20 00:44:09.820 lat (usec) : 750=25.14%, 1000=7.62% 00:44:09.820 lat (msec) : 2=0.48%, 50=66.76% 00:44:09.820 cpu : usr=95.33%, sys=4.37%, ctx=20, majf=0, minf=198 00:44:09.820 IO depths : 1=25.0%, 2=50.0%, 4=25.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:44:09.820 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:09.820 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:09.820 issued rwts: total=1444,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:09.820 latency : target=0, window=0, percentile=100.00%, depth=4 00:44:09.820 filename1: (groupid=0, jobs=1): err= 0: pid=476995: Fri Dec 6 17:59:51 2024 00:44:09.820 read: IOPS=97, BW=389KiB/s (398kB/s)(3904KiB/10037msec) 00:44:09.820 slat (nsec): min=7178, max=22109, avg=9972.32, stdev=2764.62 00:44:09.820 clat (usec): min=40747, max=44540, avg=41102.28, stdev=392.49 00:44:09.820 lat (usec): min=40756, max=44559, avg=41112.25, stdev=392.69 00:44:09.820 clat percentiles (usec): 00:44:09.820 | 1.00th=[40633], 5.00th=[41157], 10.00th=[41157], 20.00th=[41157], 00:44:09.820 | 30.00th=[41157], 40.00th=[41157], 50.00th=[41157], 60.00th=[41157], 00:44:09.820 | 70.00th=[41157], 80.00th=[41157], 90.00th=[41681], 95.00th=[42206], 00:44:09.820 | 99.00th=[42206], 99.50th=[42206], 99.90th=[44303], 99.95th=[44303], 00:44:09.820 | 99.99th=[44303] 00:44:09.820 bw ( KiB/s): min= 384, max= 416, per=40.23%, avg=388.80, stdev=11.72, samples=20 00:44:09.820 iops : min= 96, max= 104, avg=97.20, stdev= 2.93, samples=20 00:44:09.820 lat (msec) : 50=100.00% 00:44:09.820 cpu : usr=94.98%, sys=4.73%, ctx=13, majf=0, minf=78 00:44:09.820 IO depths : 1=25.0%, 2=50.0%, 4=25.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:44:09.820 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:09.820 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:09.820 issued rwts: total=976,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:09.820 latency : target=0, window=0, percentile=100.00%, depth=4 00:44:09.820 00:44:09.820 Run status group 0 (all jobs): 00:44:09.820 READ: bw=964KiB/s (988kB/s), 389KiB/s-577KiB/s (398kB/s-591kB/s), io=9680KiB (9912kB), run=10011-10037msec 00:44:10.080 17:59:51 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@96 -- # destroy_subsystems 0 1 00:44:10.080 17:59:51 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@43 -- # local sub 00:44:10.080 17:59:51 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@45 -- # for sub in "$@" 00:44:10.080 17:59:51 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@46 -- # destroy_subsystem 0 00:44:10.080 17:59:51 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@36 -- # local sub_id=0 00:44:10.080 17:59:51 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:44:10.080 17:59:51 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:10.080 17:59:51 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:44:10.080 17:59:51 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:10.080 17:59:51 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:44:10.080 17:59:51 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:10.080 17:59:51 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:44:10.080 17:59:51 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:10.080 17:59:51 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@45 -- # for sub in "$@" 00:44:10.080 17:59:51 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@46 -- # destroy_subsystem 1 00:44:10.080 17:59:51 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@36 -- # local sub_id=1 00:44:10.080 17:59:51 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:44:10.080 17:59:51 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:10.080 17:59:51 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:44:10.080 17:59:51 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:10.080 17:59:51 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null1 00:44:10.080 17:59:51 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:10.080 17:59:51 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:44:10.080 17:59:51 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:10.080 00:44:10.080 real 0m11.362s 00:44:10.080 user 0m20.470s 00:44:10.080 sys 0m1.197s 00:44:10.080 17:59:51 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1130 -- # xtrace_disable 00:44:10.080 17:59:51 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:44:10.080 ************************************ 00:44:10.080 END TEST fio_dif_1_multi_subsystems 00:44:10.080 ************************************ 00:44:10.080 17:59:51 nvmf_dif -- target/dif.sh@143 -- # run_test fio_dif_rand_params fio_dif_rand_params 00:44:10.080 17:59:51 nvmf_dif -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:44:10.080 17:59:51 nvmf_dif -- common/autotest_common.sh@1111 -- # xtrace_disable 00:44:10.080 17:59:51 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:44:10.080 ************************************ 00:44:10.080 START TEST fio_dif_rand_params 00:44:10.080 ************************************ 00:44:10.080 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1129 -- # fio_dif_rand_params 00:44:10.080 17:59:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@100 -- # local NULL_DIF 00:44:10.080 17:59:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@101 -- # local bs numjobs runtime iodepth files 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@103 -- # NULL_DIF=3 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@103 -- # bs=128k 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@103 -- # numjobs=3 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@103 -- # iodepth=3 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@103 -- # runtime=5 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@105 -- # create_subsystems 0 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@28 -- # local sub 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 0 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=0 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 3 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:10.081 bdev_null0 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:10.081 [2024-12-06 17:59:51.828741] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@106 -- # fio /dev/fd/62 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@106 -- # create_json_sub_conf 0 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@51 -- # gen_nvmf_target_json 0 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1360 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@560 -- # config=() 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # gen_fio_conf 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1341 -- # local fio_dir=/usr/src/fio 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@560 -- # local subsystem config 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1343 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@54 -- # local file 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1343 -- # local sanitizers 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@56 -- # cat 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1344 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:44:10.081 { 00:44:10.081 "params": { 00:44:10.081 "name": "Nvme$subsystem", 00:44:10.081 "trtype": "$TEST_TRANSPORT", 00:44:10.081 "traddr": "$NVMF_FIRST_TARGET_IP", 00:44:10.081 "adrfam": "ipv4", 00:44:10.081 "trsvcid": "$NVMF_PORT", 00:44:10.081 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:44:10.081 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:44:10.081 "hdgst": ${hdgst:-false}, 00:44:10.081 "ddgst": ${ddgst:-false} 00:44:10.081 }, 00:44:10.081 "method": "bdev_nvme_attach_controller" 00:44:10.081 } 00:44:10.081 EOF 00:44:10.081 )") 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # shift 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1347 -- # local asan_lib= 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # cat 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # grep libasan 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file = 1 )) 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@584 -- # jq . 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@585 -- # IFS=, 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:44:10.081 "params": { 00:44:10.081 "name": "Nvme0", 00:44:10.081 "trtype": "tcp", 00:44:10.081 "traddr": "10.0.0.2", 00:44:10.081 "adrfam": "ipv4", 00:44:10.081 "trsvcid": "4420", 00:44:10.081 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:44:10.081 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:44:10.081 "hdgst": false, 00:44:10.081 "ddgst": false 00:44:10.081 }, 00:44:10.081 "method": "bdev_nvme_attach_controller" 00:44:10.081 }' 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # asan_lib= 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # grep libclang_rt.asan 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # asan_lib= 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1356 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:44:10.081 17:59:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1356 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:44:10.338 filename0: (g=0): rw=randread, bs=(R) 128KiB-128KiB, (W) 128KiB-128KiB, (T) 128KiB-128KiB, ioengine=spdk_bdev, iodepth=3 00:44:10.339 ... 00:44:10.339 fio-3.35 00:44:10.339 Starting 3 threads 00:44:16.895 00:44:16.895 filename0: (groupid=0, jobs=1): err= 0: pid=478273: Fri Dec 6 17:59:57 2024 00:44:16.895 read: IOPS=232, BW=29.1MiB/s (30.5MB/s)(146MiB/5005msec) 00:44:16.895 slat (nsec): min=4656, max=41272, avg=15858.32, stdev=4326.04 00:44:16.895 clat (usec): min=8649, max=53746, avg=12866.24, stdev=2671.86 00:44:16.895 lat (usec): min=8667, max=53759, avg=12882.10, stdev=2671.80 00:44:16.896 clat percentiles (usec): 00:44:16.896 | 1.00th=[ 9503], 5.00th=[10290], 10.00th=[10814], 20.00th=[11338], 00:44:16.896 | 30.00th=[11731], 40.00th=[12125], 50.00th=[12649], 60.00th=[13042], 00:44:16.896 | 70.00th=[13435], 80.00th=[14222], 90.00th=[15139], 95.00th=[15795], 00:44:16.896 | 99.00th=[17433], 99.50th=[19268], 99.90th=[53216], 99.95th=[53740], 00:44:16.896 | 99.99th=[53740] 00:44:16.896 bw ( KiB/s): min=26368, max=32256, per=34.44%, avg=29772.80, stdev=1468.38, samples=10 00:44:16.896 iops : min= 206, max= 252, avg=232.60, stdev=11.47, samples=10 00:44:16.896 lat (msec) : 10=3.00%, 20=96.65%, 50=0.09%, 100=0.26% 00:44:16.896 cpu : usr=93.55%, sys=5.96%, ctx=12, majf=0, minf=68 00:44:16.896 IO depths : 1=0.1%, 2=99.9%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:44:16.896 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:16.896 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:16.896 issued rwts: total=1165,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:16.896 latency : target=0, window=0, percentile=100.00%, depth=3 00:44:16.896 filename0: (groupid=0, jobs=1): err= 0: pid=478274: Fri Dec 6 17:59:57 2024 00:44:16.896 read: IOPS=218, BW=27.3MiB/s (28.7MB/s)(138MiB/5045msec) 00:44:16.896 slat (nsec): min=4805, max=40895, avg=15534.30, stdev=4309.46 00:44:16.896 clat (usec): min=9133, max=50030, avg=13661.75, stdev=2326.70 00:44:16.896 lat (usec): min=9145, max=50043, avg=13677.29, stdev=2326.76 00:44:16.896 clat percentiles (usec): 00:44:16.896 | 1.00th=[ 9896], 5.00th=[10683], 10.00th=[11207], 20.00th=[11863], 00:44:16.896 | 30.00th=[12518], 40.00th=[13042], 50.00th=[13566], 60.00th=[14222], 00:44:16.896 | 70.00th=[14746], 80.00th=[15139], 90.00th=[15795], 95.00th=[16450], 00:44:16.896 | 99.00th=[17433], 99.50th=[18220], 99.90th=[47449], 99.95th=[50070], 00:44:16.896 | 99.99th=[50070] 00:44:16.896 bw ( KiB/s): min=26368, max=29696, per=32.60%, avg=28185.60, stdev=1245.11, samples=10 00:44:16.896 iops : min= 206, max= 232, avg=220.20, stdev= 9.73, samples=10 00:44:16.896 lat (msec) : 10=1.27%, 20=98.55%, 50=0.09%, 100=0.09% 00:44:16.896 cpu : usr=94.21%, sys=5.29%, ctx=10, majf=0, minf=100 00:44:16.896 IO depths : 1=0.2%, 2=99.8%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:44:16.896 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:16.896 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:16.896 issued rwts: total=1103,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:16.896 latency : target=0, window=0, percentile=100.00%, depth=3 00:44:16.896 filename0: (groupid=0, jobs=1): err= 0: pid=478275: Fri Dec 6 17:59:57 2024 00:44:16.896 read: IOPS=226, BW=28.3MiB/s (29.6MB/s)(143MiB/5047msec) 00:44:16.896 slat (nsec): min=5065, max=53367, avg=19872.43, stdev=4768.27 00:44:16.896 clat (usec): min=7831, max=49959, avg=13206.15, stdev=2397.41 00:44:16.896 lat (usec): min=7844, max=49979, avg=13226.02, stdev=2397.61 00:44:16.896 clat percentiles (usec): 00:44:16.896 | 1.00th=[ 9241], 5.00th=[10421], 10.00th=[10814], 20.00th=[11469], 00:44:16.896 | 30.00th=[11994], 40.00th=[12518], 50.00th=[13042], 60.00th=[13566], 00:44:16.896 | 70.00th=[14091], 80.00th=[14746], 90.00th=[15533], 95.00th=[16319], 00:44:16.896 | 99.00th=[17695], 99.50th=[19006], 99.90th=[47449], 99.95th=[50070], 00:44:16.896 | 99.99th=[50070] 00:44:16.896 bw ( KiB/s): min=27392, max=30720, per=33.73%, avg=29158.40, stdev=1075.67, samples=10 00:44:16.896 iops : min= 214, max= 240, avg=227.80, stdev= 8.40, samples=10 00:44:16.896 lat (msec) : 10=2.54%, 20=97.02%, 50=0.44% 00:44:16.896 cpu : usr=95.64%, sys=3.82%, ctx=7, majf=0, minf=93 00:44:16.896 IO depths : 1=0.3%, 2=99.7%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:44:16.896 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:16.896 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:16.896 issued rwts: total=1141,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:16.896 latency : target=0, window=0, percentile=100.00%, depth=3 00:44:16.896 00:44:16.896 Run status group 0 (all jobs): 00:44:16.896 READ: bw=84.4MiB/s (88.5MB/s), 27.3MiB/s-29.1MiB/s (28.7MB/s-30.5MB/s), io=426MiB (447MB), run=5005-5047msec 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@107 -- # destroy_subsystems 0 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@43 -- # local sub 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 0 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=0 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # NULL_DIF=2 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # bs=4k 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # numjobs=8 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # iodepth=16 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # runtime= 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # files=2 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@111 -- # create_subsystems 0 1 2 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@28 -- # local sub 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 0 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=0 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 2 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:16.896 bdev_null0 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:16.896 [2024-12-06 17:59:58.075753] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 1 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=1 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null1 64 512 --md-size 16 --dif-type 2 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:16.896 bdev_null1 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 --serial-number 53313233-1 --allow-any-host 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 bdev_null1 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 2 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=2 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null2 64 512 --md-size 16 --dif-type 2 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:16.896 bdev_null2 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 --serial-number 53313233-2 --allow-any-host 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 bdev_null2 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@112 -- # fio /dev/fd/62 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@112 -- # create_json_sub_conf 0 1 2 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@51 -- # gen_nvmf_target_json 0 1 2 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@560 -- # config=() 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@560 -- # local subsystem config 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:44:16.896 { 00:44:16.896 "params": { 00:44:16.896 "name": "Nvme$subsystem", 00:44:16.896 "trtype": "$TEST_TRANSPORT", 00:44:16.896 "traddr": "$NVMF_FIRST_TARGET_IP", 00:44:16.896 "adrfam": "ipv4", 00:44:16.896 "trsvcid": "$NVMF_PORT", 00:44:16.896 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:44:16.896 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:44:16.896 "hdgst": ${hdgst:-false}, 00:44:16.896 "ddgst": ${ddgst:-false} 00:44:16.896 }, 00:44:16.896 "method": "bdev_nvme_attach_controller" 00:44:16.896 } 00:44:16.896 EOF 00:44:16.896 )") 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1360 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1341 -- # local fio_dir=/usr/src/fio 00:44:16.896 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # gen_fio_conf 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1343 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@54 -- # local file 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1343 -- # local sanitizers 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1344 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@56 -- # cat 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # shift 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1347 -- # local asan_lib= 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # cat 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # grep libasan 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file = 1 )) 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@73 -- # cat 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:44:16.897 { 00:44:16.897 "params": { 00:44:16.897 "name": "Nvme$subsystem", 00:44:16.897 "trtype": "$TEST_TRANSPORT", 00:44:16.897 "traddr": "$NVMF_FIRST_TARGET_IP", 00:44:16.897 "adrfam": "ipv4", 00:44:16.897 "trsvcid": "$NVMF_PORT", 00:44:16.897 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:44:16.897 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:44:16.897 "hdgst": ${hdgst:-false}, 00:44:16.897 "ddgst": ${ddgst:-false} 00:44:16.897 }, 00:44:16.897 "method": "bdev_nvme_attach_controller" 00:44:16.897 } 00:44:16.897 EOF 00:44:16.897 )") 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file++ )) 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # cat 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@73 -- # cat 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file++ )) 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:44:16.897 { 00:44:16.897 "params": { 00:44:16.897 "name": "Nvme$subsystem", 00:44:16.897 "trtype": "$TEST_TRANSPORT", 00:44:16.897 "traddr": "$NVMF_FIRST_TARGET_IP", 00:44:16.897 "adrfam": "ipv4", 00:44:16.897 "trsvcid": "$NVMF_PORT", 00:44:16.897 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:44:16.897 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:44:16.897 "hdgst": ${hdgst:-false}, 00:44:16.897 "ddgst": ${ddgst:-false} 00:44:16.897 }, 00:44:16.897 "method": "bdev_nvme_attach_controller" 00:44:16.897 } 00:44:16.897 EOF 00:44:16.897 )") 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # cat 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@584 -- # jq . 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@585 -- # IFS=, 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:44:16.897 "params": { 00:44:16.897 "name": "Nvme0", 00:44:16.897 "trtype": "tcp", 00:44:16.897 "traddr": "10.0.0.2", 00:44:16.897 "adrfam": "ipv4", 00:44:16.897 "trsvcid": "4420", 00:44:16.897 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:44:16.897 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:44:16.897 "hdgst": false, 00:44:16.897 "ddgst": false 00:44:16.897 }, 00:44:16.897 "method": "bdev_nvme_attach_controller" 00:44:16.897 },{ 00:44:16.897 "params": { 00:44:16.897 "name": "Nvme1", 00:44:16.897 "trtype": "tcp", 00:44:16.897 "traddr": "10.0.0.2", 00:44:16.897 "adrfam": "ipv4", 00:44:16.897 "trsvcid": "4420", 00:44:16.897 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:44:16.897 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:44:16.897 "hdgst": false, 00:44:16.897 "ddgst": false 00:44:16.897 }, 00:44:16.897 "method": "bdev_nvme_attach_controller" 00:44:16.897 },{ 00:44:16.897 "params": { 00:44:16.897 "name": "Nvme2", 00:44:16.897 "trtype": "tcp", 00:44:16.897 "traddr": "10.0.0.2", 00:44:16.897 "adrfam": "ipv4", 00:44:16.897 "trsvcid": "4420", 00:44:16.897 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:44:16.897 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:44:16.897 "hdgst": false, 00:44:16.897 "ddgst": false 00:44:16.897 }, 00:44:16.897 "method": "bdev_nvme_attach_controller" 00:44:16.897 }' 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # asan_lib= 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # grep libclang_rt.asan 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # asan_lib= 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1356 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:44:16.897 17:59:58 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1356 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:44:16.897 filename0: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=16 00:44:16.897 ... 00:44:16.897 filename1: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=16 00:44:16.897 ... 00:44:16.897 filename2: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=16 00:44:16.897 ... 00:44:16.897 fio-3.35 00:44:16.897 Starting 24 threads 00:44:29.116 00:44:29.116 filename0: (groupid=0, jobs=1): err= 0: pid=479131: Fri Dec 6 18:00:09 2024 00:44:29.116 read: IOPS=308, BW=1234KiB/s (1263kB/s)(12.1MiB/10012msec) 00:44:29.116 slat (usec): min=9, max=106, avg=35.10, stdev=12.52 00:44:29.116 clat (msec): min=19, max=499, avg=51.56, stdev=73.52 00:44:29.116 lat (msec): min=19, max=499, avg=51.60, stdev=73.53 00:44:29.116 clat percentiles (msec): 00:44:29.116 | 1.00th=[ 33], 5.00th=[ 34], 10.00th=[ 34], 20.00th=[ 34], 00:44:29.116 | 30.00th=[ 34], 40.00th=[ 34], 50.00th=[ 34], 60.00th=[ 34], 00:44:29.116 | 70.00th=[ 34], 80.00th=[ 34], 90.00th=[ 35], 95.00th=[ 284], 00:44:29.116 | 99.00th=[ 393], 99.50th=[ 430], 99.90th=[ 430], 99.95th=[ 502], 00:44:29.116 | 99.99th=[ 502] 00:44:29.116 bw ( KiB/s): min= 128, max= 1920, per=4.09%, avg=1228.00, stdev=835.47, samples=20 00:44:29.116 iops : min= 32, max= 480, avg=307.00, stdev=208.87, samples=20 00:44:29.116 lat (msec) : 20=0.49%, 50=93.30%, 100=0.52%, 500=5.70% 00:44:29.116 cpu : usr=98.13%, sys=1.26%, ctx=46, majf=0, minf=56 00:44:29.116 IO depths : 1=6.1%, 2=12.3%, 4=25.0%, 8=50.2%, 16=6.4%, 32=0.0%, >=64=0.0% 00:44:29.116 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.116 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.116 issued rwts: total=3088,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:29.116 latency : target=0, window=0, percentile=100.00%, depth=16 00:44:29.116 filename0: (groupid=0, jobs=1): err= 0: pid=479132: Fri Dec 6 18:00:09 2024 00:44:29.116 read: IOPS=313, BW=1254KiB/s (1284kB/s)(12.3MiB/10024msec) 00:44:29.116 slat (usec): min=7, max=114, avg=31.78, stdev=13.95 00:44:29.116 clat (msec): min=23, max=492, avg=50.79, stdev=64.08 00:44:29.116 lat (msec): min=23, max=492, avg=50.82, stdev=64.08 00:44:29.116 clat percentiles (msec): 00:44:29.116 | 1.00th=[ 30], 5.00th=[ 34], 10.00th=[ 34], 20.00th=[ 34], 00:44:29.116 | 30.00th=[ 34], 40.00th=[ 34], 50.00th=[ 34], 60.00th=[ 34], 00:44:29.116 | 70.00th=[ 34], 80.00th=[ 34], 90.00th=[ 35], 95.00th=[ 255], 00:44:29.116 | 99.00th=[ 347], 99.50th=[ 384], 99.90th=[ 439], 99.95th=[ 493], 00:44:29.116 | 99.99th=[ 493] 00:44:29.116 bw ( KiB/s): min= 128, max= 1920, per=4.16%, avg=1250.40, stdev=809.71, samples=20 00:44:29.116 iops : min= 32, max= 480, avg=312.60, stdev=202.43, samples=20 00:44:29.116 lat (msec) : 50=92.17%, 100=0.73%, 250=1.62%, 500=5.47% 00:44:29.116 cpu : usr=98.54%, sys=0.99%, ctx=24, majf=0, minf=43 00:44:29.116 IO depths : 1=5.9%, 2=12.1%, 4=24.8%, 8=50.7%, 16=6.6%, 32=0.0%, >=64=0.0% 00:44:29.116 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.116 complete : 0=0.0%, 4=94.1%, 8=0.1%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.116 issued rwts: total=3142,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:29.116 latency : target=0, window=0, percentile=100.00%, depth=16 00:44:29.116 filename0: (groupid=0, jobs=1): err= 0: pid=479133: Fri Dec 6 18:00:09 2024 00:44:29.116 read: IOPS=322, BW=1289KiB/s (1320kB/s)(12.6MiB/10014msec) 00:44:29.116 slat (usec): min=5, max=123, avg=22.66, stdev=17.86 00:44:29.116 clat (usec): min=1961, max=477801, avg=49463.97, stdev=62297.98 00:44:29.116 lat (usec): min=1974, max=477895, avg=49486.62, stdev=62310.08 00:44:29.116 clat percentiles (msec): 00:44:29.116 | 1.00th=[ 12], 5.00th=[ 28], 10.00th=[ 34], 20.00th=[ 34], 00:44:29.116 | 30.00th=[ 34], 40.00th=[ 34], 50.00th=[ 34], 60.00th=[ 34], 00:44:29.116 | 70.00th=[ 34], 80.00th=[ 34], 90.00th=[ 35], 95.00th=[ 251], 00:44:29.116 | 99.00th=[ 338], 99.50th=[ 393], 99.90th=[ 409], 99.95th=[ 477], 00:44:29.116 | 99.99th=[ 477] 00:44:29.116 bw ( KiB/s): min= 128, max= 2096, per=4.28%, avg=1284.80, stdev=795.51, samples=20 00:44:29.116 iops : min= 32, max= 524, avg=321.20, stdev=198.88, samples=20 00:44:29.116 lat (msec) : 2=0.09%, 4=0.90%, 20=0.99%, 50=90.09%, 100=0.50% 00:44:29.116 lat (msec) : 250=2.35%, 500=5.08% 00:44:29.116 cpu : usr=97.80%, sys=1.41%, ctx=80, majf=0, minf=50 00:44:29.116 IO depths : 1=5.7%, 2=11.6%, 4=23.7%, 8=52.2%, 16=6.8%, 32=0.0%, >=64=0.0% 00:44:29.116 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.116 complete : 0=0.0%, 4=93.8%, 8=0.4%, 16=5.8%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.116 issued rwts: total=3228,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:29.116 latency : target=0, window=0, percentile=100.00%, depth=16 00:44:29.116 filename0: (groupid=0, jobs=1): err= 0: pid=479134: Fri Dec 6 18:00:09 2024 00:44:29.116 read: IOPS=317, BW=1271KiB/s (1301kB/s)(12.4MiB/10008msec) 00:44:29.116 slat (nsec): min=6274, max=62907, avg=29960.11, stdev=11334.60 00:44:29.116 clat (msec): min=12, max=394, avg=50.08, stdev=61.87 00:44:29.116 lat (msec): min=12, max=394, avg=50.11, stdev=61.86 00:44:29.116 clat percentiles (msec): 00:44:29.116 | 1.00th=[ 23], 5.00th=[ 33], 10.00th=[ 34], 20.00th=[ 34], 00:44:29.116 | 30.00th=[ 34], 40.00th=[ 34], 50.00th=[ 34], 60.00th=[ 34], 00:44:29.116 | 70.00th=[ 34], 80.00th=[ 34], 90.00th=[ 35], 95.00th=[ 251], 00:44:29.116 | 99.00th=[ 326], 99.50th=[ 342], 99.90th=[ 393], 99.95th=[ 393], 00:44:29.116 | 99.99th=[ 393] 00:44:29.116 bw ( KiB/s): min= 128, max= 1920, per=4.21%, avg=1265.60, stdev=811.69, samples=20 00:44:29.116 iops : min= 32, max= 480, avg=316.40, stdev=202.92, samples=20 00:44:29.116 lat (msec) : 20=0.50%, 50=92.08%, 250=2.20%, 500=5.22% 00:44:29.116 cpu : usr=98.16%, sys=1.27%, ctx=99, majf=0, minf=93 00:44:29.116 IO depths : 1=5.7%, 2=11.4%, 4=23.5%, 8=52.5%, 16=6.8%, 32=0.0%, >=64=0.0% 00:44:29.116 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.116 complete : 0=0.0%, 4=93.7%, 8=0.5%, 16=5.8%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.116 issued rwts: total=3180,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:29.116 latency : target=0, window=0, percentile=100.00%, depth=16 00:44:29.116 filename0: (groupid=0, jobs=1): err= 0: pid=479135: Fri Dec 6 18:00:09 2024 00:44:29.116 read: IOPS=309, BW=1239KiB/s (1268kB/s)(12.1MiB/10023msec) 00:44:29.116 slat (usec): min=5, max=161, avg=36.77, stdev=15.36 00:44:29.116 clat (msec): min=23, max=502, avg=51.34, stdev=70.17 00:44:29.116 lat (msec): min=23, max=502, avg=51.37, stdev=70.18 00:44:29.116 clat percentiles (msec): 00:44:29.116 | 1.00th=[ 33], 5.00th=[ 34], 10.00th=[ 34], 20.00th=[ 34], 00:44:29.116 | 30.00th=[ 34], 40.00th=[ 34], 50.00th=[ 34], 60.00th=[ 34], 00:44:29.116 | 70.00th=[ 34], 80.00th=[ 34], 90.00th=[ 35], 95.00th=[ 259], 00:44:29.116 | 99.00th=[ 384], 99.50th=[ 393], 99.90th=[ 489], 99.95th=[ 502], 00:44:29.116 | 99.99th=[ 502] 00:44:29.116 bw ( KiB/s): min= 128, max= 1920, per=4.11%, avg=1235.20, stdev=825.82, samples=20 00:44:29.116 iops : min= 32, max= 480, avg=308.80, stdev=206.45, samples=20 00:44:29.117 lat (msec) : 50=92.78%, 100=0.52%, 250=1.16%, 500=5.48%, 750=0.06% 00:44:29.117 cpu : usr=98.03%, sys=1.33%, ctx=70, majf=0, minf=46 00:44:29.117 IO depths : 1=6.1%, 2=12.3%, 4=25.0%, 8=50.2%, 16=6.4%, 32=0.0%, >=64=0.0% 00:44:29.117 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.117 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.117 issued rwts: total=3104,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:29.117 latency : target=0, window=0, percentile=100.00%, depth=16 00:44:29.117 filename0: (groupid=0, jobs=1): err= 0: pid=479136: Fri Dec 6 18:00:09 2024 00:44:29.117 read: IOPS=308, BW=1233KiB/s (1262kB/s)(12.1MiB/10019msec) 00:44:29.117 slat (usec): min=11, max=116, avg=36.26, stdev=13.31 00:44:29.117 clat (msec): min=24, max=524, avg=51.59, stdev=71.46 00:44:29.117 lat (msec): min=24, max=524, avg=51.63, stdev=71.47 00:44:29.117 clat percentiles (msec): 00:44:29.117 | 1.00th=[ 33], 5.00th=[ 34], 10.00th=[ 34], 20.00th=[ 34], 00:44:29.117 | 30.00th=[ 34], 40.00th=[ 34], 50.00th=[ 34], 60.00th=[ 34], 00:44:29.117 | 70.00th=[ 34], 80.00th=[ 34], 90.00th=[ 35], 95.00th=[ 271], 00:44:29.117 | 99.00th=[ 384], 99.50th=[ 393], 99.90th=[ 409], 99.95th=[ 527], 00:44:29.117 | 99.99th=[ 527] 00:44:29.117 bw ( KiB/s): min= 128, max= 1920, per=4.09%, avg=1228.80, stdev=836.05, samples=20 00:44:29.117 iops : min= 32, max= 480, avg=307.20, stdev=209.01, samples=20 00:44:29.117 lat (msec) : 50=93.78%, 250=0.58%, 500=5.57%, 750=0.06% 00:44:29.117 cpu : usr=96.56%, sys=2.03%, ctx=285, majf=0, minf=38 00:44:29.117 IO depths : 1=6.2%, 2=12.4%, 4=25.0%, 8=50.1%, 16=6.3%, 32=0.0%, >=64=0.0% 00:44:29.117 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.117 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.117 issued rwts: total=3088,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:29.117 latency : target=0, window=0, percentile=100.00%, depth=16 00:44:29.117 filename0: (groupid=0, jobs=1): err= 0: pid=479137: Fri Dec 6 18:00:09 2024 00:44:29.117 read: IOPS=317, BW=1272KiB/s (1302kB/s)(12.5MiB/10032msec) 00:44:29.117 slat (nsec): min=6277, max=87280, avg=32165.26, stdev=11502.76 00:44:29.117 clat (msec): min=21, max=326, avg=50.00, stdev=55.32 00:44:29.117 lat (msec): min=21, max=326, avg=50.03, stdev=55.32 00:44:29.117 clat percentiles (msec): 00:44:29.117 | 1.00th=[ 27], 5.00th=[ 34], 10.00th=[ 34], 20.00th=[ 34], 00:44:29.117 | 30.00th=[ 34], 40.00th=[ 34], 50.00th=[ 34], 60.00th=[ 34], 00:44:29.117 | 70.00th=[ 34], 80.00th=[ 34], 90.00th=[ 36], 95.00th=[ 230], 00:44:29.117 | 99.00th=[ 262], 99.50th=[ 279], 99.90th=[ 326], 99.95th=[ 326], 00:44:29.117 | 99.99th=[ 326] 00:44:29.117 bw ( KiB/s): min= 240, max= 1920, per=4.23%, avg=1269.60, stdev=784.06, samples=20 00:44:29.117 iops : min= 60, max= 480, avg=317.40, stdev=196.02, samples=20 00:44:29.117 lat (msec) : 50=90.97%, 100=0.50%, 250=4.70%, 500=3.82% 00:44:29.117 cpu : usr=98.45%, sys=1.11%, ctx=33, majf=0, minf=76 00:44:29.117 IO depths : 1=5.8%, 2=11.9%, 4=24.6%, 8=51.0%, 16=6.7%, 32=0.0%, >=64=0.0% 00:44:29.117 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.117 complete : 0=0.0%, 4=94.0%, 8=0.1%, 16=5.8%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.117 issued rwts: total=3190,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:29.117 latency : target=0, window=0, percentile=100.00%, depth=16 00:44:29.117 filename0: (groupid=0, jobs=1): err= 0: pid=479138: Fri Dec 6 18:00:09 2024 00:44:29.117 read: IOPS=314, BW=1260KiB/s (1290kB/s)(12.3MiB/10007msec) 00:44:29.117 slat (nsec): min=8456, max=66463, avg=31295.36, stdev=8948.97 00:44:29.117 clat (msec): min=19, max=351, avg=50.53, stdev=59.84 00:44:29.117 lat (msec): min=19, max=351, avg=50.56, stdev=59.84 00:44:29.117 clat percentiles (msec): 00:44:29.117 | 1.00th=[ 33], 5.00th=[ 34], 10.00th=[ 34], 20.00th=[ 34], 00:44:29.117 | 30.00th=[ 34], 40.00th=[ 34], 50.00th=[ 34], 60.00th=[ 34], 00:44:29.117 | 70.00th=[ 34], 80.00th=[ 34], 90.00th=[ 36], 95.00th=[ 249], 00:44:29.117 | 99.00th=[ 342], 99.50th=[ 342], 99.90th=[ 342], 99.95th=[ 351], 00:44:29.117 | 99.99th=[ 351] 00:44:29.117 bw ( KiB/s): min= 128, max= 1920, per=4.18%, avg=1254.40, stdev=806.87, samples=20 00:44:29.117 iops : min= 32, max= 480, avg=313.60, stdev=201.72, samples=20 00:44:29.117 lat (msec) : 20=0.51%, 50=91.37%, 100=0.51%, 250=3.11%, 500=4.51% 00:44:29.117 cpu : usr=98.38%, sys=1.19%, ctx=23, majf=0, minf=57 00:44:29.117 IO depths : 1=6.1%, 2=12.3%, 4=25.0%, 8=50.2%, 16=6.4%, 32=0.0%, >=64=0.0% 00:44:29.117 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.117 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.117 issued rwts: total=3152,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:29.117 latency : target=0, window=0, percentile=100.00%, depth=16 00:44:29.117 filename1: (groupid=0, jobs=1): err= 0: pid=479139: Fri Dec 6 18:00:09 2024 00:44:29.117 read: IOPS=326, BW=1305KiB/s (1336kB/s)(12.8MiB/10023msec) 00:44:29.117 slat (nsec): min=7973, max=85380, avg=26715.58, stdev=15501.79 00:44:29.117 clat (msec): min=12, max=270, avg=48.82, stdev=52.36 00:44:29.117 lat (msec): min=12, max=270, avg=48.85, stdev=52.36 00:44:29.117 clat percentiles (msec): 00:44:29.117 | 1.00th=[ 23], 5.00th=[ 27], 10.00th=[ 34], 20.00th=[ 34], 00:44:29.117 | 30.00th=[ 34], 40.00th=[ 34], 50.00th=[ 34], 60.00th=[ 34], 00:44:29.117 | 70.00th=[ 34], 80.00th=[ 34], 90.00th=[ 36], 95.00th=[ 205], 00:44:29.117 | 99.00th=[ 259], 99.50th=[ 268], 99.90th=[ 271], 99.95th=[ 271], 00:44:29.117 | 99.99th=[ 271] 00:44:29.117 bw ( KiB/s): min= 208, max= 2048, per=4.33%, avg=1301.60, stdev=774.09, samples=20 00:44:29.117 iops : min= 52, max= 512, avg=325.40, stdev=193.52, samples=20 00:44:29.117 lat (msec) : 20=0.98%, 50=89.54%, 100=0.92%, 250=5.81%, 500=2.75% 00:44:29.117 cpu : usr=97.34%, sys=1.72%, ctx=135, majf=0, minf=61 00:44:29.117 IO depths : 1=5.4%, 2=10.8%, 4=22.5%, 8=54.3%, 16=7.2%, 32=0.0%, >=64=0.0% 00:44:29.117 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.117 complete : 0=0.0%, 4=93.4%, 8=0.8%, 16=5.8%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.117 issued rwts: total=3270,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:29.117 latency : target=0, window=0, percentile=100.00%, depth=16 00:44:29.117 filename1: (groupid=0, jobs=1): err= 0: pid=479140: Fri Dec 6 18:00:09 2024 00:44:29.117 read: IOPS=313, BW=1255KiB/s (1285kB/s)(12.3MiB/10012msec) 00:44:29.117 slat (usec): min=3, max=136, avg=30.27, stdev=10.28 00:44:29.117 clat (msec): min=23, max=415, avg=50.73, stdev=61.68 00:44:29.117 lat (msec): min=23, max=415, avg=50.76, stdev=61.68 00:44:29.117 clat percentiles (msec): 00:44:29.117 | 1.00th=[ 25], 5.00th=[ 34], 10.00th=[ 34], 20.00th=[ 34], 00:44:29.117 | 30.00th=[ 34], 40.00th=[ 34], 50.00th=[ 34], 60.00th=[ 34], 00:44:29.117 | 70.00th=[ 34], 80.00th=[ 34], 90.00th=[ 35], 95.00th=[ 251], 00:44:29.117 | 99.00th=[ 317], 99.50th=[ 317], 99.90th=[ 388], 99.95th=[ 414], 00:44:29.117 | 99.99th=[ 414] 00:44:29.117 bw ( KiB/s): min= 128, max= 1920, per=4.16%, avg=1250.40, stdev=811.24, samples=20 00:44:29.117 iops : min= 32, max= 480, avg=312.60, stdev=202.81, samples=20 00:44:29.117 lat (msec) : 50=92.36%, 250=1.91%, 500=5.73% 00:44:29.117 cpu : usr=97.86%, sys=1.41%, ctx=88, majf=0, minf=51 00:44:29.117 IO depths : 1=5.8%, 2=11.8%, 4=24.3%, 8=51.4%, 16=6.7%, 32=0.0%, >=64=0.0% 00:44:29.117 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.117 complete : 0=0.0%, 4=93.9%, 8=0.2%, 16=5.8%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.117 issued rwts: total=3142,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:29.117 latency : target=0, window=0, percentile=100.00%, depth=16 00:44:29.117 filename1: (groupid=0, jobs=1): err= 0: pid=479141: Fri Dec 6 18:00:09 2024 00:44:29.117 read: IOPS=306, BW=1227KiB/s (1256kB/s)(12.0MiB/10018msec) 00:44:29.117 slat (nsec): min=4252, max=73479, avg=32513.47, stdev=9224.36 00:44:29.117 clat (msec): min=24, max=502, avg=51.88, stdev=76.08 00:44:29.117 lat (msec): min=24, max=502, avg=51.91, stdev=76.07 00:44:29.117 clat percentiles (msec): 00:44:29.117 | 1.00th=[ 33], 5.00th=[ 34], 10.00th=[ 34], 20.00th=[ 34], 00:44:29.117 | 30.00th=[ 34], 40.00th=[ 34], 50.00th=[ 34], 60.00th=[ 34], 00:44:29.117 | 70.00th=[ 34], 80.00th=[ 34], 90.00th=[ 35], 95.00th=[ 275], 00:44:29.117 | 99.00th=[ 397], 99.50th=[ 502], 99.90th=[ 502], 99.95th=[ 502], 00:44:29.117 | 99.99th=[ 502] 00:44:29.117 bw ( KiB/s): min= 128, max= 1920, per=4.28%, avg=1286.74, stdev=816.79, samples=19 00:44:29.117 iops : min= 32, max= 480, avg=321.68, stdev=204.20, samples=19 00:44:29.117 lat (msec) : 50=94.27%, 250=0.07%, 500=5.14%, 750=0.52% 00:44:29.117 cpu : usr=98.17%, sys=1.21%, ctx=118, majf=0, minf=38 00:44:29.117 IO depths : 1=6.2%, 2=12.4%, 4=25.0%, 8=50.1%, 16=6.3%, 32=0.0%, >=64=0.0% 00:44:29.117 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.117 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.117 issued rwts: total=3072,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:29.117 latency : target=0, window=0, percentile=100.00%, depth=16 00:44:29.117 filename1: (groupid=0, jobs=1): err= 0: pid=479142: Fri Dec 6 18:00:09 2024 00:44:29.117 read: IOPS=314, BW=1259KiB/s (1289kB/s)(12.3MiB/10017msec) 00:44:29.117 slat (usec): min=7, max=111, avg=33.71, stdev=14.23 00:44:29.117 clat (msec): min=23, max=384, avg=50.51, stdev=59.38 00:44:29.117 lat (msec): min=23, max=384, avg=50.55, stdev=59.38 00:44:29.117 clat percentiles (msec): 00:44:29.117 | 1.00th=[ 33], 5.00th=[ 34], 10.00th=[ 34], 20.00th=[ 34], 00:44:29.117 | 30.00th=[ 34], 40.00th=[ 34], 50.00th=[ 34], 60.00th=[ 34], 00:44:29.117 | 70.00th=[ 34], 80.00th=[ 34], 90.00th=[ 36], 95.00th=[ 234], 00:44:29.117 | 99.00th=[ 284], 99.50th=[ 384], 99.90th=[ 384], 99.95th=[ 384], 00:44:29.117 | 99.99th=[ 384] 00:44:29.117 bw ( KiB/s): min= 128, max= 1920, per=4.18%, avg=1254.40, stdev=802.58, samples=20 00:44:29.117 iops : min= 32, max= 480, avg=313.60, stdev=200.64, samples=20 00:44:29.117 lat (msec) : 50=91.37%, 100=0.98%, 250=3.20%, 500=4.44% 00:44:29.117 cpu : usr=97.81%, sys=1.36%, ctx=93, majf=0, minf=51 00:44:29.117 IO depths : 1=6.1%, 2=12.4%, 4=25.0%, 8=50.1%, 16=6.4%, 32=0.0%, >=64=0.0% 00:44:29.117 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.117 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.117 issued rwts: total=3152,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:29.117 latency : target=0, window=0, percentile=100.00%, depth=16 00:44:29.117 filename1: (groupid=0, jobs=1): err= 0: pid=479143: Fri Dec 6 18:00:09 2024 00:44:29.117 read: IOPS=311, BW=1247KiB/s (1277kB/s)(12.2MiB/10010msec) 00:44:29.117 slat (usec): min=8, max=100, avg=32.84, stdev=10.26 00:44:29.117 clat (msec): min=24, max=429, avg=51.04, stdev=64.05 00:44:29.118 lat (msec): min=24, max=429, avg=51.07, stdev=64.05 00:44:29.118 clat percentiles (msec): 00:44:29.118 | 1.00th=[ 33], 5.00th=[ 34], 10.00th=[ 34], 20.00th=[ 34], 00:44:29.118 | 30.00th=[ 34], 40.00th=[ 34], 50.00th=[ 34], 60.00th=[ 34], 00:44:29.118 | 70.00th=[ 34], 80.00th=[ 34], 90.00th=[ 35], 95.00th=[ 251], 00:44:29.118 | 99.00th=[ 342], 99.50th=[ 351], 99.90th=[ 351], 99.95th=[ 430], 00:44:29.118 | 99.99th=[ 430] 00:44:29.118 bw ( KiB/s): min= 128, max= 1923, per=4.14%, avg=1242.05, stdev=819.46, samples=20 00:44:29.118 iops : min= 32, max= 480, avg=310.40, stdev=204.77, samples=20 00:44:29.118 lat (msec) : 50=92.82%, 250=1.54%, 500=5.64% 00:44:29.118 cpu : usr=97.37%, sys=1.60%, ctx=109, majf=0, minf=46 00:44:29.118 IO depths : 1=5.8%, 2=12.1%, 4=25.0%, 8=50.4%, 16=6.7%, 32=0.0%, >=64=0.0% 00:44:29.118 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.118 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.118 issued rwts: total=3120,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:29.118 latency : target=0, window=0, percentile=100.00%, depth=16 00:44:29.118 filename1: (groupid=0, jobs=1): err= 0: pid=479144: Fri Dec 6 18:00:09 2024 00:44:29.118 read: IOPS=315, BW=1261KiB/s (1291kB/s)(12.3MiB/10007msec) 00:44:29.118 slat (nsec): min=5468, max=97433, avg=32532.46, stdev=11824.61 00:44:29.118 clat (msec): min=21, max=390, avg=50.47, stdev=59.52 00:44:29.118 lat (msec): min=21, max=390, avg=50.50, stdev=59.52 00:44:29.118 clat percentiles (msec): 00:44:29.118 | 1.00th=[ 27], 5.00th=[ 34], 10.00th=[ 34], 20.00th=[ 34], 00:44:29.118 | 30.00th=[ 34], 40.00th=[ 34], 50.00th=[ 34], 60.00th=[ 34], 00:44:29.118 | 70.00th=[ 34], 80.00th=[ 34], 90.00th=[ 35], 95.00th=[ 243], 00:44:29.118 | 99.00th=[ 309], 99.50th=[ 317], 99.90th=[ 393], 99.95th=[ 393], 00:44:29.118 | 99.99th=[ 393] 00:44:29.118 bw ( KiB/s): min= 144, max= 1920, per=4.08%, avg=1226.95, stdev=817.10, samples=19 00:44:29.118 iops : min= 36, max= 480, avg=306.74, stdev=204.27, samples=19 00:44:29.118 lat (msec) : 50=92.01%, 250=3.36%, 500=4.63% 00:44:29.118 cpu : usr=98.14%, sys=1.34%, ctx=37, majf=0, minf=57 00:44:29.118 IO depths : 1=5.7%, 2=11.5%, 4=23.6%, 8=52.3%, 16=6.8%, 32=0.0%, >=64=0.0% 00:44:29.118 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.118 complete : 0=0.0%, 4=93.7%, 8=0.5%, 16=5.8%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.118 issued rwts: total=3154,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:29.118 latency : target=0, window=0, percentile=100.00%, depth=16 00:44:29.118 filename1: (groupid=0, jobs=1): err= 0: pid=479145: Fri Dec 6 18:00:09 2024 00:44:29.118 read: IOPS=310, BW=1241KiB/s (1270kB/s)(12.1MiB/10008msec) 00:44:29.118 slat (usec): min=9, max=116, avg=33.28, stdev=14.54 00:44:29.118 clat (msec): min=32, max=385, avg=51.30, stdev=67.54 00:44:29.118 lat (msec): min=32, max=385, avg=51.34, stdev=67.55 00:44:29.118 clat percentiles (msec): 00:44:29.118 | 1.00th=[ 34], 5.00th=[ 34], 10.00th=[ 34], 20.00th=[ 34], 00:44:29.118 | 30.00th=[ 34], 40.00th=[ 34], 50.00th=[ 34], 60.00th=[ 34], 00:44:29.118 | 70.00th=[ 34], 80.00th=[ 34], 90.00th=[ 35], 95.00th=[ 257], 00:44:29.118 | 99.00th=[ 376], 99.50th=[ 384], 99.90th=[ 384], 99.95th=[ 384], 00:44:29.118 | 99.99th=[ 384] 00:44:29.118 bw ( KiB/s): min= 128, max= 1920, per=4.11%, avg=1235.20, stdev=825.65, samples=20 00:44:29.118 iops : min= 32, max= 480, avg=308.80, stdev=206.41, samples=20 00:44:29.118 lat (msec) : 50=92.78%, 100=0.52%, 250=0.52%, 500=6.19% 00:44:29.118 cpu : usr=98.16%, sys=1.30%, ctx=25, majf=0, minf=67 00:44:29.118 IO depths : 1=6.2%, 2=12.5%, 4=25.0%, 8=50.0%, 16=6.2%, 32=0.0%, >=64=0.0% 00:44:29.118 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.118 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.118 issued rwts: total=3104,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:29.118 latency : target=0, window=0, percentile=100.00%, depth=16 00:44:29.118 filename1: (groupid=0, jobs=1): err= 0: pid=479146: Fri Dec 6 18:00:09 2024 00:44:29.118 read: IOPS=311, BW=1246KiB/s (1275kB/s)(12.2MiB/10007msec) 00:44:29.118 slat (usec): min=8, max=140, avg=32.16, stdev=12.97 00:44:29.118 clat (msec): min=19, max=498, avg=51.07, stdev=66.82 00:44:29.118 lat (msec): min=19, max=498, avg=51.10, stdev=66.82 00:44:29.118 clat percentiles (msec): 00:44:29.118 | 1.00th=[ 33], 5.00th=[ 34], 10.00th=[ 34], 20.00th=[ 34], 00:44:29.118 | 30.00th=[ 34], 40.00th=[ 34], 50.00th=[ 34], 60.00th=[ 34], 00:44:29.118 | 70.00th=[ 34], 80.00th=[ 34], 90.00th=[ 35], 95.00th=[ 257], 00:44:29.118 | 99.00th=[ 347], 99.50th=[ 384], 99.90th=[ 439], 99.95th=[ 498], 00:44:29.118 | 99.99th=[ 498] 00:44:29.118 bw ( KiB/s): min= 128, max= 1920, per=4.13%, avg=1240.00, stdev=824.38, samples=20 00:44:29.118 iops : min= 32, max= 480, avg=310.00, stdev=206.09, samples=20 00:44:29.118 lat (msec) : 20=0.48%, 50=92.65%, 100=0.19%, 250=1.16%, 500=5.52% 00:44:29.118 cpu : usr=97.92%, sys=1.33%, ctx=32, majf=0, minf=54 00:44:29.118 IO depths : 1=6.0%, 2=12.1%, 4=24.7%, 8=50.6%, 16=6.5%, 32=0.0%, >=64=0.0% 00:44:29.118 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.118 complete : 0=0.0%, 4=94.0%, 8=0.1%, 16=5.8%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.118 issued rwts: total=3116,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:29.118 latency : target=0, window=0, percentile=100.00%, depth=16 00:44:29.118 filename2: (groupid=0, jobs=1): err= 0: pid=479147: Fri Dec 6 18:00:09 2024 00:44:29.118 read: IOPS=309, BW=1240KiB/s (1270kB/s)(12.1MiB/10013msec) 00:44:29.118 slat (usec): min=4, max=112, avg=30.00, stdev=15.10 00:44:29.118 clat (msec): min=23, max=504, avg=51.38, stdev=67.75 00:44:29.118 lat (msec): min=23, max=504, avg=51.41, stdev=67.76 00:44:29.118 clat percentiles (msec): 00:44:29.118 | 1.00th=[ 33], 5.00th=[ 34], 10.00th=[ 34], 20.00th=[ 34], 00:44:29.118 | 30.00th=[ 34], 40.00th=[ 34], 50.00th=[ 34], 60.00th=[ 34], 00:44:29.118 | 70.00th=[ 34], 80.00th=[ 34], 90.00th=[ 35], 95.00th=[ 253], 00:44:29.118 | 99.00th=[ 380], 99.50th=[ 384], 99.90th=[ 481], 99.95th=[ 506], 00:44:29.118 | 99.99th=[ 506] 00:44:29.118 bw ( KiB/s): min= 128, max= 1920, per=4.11%, avg=1235.20, stdev=827.63, samples=20 00:44:29.118 iops : min= 32, max= 480, avg=308.80, stdev=206.91, samples=20 00:44:29.118 lat (msec) : 50=92.85%, 100=0.45%, 250=1.68%, 500=4.96%, 750=0.06% 00:44:29.118 cpu : usr=97.34%, sys=1.70%, ctx=157, majf=0, minf=48 00:44:29.118 IO depths : 1=5.9%, 2=12.1%, 4=25.0%, 8=50.4%, 16=6.6%, 32=0.0%, >=64=0.0% 00:44:29.118 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.118 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.118 issued rwts: total=3104,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:29.118 latency : target=0, window=0, percentile=100.00%, depth=16 00:44:29.118 filename2: (groupid=0, jobs=1): err= 0: pid=479148: Fri Dec 6 18:00:09 2024 00:44:29.118 read: IOPS=317, BW=1270KiB/s (1300kB/s)(12.4MiB/10032msec) 00:44:29.118 slat (usec): min=8, max=111, avg=14.33, stdev=11.75 00:44:29.118 clat (msec): min=12, max=343, avg=50.28, stdev=59.67 00:44:29.118 lat (msec): min=12, max=343, avg=50.30, stdev=59.68 00:44:29.118 clat percentiles (msec): 00:44:29.118 | 1.00th=[ 19], 5.00th=[ 34], 10.00th=[ 34], 20.00th=[ 34], 00:44:29.118 | 30.00th=[ 34], 40.00th=[ 34], 50.00th=[ 34], 60.00th=[ 34], 00:44:29.118 | 70.00th=[ 34], 80.00th=[ 35], 90.00th=[ 36], 95.00th=[ 251], 00:44:29.118 | 99.00th=[ 321], 99.50th=[ 338], 99.90th=[ 338], 99.95th=[ 342], 00:44:29.118 | 99.99th=[ 342] 00:44:29.118 bw ( KiB/s): min= 128, max= 1920, per=4.22%, avg=1267.20, stdev=800.87, samples=20 00:44:29.118 iops : min= 32, max= 480, avg=316.80, stdev=200.22, samples=20 00:44:29.118 lat (msec) : 20=1.01%, 50=90.52%, 100=1.38%, 250=2.14%, 500=4.96% 00:44:29.118 cpu : usr=98.18%, sys=1.34%, ctx=41, majf=0, minf=64 00:44:29.118 IO depths : 1=5.9%, 2=12.2%, 4=25.0%, 8=50.3%, 16=6.6%, 32=0.0%, >=64=0.0% 00:44:29.118 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.118 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.118 issued rwts: total=3184,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:29.118 latency : target=0, window=0, percentile=100.00%, depth=16 00:44:29.118 filename2: (groupid=0, jobs=1): err= 0: pid=479149: Fri Dec 6 18:00:09 2024 00:44:29.118 read: IOPS=318, BW=1273KiB/s (1303kB/s)(12.4MiB/10008msec) 00:44:29.118 slat (nsec): min=6679, max=72252, avg=24780.99, stdev=11218.35 00:44:29.118 clat (msec): min=27, max=277, avg=50.09, stdev=52.89 00:44:29.118 lat (msec): min=27, max=277, avg=50.11, stdev=52.88 00:44:29.118 clat percentiles (msec): 00:44:29.118 | 1.00th=[ 34], 5.00th=[ 34], 10.00th=[ 34], 20.00th=[ 34], 00:44:29.118 | 30.00th=[ 34], 40.00th=[ 34], 50.00th=[ 34], 60.00th=[ 34], 00:44:29.118 | 70.00th=[ 34], 80.00th=[ 35], 90.00th=[ 39], 95.00th=[ 209], 00:44:29.118 | 99.00th=[ 262], 99.50th=[ 271], 99.90th=[ 279], 99.95th=[ 279], 00:44:29.118 | 99.99th=[ 279] 00:44:29.118 bw ( KiB/s): min= 240, max= 1920, per=4.22%, avg=1267.20, stdev=783.62, samples=20 00:44:29.118 iops : min= 60, max= 480, avg=316.80, stdev=195.91, samples=20 00:44:29.118 lat (msec) : 50=90.45%, 100=0.50%, 250=6.53%, 500=2.51% 00:44:29.118 cpu : usr=98.49%, sys=1.04%, ctx=26, majf=0, minf=62 00:44:29.118 IO depths : 1=5.7%, 2=11.9%, 4=25.0%, 8=50.6%, 16=6.8%, 32=0.0%, >=64=0.0% 00:44:29.118 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.118 complete : 0=0.0%, 4=94.2%, 8=0.0%, 16=5.8%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.118 issued rwts: total=3184,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:29.118 latency : target=0, window=0, percentile=100.00%, depth=16 00:44:29.118 filename2: (groupid=0, jobs=1): err= 0: pid=479150: Fri Dec 6 18:00:09 2024 00:44:29.118 read: IOPS=314, BW=1258KiB/s (1289kB/s)(12.3MiB/10019msec) 00:44:29.118 slat (nsec): min=8030, max=77401, avg=32414.37, stdev=9483.28 00:44:29.118 clat (msec): min=24, max=344, avg=50.56, stdev=58.07 00:44:29.118 lat (msec): min=24, max=344, avg=50.59, stdev=58.06 00:44:29.118 clat percentiles (msec): 00:44:29.118 | 1.00th=[ 33], 5.00th=[ 34], 10.00th=[ 34], 20.00th=[ 34], 00:44:29.118 | 30.00th=[ 34], 40.00th=[ 34], 50.00th=[ 34], 60.00th=[ 34], 00:44:29.118 | 70.00th=[ 34], 80.00th=[ 34], 90.00th=[ 36], 95.00th=[ 249], 00:44:29.118 | 99.00th=[ 284], 99.50th=[ 288], 99.90th=[ 288], 99.95th=[ 347], 00:44:29.118 | 99.99th=[ 347] 00:44:29.118 bw ( KiB/s): min= 240, max= 1920, per=4.18%, avg=1254.40, stdev=802.04, samples=20 00:44:29.118 iops : min= 60, max= 480, avg=313.60, stdev=200.51, samples=20 00:44:29.118 lat (msec) : 50=91.88%, 250=3.62%, 500=4.51% 00:44:29.118 cpu : usr=97.29%, sys=1.73%, ctx=164, majf=0, minf=57 00:44:29.119 IO depths : 1=6.1%, 2=12.3%, 4=25.0%, 8=50.2%, 16=6.4%, 32=0.0%, >=64=0.0% 00:44:29.119 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.119 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.119 issued rwts: total=3152,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:29.119 latency : target=0, window=0, percentile=100.00%, depth=16 00:44:29.119 filename2: (groupid=0, jobs=1): err= 0: pid=479151: Fri Dec 6 18:00:09 2024 00:44:29.119 read: IOPS=308, BW=1234KiB/s (1264kB/s)(12.1MiB/10008msec) 00:44:29.119 slat (usec): min=8, max=138, avg=32.09, stdev=11.99 00:44:29.119 clat (msec): min=12, max=503, avg=51.55, stdev=73.73 00:44:29.119 lat (msec): min=12, max=503, avg=51.58, stdev=73.73 00:44:29.119 clat percentiles (msec): 00:44:29.119 | 1.00th=[ 30], 5.00th=[ 34], 10.00th=[ 34], 20.00th=[ 34], 00:44:29.119 | 30.00th=[ 34], 40.00th=[ 34], 50.00th=[ 34], 60.00th=[ 34], 00:44:29.119 | 70.00th=[ 34], 80.00th=[ 34], 90.00th=[ 35], 95.00th=[ 271], 00:44:29.119 | 99.00th=[ 393], 99.50th=[ 422], 99.90th=[ 493], 99.95th=[ 506], 00:44:29.119 | 99.99th=[ 506] 00:44:29.119 bw ( KiB/s): min= 128, max= 1920, per=4.09%, avg=1228.80, stdev=846.68, samples=20 00:44:29.119 iops : min= 32, max= 480, avg=307.20, stdev=211.67, samples=20 00:44:29.119 lat (msec) : 20=0.52%, 50=93.26%, 250=0.58%, 500=5.57%, 750=0.06% 00:44:29.119 cpu : usr=98.61%, sys=0.95%, ctx=52, majf=0, minf=51 00:44:29.119 IO depths : 1=6.1%, 2=12.3%, 4=25.0%, 8=50.2%, 16=6.4%, 32=0.0%, >=64=0.0% 00:44:29.119 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.119 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.119 issued rwts: total=3088,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:29.119 latency : target=0, window=0, percentile=100.00%, depth=16 00:44:29.119 filename2: (groupid=0, jobs=1): err= 0: pid=479152: Fri Dec 6 18:00:09 2024 00:44:29.119 read: IOPS=307, BW=1228KiB/s (1258kB/s)(12.0MiB/10005msec) 00:44:29.119 slat (nsec): min=6247, max=58086, avg=31141.47, stdev=7511.12 00:44:29.119 clat (msec): min=24, max=489, avg=51.82, stdev=75.52 00:44:29.119 lat (msec): min=24, max=489, avg=51.85, stdev=75.52 00:44:29.119 clat percentiles (msec): 00:44:29.119 | 1.00th=[ 33], 5.00th=[ 34], 10.00th=[ 34], 20.00th=[ 34], 00:44:29.119 | 30.00th=[ 34], 40.00th=[ 34], 50.00th=[ 34], 60.00th=[ 34], 00:44:29.119 | 70.00th=[ 34], 80.00th=[ 34], 90.00th=[ 35], 95.00th=[ 275], 00:44:29.119 | 99.00th=[ 397], 99.50th=[ 489], 99.90th=[ 489], 99.95th=[ 489], 00:44:29.119 | 99.99th=[ 489] 00:44:29.119 bw ( KiB/s): min= 128, max= 1920, per=3.97%, avg=1192.42, stdev=856.58, samples=19 00:44:29.119 iops : min= 32, max= 480, avg=298.11, stdev=214.15, samples=19 00:44:29.119 lat (msec) : 50=94.27%, 250=0.07%, 500=5.66% 00:44:29.119 cpu : usr=97.99%, sys=1.47%, ctx=33, majf=0, minf=54 00:44:29.119 IO depths : 1=6.2%, 2=12.4%, 4=25.0%, 8=50.1%, 16=6.3%, 32=0.0%, >=64=0.0% 00:44:29.119 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.119 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.119 issued rwts: total=3072,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:29.119 latency : target=0, window=0, percentile=100.00%, depth=16 00:44:29.119 filename2: (groupid=0, jobs=1): err= 0: pid=479153: Fri Dec 6 18:00:09 2024 00:44:29.119 read: IOPS=314, BW=1260KiB/s (1290kB/s)(12.3MiB/10007msec) 00:44:29.119 slat (nsec): min=8367, max=97886, avg=33113.71, stdev=9690.81 00:44:29.119 clat (msec): min=19, max=422, avg=50.51, stdev=59.57 00:44:29.119 lat (msec): min=19, max=422, avg=50.54, stdev=59.57 00:44:29.119 clat percentiles (msec): 00:44:29.119 | 1.00th=[ 33], 5.00th=[ 34], 10.00th=[ 34], 20.00th=[ 34], 00:44:29.119 | 30.00th=[ 34], 40.00th=[ 34], 50.00th=[ 34], 60.00th=[ 34], 00:44:29.119 | 70.00th=[ 34], 80.00th=[ 34], 90.00th=[ 36], 95.00th=[ 249], 00:44:29.119 | 99.00th=[ 326], 99.50th=[ 342], 99.90th=[ 342], 99.95th=[ 422], 00:44:29.119 | 99.99th=[ 422] 00:44:29.119 bw ( KiB/s): min= 128, max= 1920, per=4.06%, avg=1219.37, stdev=813.06, samples=19 00:44:29.119 iops : min= 32, max= 480, avg=304.84, stdev=203.26, samples=19 00:44:29.119 lat (msec) : 20=0.51%, 50=91.43%, 100=0.44%, 250=3.52%, 500=4.09% 00:44:29.119 cpu : usr=98.58%, sys=0.99%, ctx=26, majf=0, minf=52 00:44:29.119 IO depths : 1=5.7%, 2=12.0%, 4=25.0%, 8=50.5%, 16=6.8%, 32=0.0%, >=64=0.0% 00:44:29.119 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.119 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.119 issued rwts: total=3152,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:29.119 latency : target=0, window=0, percentile=100.00%, depth=16 00:44:29.119 filename2: (groupid=0, jobs=1): err= 0: pid=479154: Fri Dec 6 18:00:09 2024 00:44:29.119 read: IOPS=308, BW=1234KiB/s (1264kB/s)(12.1MiB/10007msec) 00:44:29.119 slat (usec): min=8, max=107, avg=34.59, stdev=12.02 00:44:29.119 clat (msec): min=19, max=527, avg=51.54, stdev=72.16 00:44:29.119 lat (msec): min=19, max=527, avg=51.57, stdev=72.16 00:44:29.119 clat percentiles (msec): 00:44:29.119 | 1.00th=[ 33], 5.00th=[ 34], 10.00th=[ 34], 20.00th=[ 34], 00:44:29.119 | 30.00th=[ 34], 40.00th=[ 34], 50.00th=[ 34], 60.00th=[ 34], 00:44:29.119 | 70.00th=[ 34], 80.00th=[ 34], 90.00th=[ 35], 95.00th=[ 259], 00:44:29.119 | 99.00th=[ 405], 99.50th=[ 422], 99.90th=[ 426], 99.95th=[ 527], 00:44:29.119 | 99.99th=[ 527] 00:44:29.119 bw ( KiB/s): min= 128, max= 1920, per=3.97%, avg=1192.42, stdev=842.66, samples=19 00:44:29.119 iops : min= 32, max= 480, avg=298.11, stdev=210.66, samples=19 00:44:29.119 lat (msec) : 20=0.49%, 50=93.30%, 250=1.10%, 500=5.05%, 750=0.06% 00:44:29.119 cpu : usr=97.75%, sys=1.52%, ctx=82, majf=0, minf=47 00:44:29.119 IO depths : 1=6.1%, 2=12.3%, 4=25.0%, 8=50.2%, 16=6.4%, 32=0.0%, >=64=0.0% 00:44:29.119 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.119 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:29.119 issued rwts: total=3088,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:29.119 latency : target=0, window=0, percentile=100.00%, depth=16 00:44:29.119 00:44:29.119 Run status group 0 (all jobs): 00:44:29.119 READ: bw=29.3MiB/s (30.8MB/s), 1227KiB/s-1305KiB/s (1256kB/s-1336kB/s), io=294MiB (309MB), run=10005-10032msec 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@113 -- # destroy_subsystems 0 1 2 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@43 -- # local sub 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 0 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=0 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 1 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=1 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null1 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 2 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=2 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode2 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null2 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # NULL_DIF=1 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # bs=8k,16k,128k 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # numjobs=2 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # iodepth=8 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # runtime=5 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # files=1 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@117 -- # create_subsystems 0 1 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@28 -- # local sub 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 0 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=0 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 1 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:29.119 bdev_null0 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:29.119 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:29.120 [2024-12-06 18:00:09.606475] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 1 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=1 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null1 64 512 --md-size 16 --dif-type 1 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:29.120 bdev_null1 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 --serial-number 53313233-1 --allow-any-host 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 bdev_null1 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@118 -- # fio /dev/fd/62 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@118 -- # create_json_sub_conf 0 1 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@51 -- # gen_nvmf_target_json 0 1 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@560 -- # config=() 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@560 -- # local subsystem config 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:44:29.120 { 00:44:29.120 "params": { 00:44:29.120 "name": "Nvme$subsystem", 00:44:29.120 "trtype": "$TEST_TRANSPORT", 00:44:29.120 "traddr": "$NVMF_FIRST_TARGET_IP", 00:44:29.120 "adrfam": "ipv4", 00:44:29.120 "trsvcid": "$NVMF_PORT", 00:44:29.120 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:44:29.120 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:44:29.120 "hdgst": ${hdgst:-false}, 00:44:29.120 "ddgst": ${ddgst:-false} 00:44:29.120 }, 00:44:29.120 "method": "bdev_nvme_attach_controller" 00:44:29.120 } 00:44:29.120 EOF 00:44:29.120 )") 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1360 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # gen_fio_conf 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1341 -- # local fio_dir=/usr/src/fio 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@54 -- # local file 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@56 -- # cat 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1343 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1343 -- # local sanitizers 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1344 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # shift 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # cat 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1347 -- # local asan_lib= 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file = 1 )) 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # grep libasan 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@73 -- # cat 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:44:29.120 { 00:44:29.120 "params": { 00:44:29.120 "name": "Nvme$subsystem", 00:44:29.120 "trtype": "$TEST_TRANSPORT", 00:44:29.120 "traddr": "$NVMF_FIRST_TARGET_IP", 00:44:29.120 "adrfam": "ipv4", 00:44:29.120 "trsvcid": "$NVMF_PORT", 00:44:29.120 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:44:29.120 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:44:29.120 "hdgst": ${hdgst:-false}, 00:44:29.120 "ddgst": ${ddgst:-false} 00:44:29.120 }, 00:44:29.120 "method": "bdev_nvme_attach_controller" 00:44:29.120 } 00:44:29.120 EOF 00:44:29.120 )") 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # cat 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file++ )) 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@584 -- # jq . 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@585 -- # IFS=, 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:44:29.120 "params": { 00:44:29.120 "name": "Nvme0", 00:44:29.120 "trtype": "tcp", 00:44:29.120 "traddr": "10.0.0.2", 00:44:29.120 "adrfam": "ipv4", 00:44:29.120 "trsvcid": "4420", 00:44:29.120 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:44:29.120 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:44:29.120 "hdgst": false, 00:44:29.120 "ddgst": false 00:44:29.120 }, 00:44:29.120 "method": "bdev_nvme_attach_controller" 00:44:29.120 },{ 00:44:29.120 "params": { 00:44:29.120 "name": "Nvme1", 00:44:29.120 "trtype": "tcp", 00:44:29.120 "traddr": "10.0.0.2", 00:44:29.120 "adrfam": "ipv4", 00:44:29.120 "trsvcid": "4420", 00:44:29.120 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:44:29.120 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:44:29.120 "hdgst": false, 00:44:29.120 "ddgst": false 00:44:29.120 }, 00:44:29.120 "method": "bdev_nvme_attach_controller" 00:44:29.120 }' 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # asan_lib= 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # grep libclang_rt.asan 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1349 -- # asan_lib= 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1356 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:44:29.120 18:00:09 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1356 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:44:29.120 filename0: (g=0): rw=randread, bs=(R) 8192B-8192B, (W) 16.0KiB-16.0KiB, (T) 128KiB-128KiB, ioengine=spdk_bdev, iodepth=8 00:44:29.120 ... 00:44:29.120 filename1: (g=0): rw=randread, bs=(R) 8192B-8192B, (W) 16.0KiB-16.0KiB, (T) 128KiB-128KiB, ioengine=spdk_bdev, iodepth=8 00:44:29.120 ... 00:44:29.120 fio-3.35 00:44:29.120 Starting 4 threads 00:44:34.414 00:44:34.414 filename0: (groupid=0, jobs=1): err= 0: pid=481148: Fri Dec 6 18:00:15 2024 00:44:34.414 read: IOPS=1799, BW=14.1MiB/s (14.7MB/s)(70.3MiB/5001msec) 00:44:34.414 slat (nsec): min=4248, max=74596, avg=18736.37, stdev=10810.55 00:44:34.414 clat (usec): min=841, max=7933, avg=4379.49, stdev=697.06 00:44:34.414 lat (usec): min=854, max=7941, avg=4398.22, stdev=696.41 00:44:34.414 clat percentiles (usec): 00:44:34.414 | 1.00th=[ 2769], 5.00th=[ 3523], 10.00th=[ 3785], 20.00th=[ 4047], 00:44:34.414 | 30.00th=[ 4146], 40.00th=[ 4228], 50.00th=[ 4293], 60.00th=[ 4359], 00:44:34.414 | 70.00th=[ 4424], 80.00th=[ 4621], 90.00th=[ 5211], 95.00th=[ 5735], 00:44:34.414 | 99.00th=[ 6849], 99.50th=[ 7242], 99.90th=[ 7635], 99.95th=[ 7635], 00:44:34.414 | 99.99th=[ 7963] 00:44:34.414 bw ( KiB/s): min=14048, max=14656, per=24.22%, avg=14369.78, stdev=183.73, samples=9 00:44:34.414 iops : min= 1756, max= 1832, avg=1796.22, stdev=22.97, samples=9 00:44:34.414 lat (usec) : 1000=0.10% 00:44:34.414 lat (msec) : 2=0.44%, 4=16.28%, 10=83.18% 00:44:34.414 cpu : usr=95.18%, sys=3.88%, ctx=65, majf=0, minf=38 00:44:34.414 IO depths : 1=0.1%, 2=14.2%, 4=58.0%, 8=27.6%, 16=0.0%, 32=0.0%, >=64=0.0% 00:44:34.414 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:34.414 complete : 0=0.0%, 4=92.2%, 8=7.8%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:34.414 issued rwts: total=8999,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:34.414 latency : target=0, window=0, percentile=100.00%, depth=8 00:44:34.414 filename0: (groupid=0, jobs=1): err= 0: pid=481149: Fri Dec 6 18:00:15 2024 00:44:34.414 read: IOPS=1921, BW=15.0MiB/s (15.7MB/s)(75.1MiB/5002msec) 00:44:34.414 slat (nsec): min=4153, max=70508, avg=14038.90, stdev=8720.22 00:44:34.414 clat (usec): min=972, max=7709, avg=4114.40, stdev=555.70 00:44:34.414 lat (usec): min=979, max=7720, avg=4128.44, stdev=556.06 00:44:34.414 clat percentiles (usec): 00:44:34.414 | 1.00th=[ 2376], 5.00th=[ 3261], 10.00th=[ 3490], 20.00th=[ 3785], 00:44:34.414 | 30.00th=[ 3949], 40.00th=[ 4080], 50.00th=[ 4178], 60.00th=[ 4228], 00:44:34.414 | 70.00th=[ 4293], 80.00th=[ 4424], 90.00th=[ 4555], 95.00th=[ 4817], 00:44:34.414 | 99.00th=[ 5932], 99.50th=[ 6587], 99.90th=[ 7308], 99.95th=[ 7373], 00:44:34.414 | 99.99th=[ 7701] 00:44:34.414 bw ( KiB/s): min=14976, max=15728, per=25.90%, avg=15367.90, stdev=282.13, samples=10 00:44:34.414 iops : min= 1872, max= 1966, avg=1920.90, stdev=35.17, samples=10 00:44:34.414 lat (usec) : 1000=0.04% 00:44:34.414 lat (msec) : 2=0.35%, 4=32.35%, 10=67.26% 00:44:34.414 cpu : usr=94.86%, sys=4.60%, ctx=10, majf=0, minf=50 00:44:34.414 IO depths : 1=0.4%, 2=14.7%, 4=57.5%, 8=27.4%, 16=0.0%, 32=0.0%, >=64=0.0% 00:44:34.414 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:34.414 complete : 0=0.0%, 4=92.2%, 8=7.8%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:34.414 issued rwts: total=9611,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:34.414 latency : target=0, window=0, percentile=100.00%, depth=8 00:44:34.414 filename1: (groupid=0, jobs=1): err= 0: pid=481150: Fri Dec 6 18:00:15 2024 00:44:34.414 read: IOPS=1854, BW=14.5MiB/s (15.2MB/s)(72.5MiB/5003msec) 00:44:34.414 slat (nsec): min=4450, max=74603, avg=16813.42, stdev=10304.40 00:44:34.414 clat (usec): min=1023, max=8038, avg=4255.66, stdev=622.09 00:44:34.414 lat (usec): min=1041, max=8074, avg=4272.47, stdev=621.94 00:44:34.414 clat percentiles (usec): 00:44:34.414 | 1.00th=[ 2540], 5.00th=[ 3425], 10.00th=[ 3687], 20.00th=[ 3916], 00:44:34.414 | 30.00th=[ 4080], 40.00th=[ 4146], 50.00th=[ 4228], 60.00th=[ 4293], 00:44:34.414 | 70.00th=[ 4359], 80.00th=[ 4490], 90.00th=[ 4883], 95.00th=[ 5276], 00:44:34.414 | 99.00th=[ 6521], 99.50th=[ 7046], 99.90th=[ 7570], 99.95th=[ 7635], 00:44:34.414 | 99.99th=[ 8029] 00:44:34.414 bw ( KiB/s): min=14496, max=15440, per=25.00%, avg=14830.40, stdev=252.58, samples=10 00:44:34.414 iops : min= 1812, max= 1930, avg=1853.80, stdev=31.57, samples=10 00:44:34.414 lat (msec) : 2=0.36%, 4=23.71%, 10=75.93% 00:44:34.414 cpu : usr=95.32%, sys=4.18%, ctx=6, majf=0, minf=41 00:44:34.414 IO depths : 1=0.4%, 2=14.3%, 4=58.2%, 8=27.1%, 16=0.0%, 32=0.0%, >=64=0.0% 00:44:34.414 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:34.414 complete : 0=0.0%, 4=92.1%, 8=7.9%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:34.414 issued rwts: total=9277,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:34.414 latency : target=0, window=0, percentile=100.00%, depth=8 00:44:34.414 filename1: (groupid=0, jobs=1): err= 0: pid=481151: Fri Dec 6 18:00:15 2024 00:44:34.414 read: IOPS=1842, BW=14.4MiB/s (15.1MB/s)(72.0MiB/5001msec) 00:44:34.414 slat (nsec): min=4104, max=71618, avg=19323.93, stdev=9624.24 00:44:34.414 clat (usec): min=934, max=9335, avg=4272.57, stdev=670.56 00:44:34.414 lat (usec): min=953, max=9347, avg=4291.90, stdev=670.37 00:44:34.414 clat percentiles (usec): 00:44:34.414 | 1.00th=[ 2474], 5.00th=[ 3425], 10.00th=[ 3687], 20.00th=[ 3949], 00:44:34.414 | 30.00th=[ 4080], 40.00th=[ 4178], 50.00th=[ 4228], 60.00th=[ 4293], 00:44:34.414 | 70.00th=[ 4359], 80.00th=[ 4490], 90.00th=[ 4948], 95.00th=[ 5473], 00:44:34.414 | 99.00th=[ 6849], 99.50th=[ 7177], 99.90th=[ 7570], 99.95th=[ 7701], 00:44:34.414 | 99.99th=[ 9372] 00:44:34.414 bw ( KiB/s): min=14128, max=15248, per=24.81%, avg=14720.00, stdev=354.63, samples=9 00:44:34.414 iops : min= 1766, max= 1906, avg=1840.00, stdev=44.33, samples=9 00:44:34.414 lat (usec) : 1000=0.08% 00:44:34.414 lat (msec) : 2=0.48%, 4=22.05%, 10=77.39% 00:44:34.414 cpu : usr=96.00%, sys=3.46%, ctx=14, majf=0, minf=46 00:44:34.414 IO depths : 1=0.2%, 2=16.6%, 4=55.9%, 8=27.3%, 16=0.0%, 32=0.0%, >=64=0.0% 00:44:34.414 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:34.414 complete : 0=0.0%, 4=92.0%, 8=8.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:34.414 issued rwts: total=9214,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:34.414 latency : target=0, window=0, percentile=100.00%, depth=8 00:44:34.414 00:44:34.414 Run status group 0 (all jobs): 00:44:34.414 READ: bw=57.9MiB/s (60.7MB/s), 14.1MiB/s-15.0MiB/s (14.7MB/s-15.7MB/s), io=290MiB (304MB), run=5001-5003msec 00:44:34.414 18:00:16 nvmf_dif.fio_dif_rand_params -- target/dif.sh@119 -- # destroy_subsystems 0 1 00:44:34.414 18:00:16 nvmf_dif.fio_dif_rand_params -- target/dif.sh@43 -- # local sub 00:44:34.414 18:00:16 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:44:34.414 18:00:16 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 0 00:44:34.414 18:00:16 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=0 00:44:34.414 18:00:16 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:44:34.414 18:00:16 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:34.414 18:00:16 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:34.414 18:00:16 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:34.414 18:00:16 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:44:34.414 18:00:16 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:34.414 18:00:16 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:34.414 18:00:16 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:34.414 18:00:16 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:44:34.414 18:00:16 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 1 00:44:34.414 18:00:16 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=1 00:44:34.414 18:00:16 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:44:34.414 18:00:16 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:34.414 18:00:16 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:34.414 18:00:16 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:34.414 18:00:16 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null1 00:44:34.414 18:00:16 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:34.414 18:00:16 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:34.414 18:00:16 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:34.414 00:44:34.414 real 0m24.347s 00:44:34.414 user 4m33.600s 00:44:34.414 sys 0m5.858s 00:44:34.414 18:00:16 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1130 -- # xtrace_disable 00:44:34.414 18:00:16 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:44:34.414 ************************************ 00:44:34.414 END TEST fio_dif_rand_params 00:44:34.414 ************************************ 00:44:34.414 18:00:16 nvmf_dif -- target/dif.sh@144 -- # run_test fio_dif_digest fio_dif_digest 00:44:34.414 18:00:16 nvmf_dif -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:44:34.414 18:00:16 nvmf_dif -- common/autotest_common.sh@1111 -- # xtrace_disable 00:44:34.414 18:00:16 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:44:34.414 ************************************ 00:44:34.414 START TEST fio_dif_digest 00:44:34.414 ************************************ 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1129 -- # fio_dif_digest 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- target/dif.sh@123 -- # local NULL_DIF 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- target/dif.sh@124 -- # local bs numjobs runtime iodepth files 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- target/dif.sh@125 -- # local hdgst ddgst 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- target/dif.sh@127 -- # NULL_DIF=3 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- target/dif.sh@127 -- # bs=128k,128k,128k 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- target/dif.sh@127 -- # numjobs=3 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- target/dif.sh@127 -- # iodepth=3 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- target/dif.sh@127 -- # runtime=10 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- target/dif.sh@128 -- # hdgst=true 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- target/dif.sh@128 -- # ddgst=true 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- target/dif.sh@130 -- # create_subsystems 0 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- target/dif.sh@28 -- # local sub 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- target/dif.sh@30 -- # for sub in "$@" 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- target/dif.sh@31 -- # create_subsystem 0 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- target/dif.sh@18 -- # local sub_id=0 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 3 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:44:34.414 bdev_null0 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:44:34.414 [2024-12-06 18:00:16.230588] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- target/dif.sh@131 -- # fio /dev/fd/62 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- target/dif.sh@131 -- # create_json_sub_conf 0 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- target/dif.sh@51 -- # gen_nvmf_target_json 0 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- nvmf/common.sh@560 -- # config=() 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- nvmf/common.sh@560 -- # local subsystem config 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- nvmf/common.sh@562 -- # for subsystem in "${@:-1}" 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- nvmf/common.sh@582 -- # config+=("$(cat <<-EOF 00:44:34.414 { 00:44:34.414 "params": { 00:44:34.414 "name": "Nvme$subsystem", 00:44:34.414 "trtype": "$TEST_TRANSPORT", 00:44:34.414 "traddr": "$NVMF_FIRST_TARGET_IP", 00:44:34.414 "adrfam": "ipv4", 00:44:34.414 "trsvcid": "$NVMF_PORT", 00:44:34.414 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:44:34.414 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:44:34.414 "hdgst": ${hdgst:-false}, 00:44:34.414 "ddgst": ${ddgst:-false} 00:44:34.414 }, 00:44:34.414 "method": "bdev_nvme_attach_controller" 00:44:34.414 } 00:44:34.414 EOF 00:44:34.414 )") 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1360 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1341 -- # local fio_dir=/usr/src/fio 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1343 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1343 -- # local sanitizers 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1344 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:44:34.414 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1345 -- # shift 00:44:34.415 18:00:16 nvmf_dif.fio_dif_digest -- target/dif.sh@82 -- # gen_fio_conf 00:44:34.415 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1347 -- # local asan_lib= 00:44:34.415 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:44:34.415 18:00:16 nvmf_dif.fio_dif_digest -- target/dif.sh@54 -- # local file 00:44:34.415 18:00:16 nvmf_dif.fio_dif_digest -- target/dif.sh@56 -- # cat 00:44:34.415 18:00:16 nvmf_dif.fio_dif_digest -- nvmf/common.sh@582 -- # cat 00:44:34.415 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:44:34.415 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1349 -- # grep libasan 00:44:34.415 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:44:34.415 18:00:16 nvmf_dif.fio_dif_digest -- target/dif.sh@72 -- # (( file = 1 )) 00:44:34.415 18:00:16 nvmf_dif.fio_dif_digest -- target/dif.sh@72 -- # (( file <= files )) 00:44:34.415 18:00:16 nvmf_dif.fio_dif_digest -- nvmf/common.sh@584 -- # jq . 00:44:34.415 18:00:16 nvmf_dif.fio_dif_digest -- nvmf/common.sh@585 -- # IFS=, 00:44:34.415 18:00:16 nvmf_dif.fio_dif_digest -- nvmf/common.sh@586 -- # printf '%s\n' '{ 00:44:34.415 "params": { 00:44:34.415 "name": "Nvme0", 00:44:34.415 "trtype": "tcp", 00:44:34.415 "traddr": "10.0.0.2", 00:44:34.415 "adrfam": "ipv4", 00:44:34.415 "trsvcid": "4420", 00:44:34.415 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:44:34.415 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:44:34.415 "hdgst": true, 00:44:34.415 "ddgst": true 00:44:34.415 }, 00:44:34.415 "method": "bdev_nvme_attach_controller" 00:44:34.415 }' 00:44:34.672 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1349 -- # asan_lib= 00:44:34.672 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:44:34.672 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1348 -- # for sanitizer in "${sanitizers[@]}" 00:44:34.672 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1349 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:44:34.672 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1349 -- # grep libclang_rt.asan 00:44:34.672 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1349 -- # awk '{print $3}' 00:44:34.672 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1349 -- # asan_lib= 00:44:34.672 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1350 -- # [[ -n '' ]] 00:44:34.672 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1356 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:44:34.672 18:00:16 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1356 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:44:34.672 filename0: (g=0): rw=randread, bs=(R) 128KiB-128KiB, (W) 128KiB-128KiB, (T) 128KiB-128KiB, ioengine=spdk_bdev, iodepth=3 00:44:34.672 ... 00:44:34.672 fio-3.35 00:44:34.672 Starting 3 threads 00:44:46.931 00:44:46.931 filename0: (groupid=0, jobs=1): err= 0: pid=481920: Fri Dec 6 18:00:27 2024 00:44:46.931 read: IOPS=211, BW=26.4MiB/s (27.7MB/s)(265MiB/10047msec) 00:44:46.931 slat (nsec): min=4413, max=36677, avg=14763.03, stdev=2301.85 00:44:46.931 clat (usec): min=11092, max=55326, avg=14173.60, stdev=2178.76 00:44:46.931 lat (usec): min=11106, max=55339, avg=14188.36, stdev=2178.76 00:44:46.931 clat percentiles (usec): 00:44:46.931 | 1.00th=[11863], 5.00th=[12518], 10.00th=[12780], 20.00th=[13304], 00:44:46.931 | 30.00th=[13566], 40.00th=[13829], 50.00th=[14091], 60.00th=[14353], 00:44:46.931 | 70.00th=[14615], 80.00th=[14877], 90.00th=[15270], 95.00th=[15664], 00:44:46.931 | 99.00th=[16909], 99.50th=[17433], 99.90th=[54264], 99.95th=[55313], 00:44:46.931 | 99.99th=[55313] 00:44:46.931 bw ( KiB/s): min=25600, max=27904, per=34.80%, avg=27110.40, stdev=637.43, samples=20 00:44:46.931 iops : min= 200, max= 218, avg=211.80, stdev= 4.98, samples=20 00:44:46.931 lat (msec) : 20=99.62%, 50=0.14%, 100=0.24% 00:44:46.931 cpu : usr=93.53%, sys=5.85%, ctx=96, majf=0, minf=160 00:44:46.931 IO depths : 1=0.2%, 2=99.8%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:44:46.931 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:46.931 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:46.931 issued rwts: total=2121,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:46.931 latency : target=0, window=0, percentile=100.00%, depth=3 00:44:46.931 filename0: (groupid=0, jobs=1): err= 0: pid=481921: Fri Dec 6 18:00:27 2024 00:44:46.931 read: IOPS=197, BW=24.7MiB/s (25.9MB/s)(248MiB/10046msec) 00:44:46.931 slat (nsec): min=5328, max=49692, avg=15786.10, stdev=3326.20 00:44:46.931 clat (usec): min=9546, max=50553, avg=15164.63, stdev=1449.92 00:44:46.931 lat (usec): min=9561, max=50572, avg=15180.42, stdev=1450.09 00:44:46.931 clat percentiles (usec): 00:44:46.931 | 1.00th=[12649], 5.00th=[13566], 10.00th=[13960], 20.00th=[14353], 00:44:46.931 | 30.00th=[14615], 40.00th=[14877], 50.00th=[15139], 60.00th=[15401], 00:44:46.931 | 70.00th=[15664], 80.00th=[15926], 90.00th=[16450], 95.00th=[16712], 00:44:46.931 | 99.00th=[17433], 99.50th=[17695], 99.90th=[45351], 99.95th=[50594], 00:44:46.931 | 99.99th=[50594] 00:44:46.931 bw ( KiB/s): min=24576, max=26112, per=32.53%, avg=25344.00, stdev=462.44, samples=20 00:44:46.931 iops : min= 192, max= 204, avg=198.00, stdev= 3.61, samples=20 00:44:46.931 lat (msec) : 10=0.10%, 20=99.80%, 50=0.05%, 100=0.05% 00:44:46.931 cpu : usr=91.53%, sys=6.75%, ctx=265, majf=0, minf=118 00:44:46.931 IO depths : 1=0.1%, 2=99.9%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:44:46.931 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:46.931 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:46.931 issued rwts: total=1982,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:46.931 latency : target=0, window=0, percentile=100.00%, depth=3 00:44:46.931 filename0: (groupid=0, jobs=1): err= 0: pid=481922: Fri Dec 6 18:00:27 2024 00:44:46.931 read: IOPS=200, BW=25.0MiB/s (26.2MB/s)(252MiB/10047msec) 00:44:46.931 slat (nsec): min=4545, max=41590, avg=15900.70, stdev=3032.84 00:44:46.931 clat (usec): min=8293, max=51080, avg=14938.93, stdev=1538.97 00:44:46.931 lat (usec): min=8307, max=51102, avg=14954.83, stdev=1538.95 00:44:46.931 clat percentiles (usec): 00:44:46.931 | 1.00th=[12125], 5.00th=[13304], 10.00th=[13698], 20.00th=[14091], 00:44:46.931 | 30.00th=[14484], 40.00th=[14615], 50.00th=[14877], 60.00th=[15139], 00:44:46.931 | 70.00th=[15401], 80.00th=[15664], 90.00th=[16188], 95.00th=[16581], 00:44:46.931 | 99.00th=[17433], 99.50th=[17695], 99.90th=[19006], 99.95th=[47973], 00:44:46.931 | 99.99th=[51119] 00:44:46.931 bw ( KiB/s): min=24832, max=27136, per=33.02%, avg=25728.00, stdev=458.70, samples=20 00:44:46.931 iops : min= 194, max= 212, avg=201.00, stdev= 3.58, samples=20 00:44:46.931 lat (msec) : 10=0.55%, 20=99.35%, 50=0.05%, 100=0.05% 00:44:46.931 cpu : usr=92.88%, sys=6.16%, ctx=91, majf=0, minf=141 00:44:46.931 IO depths : 1=0.1%, 2=100.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:44:46.931 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:46.931 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:44:46.931 issued rwts: total=2012,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:44:46.931 latency : target=0, window=0, percentile=100.00%, depth=3 00:44:46.931 00:44:46.931 Run status group 0 (all jobs): 00:44:46.931 READ: bw=76.1MiB/s (79.8MB/s), 24.7MiB/s-26.4MiB/s (25.9MB/s-27.7MB/s), io=764MiB (802MB), run=10046-10047msec 00:44:46.931 18:00:27 nvmf_dif.fio_dif_digest -- target/dif.sh@132 -- # destroy_subsystems 0 00:44:46.931 18:00:27 nvmf_dif.fio_dif_digest -- target/dif.sh@43 -- # local sub 00:44:46.931 18:00:27 nvmf_dif.fio_dif_digest -- target/dif.sh@45 -- # for sub in "$@" 00:44:46.931 18:00:27 nvmf_dif.fio_dif_digest -- target/dif.sh@46 -- # destroy_subsystem 0 00:44:46.931 18:00:27 nvmf_dif.fio_dif_digest -- target/dif.sh@36 -- # local sub_id=0 00:44:46.931 18:00:27 nvmf_dif.fio_dif_digest -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:44:46.931 18:00:27 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:46.931 18:00:27 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:44:46.931 18:00:27 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:46.931 18:00:27 nvmf_dif.fio_dif_digest -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:44:46.931 18:00:27 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:46.931 18:00:27 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:44:46.931 18:00:27 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:46.931 00:44:46.931 real 0m11.150s 00:44:46.931 user 0m29.156s 00:44:46.931 sys 0m2.160s 00:44:46.931 18:00:27 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1130 -- # xtrace_disable 00:44:46.931 18:00:27 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:44:46.931 ************************************ 00:44:46.931 END TEST fio_dif_digest 00:44:46.931 ************************************ 00:44:46.931 18:00:27 nvmf_dif -- target/dif.sh@146 -- # trap - SIGINT SIGTERM EXIT 00:44:46.931 18:00:27 nvmf_dif -- target/dif.sh@147 -- # nvmftestfini 00:44:46.931 18:00:27 nvmf_dif -- nvmf/common.sh@516 -- # nvmfcleanup 00:44:46.931 18:00:27 nvmf_dif -- nvmf/common.sh@121 -- # sync 00:44:46.931 18:00:27 nvmf_dif -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:44:46.931 18:00:27 nvmf_dif -- nvmf/common.sh@124 -- # set +e 00:44:46.931 18:00:27 nvmf_dif -- nvmf/common.sh@125 -- # for i in {1..20} 00:44:46.931 18:00:27 nvmf_dif -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:44:46.931 rmmod nvme_tcp 00:44:46.931 rmmod nvme_fabrics 00:44:46.931 rmmod nvme_keyring 00:44:46.931 18:00:27 nvmf_dif -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:44:46.931 18:00:27 nvmf_dif -- nvmf/common.sh@128 -- # set -e 00:44:46.931 18:00:27 nvmf_dif -- nvmf/common.sh@129 -- # return 0 00:44:46.931 18:00:27 nvmf_dif -- nvmf/common.sh@517 -- # '[' -n 475257 ']' 00:44:46.931 18:00:27 nvmf_dif -- nvmf/common.sh@518 -- # killprocess 475257 00:44:46.931 18:00:27 nvmf_dif -- common/autotest_common.sh@954 -- # '[' -z 475257 ']' 00:44:46.931 18:00:27 nvmf_dif -- common/autotest_common.sh@958 -- # kill -0 475257 00:44:46.931 18:00:27 nvmf_dif -- common/autotest_common.sh@959 -- # uname 00:44:46.931 18:00:27 nvmf_dif -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:44:46.931 18:00:27 nvmf_dif -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 475257 00:44:46.931 18:00:27 nvmf_dif -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:44:46.931 18:00:27 nvmf_dif -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:44:46.931 18:00:27 nvmf_dif -- common/autotest_common.sh@972 -- # echo 'killing process with pid 475257' 00:44:46.931 killing process with pid 475257 00:44:46.931 18:00:27 nvmf_dif -- common/autotest_common.sh@973 -- # kill 475257 00:44:46.931 18:00:27 nvmf_dif -- common/autotest_common.sh@978 -- # wait 475257 00:44:46.931 18:00:27 nvmf_dif -- nvmf/common.sh@520 -- # '[' iso == iso ']' 00:44:46.931 18:00:27 nvmf_dif -- nvmf/common.sh@521 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:44:47.193 Waiting for block devices as requested 00:44:47.193 0000:88:00.0 (8086 0a54): vfio-pci -> nvme 00:44:47.193 0000:00:04.7 (8086 0e27): vfio-pci -> ioatdma 00:44:47.452 0000:00:04.6 (8086 0e26): vfio-pci -> ioatdma 00:44:47.452 0000:00:04.5 (8086 0e25): vfio-pci -> ioatdma 00:44:47.710 0000:00:04.4 (8086 0e24): vfio-pci -> ioatdma 00:44:47.710 0000:00:04.3 (8086 0e23): vfio-pci -> ioatdma 00:44:47.710 0000:00:04.2 (8086 0e22): vfio-pci -> ioatdma 00:44:47.710 0000:00:04.1 (8086 0e21): vfio-pci -> ioatdma 00:44:47.969 0000:00:04.0 (8086 0e20): vfio-pci -> ioatdma 00:44:47.969 0000:80:04.7 (8086 0e27): vfio-pci -> ioatdma 00:44:47.969 0000:80:04.6 (8086 0e26): vfio-pci -> ioatdma 00:44:47.969 0000:80:04.5 (8086 0e25): vfio-pci -> ioatdma 00:44:48.228 0000:80:04.4 (8086 0e24): vfio-pci -> ioatdma 00:44:48.228 0000:80:04.3 (8086 0e23): vfio-pci -> ioatdma 00:44:48.228 0000:80:04.2 (8086 0e22): vfio-pci -> ioatdma 00:44:48.228 0000:80:04.1 (8086 0e21): vfio-pci -> ioatdma 00:44:48.228 0000:80:04.0 (8086 0e20): vfio-pci -> ioatdma 00:44:48.485 18:00:30 nvmf_dif -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:44:48.485 18:00:30 nvmf_dif -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:44:48.485 18:00:30 nvmf_dif -- nvmf/common.sh@297 -- # iptr 00:44:48.485 18:00:30 nvmf_dif -- nvmf/common.sh@791 -- # iptables-save 00:44:48.485 18:00:30 nvmf_dif -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:44:48.485 18:00:30 nvmf_dif -- nvmf/common.sh@791 -- # iptables-restore 00:44:48.485 18:00:30 nvmf_dif -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:44:48.485 18:00:30 nvmf_dif -- nvmf/common.sh@302 -- # remove_spdk_ns 00:44:48.485 18:00:30 nvmf_dif -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:44:48.485 18:00:30 nvmf_dif -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:44:48.485 18:00:30 nvmf_dif -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:44:51.016 18:00:32 nvmf_dif -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:44:51.017 00:44:51.017 real 1m7.262s 00:44:51.017 user 6m30.795s 00:44:51.017 sys 0m17.241s 00:44:51.017 18:00:32 nvmf_dif -- common/autotest_common.sh@1130 -- # xtrace_disable 00:44:51.017 18:00:32 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:44:51.017 ************************************ 00:44:51.017 END TEST nvmf_dif 00:44:51.017 ************************************ 00:44:51.017 18:00:32 -- spdk/autotest.sh@290 -- # run_test nvmf_abort_qd_sizes /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort_qd_sizes.sh 00:44:51.017 18:00:32 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:44:51.017 18:00:32 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:44:51.017 18:00:32 -- common/autotest_common.sh@10 -- # set +x 00:44:51.017 ************************************ 00:44:51.017 START TEST nvmf_abort_qd_sizes 00:44:51.017 ************************************ 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort_qd_sizes.sh 00:44:51.017 * Looking for test storage... 00:44:51.017 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- common/autotest_common.sh@1711 -- # lcov --version 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@333 -- # local ver1 ver1_l 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@334 -- # local ver2 ver2_l 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@336 -- # IFS=.-: 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@336 -- # read -ra ver1 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@337 -- # IFS=.-: 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@337 -- # read -ra ver2 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@338 -- # local 'op=<' 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@340 -- # ver1_l=2 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@341 -- # ver2_l=1 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@344 -- # case "$op" in 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@345 -- # : 1 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@364 -- # (( v = 0 )) 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@365 -- # decimal 1 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@353 -- # local d=1 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@355 -- # echo 1 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@365 -- # ver1[v]=1 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@366 -- # decimal 2 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@353 -- # local d=2 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@355 -- # echo 2 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@366 -- # ver2[v]=2 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@368 -- # return 0 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:44:51.017 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:44:51.017 --rc genhtml_branch_coverage=1 00:44:51.017 --rc genhtml_function_coverage=1 00:44:51.017 --rc genhtml_legend=1 00:44:51.017 --rc geninfo_all_blocks=1 00:44:51.017 --rc geninfo_unexecuted_blocks=1 00:44:51.017 00:44:51.017 ' 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:44:51.017 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:44:51.017 --rc genhtml_branch_coverage=1 00:44:51.017 --rc genhtml_function_coverage=1 00:44:51.017 --rc genhtml_legend=1 00:44:51.017 --rc geninfo_all_blocks=1 00:44:51.017 --rc geninfo_unexecuted_blocks=1 00:44:51.017 00:44:51.017 ' 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:44:51.017 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:44:51.017 --rc genhtml_branch_coverage=1 00:44:51.017 --rc genhtml_function_coverage=1 00:44:51.017 --rc genhtml_legend=1 00:44:51.017 --rc geninfo_all_blocks=1 00:44:51.017 --rc geninfo_unexecuted_blocks=1 00:44:51.017 00:44:51.017 ' 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:44:51.017 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:44:51.017 --rc genhtml_branch_coverage=1 00:44:51.017 --rc genhtml_function_coverage=1 00:44:51.017 --rc genhtml_legend=1 00:44:51.017 --rc geninfo_all_blocks=1 00:44:51.017 --rc geninfo_unexecuted_blocks=1 00:44:51.017 00:44:51.017 ' 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@14 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@7 -- # uname -s 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@15 -- # shopt -s extglob 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- paths/export.sh@5 -- # export PATH 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@51 -- # : 0 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:44:51.017 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@55 -- # have_pci_nics=0 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@70 -- # nvmftestinit 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@469 -- # '[' -z tcp ']' 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@474 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@476 -- # prepare_net_devs 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@438 -- # local -g is_hw=no 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@440 -- # remove_spdk_ns 00:44:51.017 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:44:51.018 18:00:32 nvmf_abort_qd_sizes -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:44:51.018 18:00:32 nvmf_abort_qd_sizes -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:44:51.018 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@442 -- # [[ phy != virt ]] 00:44:51.018 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@442 -- # gather_supported_nvmf_pci_devs 00:44:51.018 18:00:32 nvmf_abort_qd_sizes -- nvmf/common.sh@309 -- # xtrace_disable 00:44:51.018 18:00:32 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@315 -- # pci_devs=() 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@315 -- # local -a pci_devs 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@316 -- # pci_net_devs=() 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@317 -- # pci_drivers=() 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@317 -- # local -A pci_drivers 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@319 -- # net_devs=() 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@319 -- # local -ga net_devs 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@320 -- # e810=() 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@320 -- # local -ga e810 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@321 -- # x722=() 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@321 -- # local -ga x722 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@322 -- # mlx=() 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@322 -- # local -ga mlx 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.0 (0x8086 - 0x159b)' 00:44:52.920 Found 0000:0a:00.0 (0x8086 - 0x159b) 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@367 -- # echo 'Found 0000:0a:00.1 (0x8086 - 0x159b)' 00:44:52.920 Found 0000:0a:00.1 (0x8086 - 0x159b) 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:44:52.920 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@418 -- # [[ up == up ]] 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.0: cvl_0_0' 00:44:52.921 Found net devices under 0000:0a:00.0: cvl_0_0 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@410 -- # for pci in "${pci_devs[@]}" 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@411 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@416 -- # [[ tcp == tcp ]] 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@417 -- # for net_dev in "${!pci_net_devs[@]}" 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@418 -- # [[ up == up ]] 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@422 -- # (( 1 == 0 )) 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@427 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@428 -- # echo 'Found net devices under 0000:0a:00.1: cvl_0_1' 00:44:52.921 Found net devices under 0000:0a:00.1: cvl_0_1 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@429 -- # net_devs+=("${pci_net_devs[@]}") 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@432 -- # (( 2 == 0 )) 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@442 -- # is_hw=yes 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@444 -- # [[ yes == yes ]] 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@445 -- # [[ tcp == tcp ]] 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@446 -- # nvmf_tcp_init 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@790 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:44:52.921 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:44:52.921 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.229 ms 00:44:52.921 00:44:52.921 --- 10.0.0.2 ping statistics --- 00:44:52.921 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:44:52.921 rtt min/avg/max/mdev = 0.229/0.229/0.229/0.000 ms 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:44:52.921 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:44:52.921 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.191 ms 00:44:52.921 00:44:52.921 --- 10.0.0.1 ping statistics --- 00:44:52.921 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:44:52.921 rtt min/avg/max/mdev = 0.191/0.191/0.191/0.000 ms 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@450 -- # return 0 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@478 -- # '[' iso == iso ']' 00:44:52.921 18:00:34 nvmf_abort_qd_sizes -- nvmf/common.sh@479 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:44:54.327 0000:00:04.7 (8086 0e27): ioatdma -> vfio-pci 00:44:54.327 0000:00:04.6 (8086 0e26): ioatdma -> vfio-pci 00:44:54.327 0000:00:04.5 (8086 0e25): ioatdma -> vfio-pci 00:44:54.327 0000:00:04.4 (8086 0e24): ioatdma -> vfio-pci 00:44:54.327 0000:00:04.3 (8086 0e23): ioatdma -> vfio-pci 00:44:54.327 0000:00:04.2 (8086 0e22): ioatdma -> vfio-pci 00:44:54.327 0000:00:04.1 (8086 0e21): ioatdma -> vfio-pci 00:44:54.327 0000:00:04.0 (8086 0e20): ioatdma -> vfio-pci 00:44:54.327 0000:80:04.7 (8086 0e27): ioatdma -> vfio-pci 00:44:54.327 0000:80:04.6 (8086 0e26): ioatdma -> vfio-pci 00:44:54.327 0000:80:04.5 (8086 0e25): ioatdma -> vfio-pci 00:44:54.327 0000:80:04.4 (8086 0e24): ioatdma -> vfio-pci 00:44:54.327 0000:80:04.3 (8086 0e23): ioatdma -> vfio-pci 00:44:54.327 0000:80:04.2 (8086 0e22): ioatdma -> vfio-pci 00:44:54.327 0000:80:04.1 (8086 0e21): ioatdma -> vfio-pci 00:44:54.327 0000:80:04.0 (8086 0e20): ioatdma -> vfio-pci 00:44:55.266 0000:88:00.0 (8086 0a54): nvme -> vfio-pci 00:44:55.266 18:00:37 nvmf_abort_qd_sizes -- nvmf/common.sh@482 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:44:55.266 18:00:37 nvmf_abort_qd_sizes -- nvmf/common.sh@483 -- # [[ tcp == \r\d\m\a ]] 00:44:55.266 18:00:37 nvmf_abort_qd_sizes -- nvmf/common.sh@492 -- # [[ tcp == \t\c\p ]] 00:44:55.266 18:00:37 nvmf_abort_qd_sizes -- nvmf/common.sh@493 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:44:55.266 18:00:37 nvmf_abort_qd_sizes -- nvmf/common.sh@496 -- # '[' tcp == tcp ']' 00:44:55.266 18:00:37 nvmf_abort_qd_sizes -- nvmf/common.sh@502 -- # modprobe nvme-tcp 00:44:55.525 18:00:37 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@71 -- # nvmfappstart -m 0xf 00:44:55.525 18:00:37 nvmf_abort_qd_sizes -- nvmf/common.sh@507 -- # timing_enter start_nvmf_tgt 00:44:55.525 18:00:37 nvmf_abort_qd_sizes -- common/autotest_common.sh@726 -- # xtrace_disable 00:44:55.525 18:00:37 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:44:55.525 18:00:37 nvmf_abort_qd_sizes -- nvmf/common.sh@509 -- # nvmfpid=486822 00:44:55.525 18:00:37 nvmf_abort_qd_sizes -- nvmf/common.sh@508 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xf 00:44:55.525 18:00:37 nvmf_abort_qd_sizes -- nvmf/common.sh@510 -- # waitforlisten 486822 00:44:55.525 18:00:37 nvmf_abort_qd_sizes -- common/autotest_common.sh@835 -- # '[' -z 486822 ']' 00:44:55.525 18:00:37 nvmf_abort_qd_sizes -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:44:55.525 18:00:37 nvmf_abort_qd_sizes -- common/autotest_common.sh@840 -- # local max_retries=100 00:44:55.525 18:00:37 nvmf_abort_qd_sizes -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:44:55.525 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:44:55.525 18:00:37 nvmf_abort_qd_sizes -- common/autotest_common.sh@844 -- # xtrace_disable 00:44:55.525 18:00:37 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:44:55.525 [2024-12-06 18:00:37.156579] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:44:55.525 [2024-12-06 18:00:37.156683] [ DPDK EAL parameters: nvmf -c 0xf --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:44:55.525 [2024-12-06 18:00:37.227980] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:44:55.525 [2024-12-06 18:00:37.272285] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:44:55.525 [2024-12-06 18:00:37.272345] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:44:55.525 [2024-12-06 18:00:37.272367] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:44:55.525 [2024-12-06 18:00:37.272378] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:44:55.525 [2024-12-06 18:00:37.272387] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:44:55.525 [2024-12-06 18:00:37.273865] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:44:55.525 [2024-12-06 18:00:37.273893] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:44:55.525 [2024-12-06 18:00:37.273952] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:44:55.525 [2024-12-06 18:00:37.273955] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:44:55.783 18:00:37 nvmf_abort_qd_sizes -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:44:55.783 18:00:37 nvmf_abort_qd_sizes -- common/autotest_common.sh@868 -- # return 0 00:44:55.783 18:00:37 nvmf_abort_qd_sizes -- nvmf/common.sh@511 -- # timing_exit start_nvmf_tgt 00:44:55.783 18:00:37 nvmf_abort_qd_sizes -- common/autotest_common.sh@732 -- # xtrace_disable 00:44:55.783 18:00:37 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:44:55.783 18:00:37 nvmf_abort_qd_sizes -- nvmf/common.sh@512 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:44:55.783 18:00:37 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@73 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini || :; clean_kernel_target' SIGINT SIGTERM EXIT 00:44:55.783 18:00:37 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@75 -- # mapfile -t nvmes 00:44:55.783 18:00:37 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@75 -- # nvme_in_userspace 00:44:55.783 18:00:37 nvmf_abort_qd_sizes -- scripts/common.sh@312 -- # local bdf bdfs 00:44:55.783 18:00:37 nvmf_abort_qd_sizes -- scripts/common.sh@313 -- # local nvmes 00:44:55.783 18:00:37 nvmf_abort_qd_sizes -- scripts/common.sh@315 -- # [[ -n 0000:88:00.0 ]] 00:44:55.783 18:00:37 nvmf_abort_qd_sizes -- scripts/common.sh@316 -- # nvmes=(${pci_bus_cache["0x010802"]}) 00:44:55.783 18:00:37 nvmf_abort_qd_sizes -- scripts/common.sh@321 -- # for bdf in "${nvmes[@]}" 00:44:55.783 18:00:37 nvmf_abort_qd_sizes -- scripts/common.sh@322 -- # [[ -e /sys/bus/pci/drivers/nvme/0000:88:00.0 ]] 00:44:55.783 18:00:37 nvmf_abort_qd_sizes -- scripts/common.sh@323 -- # uname -s 00:44:55.784 18:00:37 nvmf_abort_qd_sizes -- scripts/common.sh@323 -- # [[ Linux == FreeBSD ]] 00:44:55.784 18:00:37 nvmf_abort_qd_sizes -- scripts/common.sh@326 -- # bdfs+=("$bdf") 00:44:55.784 18:00:37 nvmf_abort_qd_sizes -- scripts/common.sh@328 -- # (( 1 )) 00:44:55.784 18:00:37 nvmf_abort_qd_sizes -- scripts/common.sh@329 -- # printf '%s\n' 0000:88:00.0 00:44:55.784 18:00:37 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@76 -- # (( 1 > 0 )) 00:44:55.784 18:00:37 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@78 -- # nvme=0000:88:00.0 00:44:55.784 18:00:37 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@80 -- # run_test spdk_target_abort spdk_target 00:44:55.784 18:00:37 nvmf_abort_qd_sizes -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:44:55.784 18:00:37 nvmf_abort_qd_sizes -- common/autotest_common.sh@1111 -- # xtrace_disable 00:44:55.784 18:00:37 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:44:55.784 ************************************ 00:44:55.784 START TEST spdk_target_abort 00:44:55.784 ************************************ 00:44:55.784 18:00:37 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@1129 -- # spdk_target 00:44:55.784 18:00:37 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@43 -- # local name=spdk_target 00:44:55.784 18:00:37 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@45 -- # rpc_cmd bdev_nvme_attach_controller -t pcie -a 0000:88:00.0 -b spdk_target 00:44:55.784 18:00:37 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:55.784 18:00:37 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:44:59.062 spdk_targetn1 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@47 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:44:59.062 [2024-12-06 18:00:40.288284] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@48 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:testnqn -a -s SPDKISFASTANDAWESOME 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@49 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:testnqn spdk_targetn1 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@50 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:testnqn -t tcp -a 10.0.0.2 -s 4420 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:44:59.062 [2024-12-06 18:00:40.336627] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@52 -- # rabort tcp IPv4 10.0.0.2 4420 nqn.2016-06.io.spdk:testnqn 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@17 -- # local trtype=tcp 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@18 -- # local adrfam=IPv4 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@19 -- # local traddr=10.0.0.2 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@20 -- # local trsvcid=4420 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@21 -- # local subnqn=nqn.2016-06.io.spdk:testnqn 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@23 -- # local qds qd 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@24 -- # local target r 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@26 -- # qds=(4 24 64) 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@29 -- # target=trtype:tcp 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4' 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.2' 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:44:59.062 18:00:40 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 4 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:45:02.357 Initializing NVMe Controllers 00:45:02.357 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:testnqn 00:45:02.357 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:45:02.357 Initialization complete. Launching workers. 00:45:02.357 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 12947, failed: 0 00:45:02.357 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 1148, failed to submit 11799 00:45:02.357 success 771, unsuccessful 377, failed 0 00:45:02.357 18:00:43 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:45:02.357 18:00:43 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 24 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:45:05.633 Initializing NVMe Controllers 00:45:05.633 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:testnqn 00:45:05.633 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:45:05.633 Initialization complete. Launching workers. 00:45:05.633 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 8526, failed: 0 00:45:05.633 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 1262, failed to submit 7264 00:45:05.633 success 305, unsuccessful 957, failed 0 00:45:05.633 18:00:46 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:45:05.633 18:00:46 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 64 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:45:08.917 Initializing NVMe Controllers 00:45:08.917 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:testnqn 00:45:08.917 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:45:08.917 Initialization complete. Launching workers. 00:45:08.917 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 31846, failed: 0 00:45:08.917 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 2715, failed to submit 29131 00:45:08.917 success 487, unsuccessful 2228, failed 0 00:45:08.917 18:00:50 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@54 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:testnqn 00:45:08.917 18:00:50 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:45:08.917 18:00:50 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:45:08.917 18:00:50 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:45:08.917 18:00:50 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@55 -- # rpc_cmd bdev_nvme_detach_controller spdk_target 00:45:08.917 18:00:50 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@563 -- # xtrace_disable 00:45:08.917 18:00:50 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:45:09.848 18:00:51 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:45:09.848 18:00:51 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@61 -- # killprocess 486822 00:45:09.848 18:00:51 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@954 -- # '[' -z 486822 ']' 00:45:09.848 18:00:51 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@958 -- # kill -0 486822 00:45:09.848 18:00:51 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@959 -- # uname 00:45:09.848 18:00:51 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:45:09.848 18:00:51 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 486822 00:45:10.106 18:00:51 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:45:10.106 18:00:51 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:45:10.106 18:00:51 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@972 -- # echo 'killing process with pid 486822' 00:45:10.106 killing process with pid 486822 00:45:10.106 18:00:51 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@973 -- # kill 486822 00:45:10.106 18:00:51 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@978 -- # wait 486822 00:45:10.106 00:45:10.106 real 0m14.451s 00:45:10.106 user 0m54.992s 00:45:10.106 sys 0m2.475s 00:45:10.106 18:00:51 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@1130 -- # xtrace_disable 00:45:10.106 18:00:51 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:45:10.106 ************************************ 00:45:10.106 END TEST spdk_target_abort 00:45:10.106 ************************************ 00:45:10.106 18:00:51 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@81 -- # run_test kernel_target_abort kernel_target 00:45:10.106 18:00:51 nvmf_abort_qd_sizes -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:45:10.106 18:00:51 nvmf_abort_qd_sizes -- common/autotest_common.sh@1111 -- # xtrace_disable 00:45:10.106 18:00:51 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:45:10.365 ************************************ 00:45:10.365 START TEST kernel_target_abort 00:45:10.365 ************************************ 00:45:10.365 18:00:51 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@1129 -- # kernel_target 00:45:10.365 18:00:51 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@65 -- # get_main_ns_ip 00:45:10.365 18:00:51 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@769 -- # local ip 00:45:10.365 18:00:51 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@770 -- # ip_candidates=() 00:45:10.365 18:00:51 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@770 -- # local -A ip_candidates 00:45:10.365 18:00:51 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@772 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:45:10.365 18:00:51 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@773 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:45:10.365 18:00:51 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@775 -- # [[ -z tcp ]] 00:45:10.365 18:00:51 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@775 -- # [[ -z NVMF_INITIATOR_IP ]] 00:45:10.365 18:00:51 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@776 -- # ip=NVMF_INITIATOR_IP 00:45:10.365 18:00:51 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@778 -- # [[ -z 10.0.0.1 ]] 00:45:10.365 18:00:51 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@783 -- # echo 10.0.0.1 00:45:10.365 18:00:51 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@65 -- # configure_kernel_target nqn.2016-06.io.spdk:testnqn 10.0.0.1 00:45:10.365 18:00:51 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@660 -- # local kernel_name=nqn.2016-06.io.spdk:testnqn kernel_target_ip=10.0.0.1 00:45:10.365 18:00:51 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@662 -- # nvmet=/sys/kernel/config/nvmet 00:45:10.365 18:00:51 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@663 -- # kernel_subsystem=/sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:45:10.365 18:00:51 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@664 -- # kernel_namespace=/sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:45:10.365 18:00:51 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@665 -- # kernel_port=/sys/kernel/config/nvmet/ports/1 00:45:10.365 18:00:51 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@667 -- # local block nvme 00:45:10.365 18:00:51 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@669 -- # [[ ! -e /sys/module/nvmet ]] 00:45:10.365 18:00:51 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@670 -- # modprobe nvmet 00:45:10.365 18:00:51 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@673 -- # [[ -e /sys/kernel/config/nvmet ]] 00:45:10.365 18:00:51 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@675 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:45:11.301 Waiting for block devices as requested 00:45:11.559 0000:88:00.0 (8086 0a54): vfio-pci -> nvme 00:45:11.559 0000:00:04.7 (8086 0e27): vfio-pci -> ioatdma 00:45:11.559 0000:00:04.6 (8086 0e26): vfio-pci -> ioatdma 00:45:11.817 0000:00:04.5 (8086 0e25): vfio-pci -> ioatdma 00:45:11.817 0000:00:04.4 (8086 0e24): vfio-pci -> ioatdma 00:45:11.817 0000:00:04.3 (8086 0e23): vfio-pci -> ioatdma 00:45:12.076 0000:00:04.2 (8086 0e22): vfio-pci -> ioatdma 00:45:12.076 0000:00:04.1 (8086 0e21): vfio-pci -> ioatdma 00:45:12.076 0000:00:04.0 (8086 0e20): vfio-pci -> ioatdma 00:45:12.076 0000:80:04.7 (8086 0e27): vfio-pci -> ioatdma 00:45:12.335 0000:80:04.6 (8086 0e26): vfio-pci -> ioatdma 00:45:12.335 0000:80:04.5 (8086 0e25): vfio-pci -> ioatdma 00:45:12.335 0000:80:04.4 (8086 0e24): vfio-pci -> ioatdma 00:45:12.335 0000:80:04.3 (8086 0e23): vfio-pci -> ioatdma 00:45:12.593 0000:80:04.2 (8086 0e22): vfio-pci -> ioatdma 00:45:12.593 0000:80:04.1 (8086 0e21): vfio-pci -> ioatdma 00:45:12.593 0000:80:04.0 (8086 0e20): vfio-pci -> ioatdma 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@678 -- # for block in /sys/block/nvme* 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@679 -- # [[ -e /sys/block/nvme0n1 ]] 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@680 -- # is_block_zoned nvme0n1 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@1650 -- # local device=nvme0n1 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@1652 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@1653 -- # [[ none != none ]] 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@681 -- # block_in_use nvme0n1 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- scripts/common.sh@381 -- # local block=nvme0n1 pt 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- scripts/common.sh@390 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdk-gpt.py nvme0n1 00:45:12.852 No valid GPT data, bailing 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- scripts/common.sh@394 -- # pt= 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- scripts/common.sh@395 -- # return 1 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@681 -- # nvme=/dev/nvme0n1 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@684 -- # [[ -b /dev/nvme0n1 ]] 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@686 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@687 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@688 -- # mkdir /sys/kernel/config/nvmet/ports/1 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@693 -- # echo SPDK-nqn.2016-06.io.spdk:testnqn 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@695 -- # echo 1 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@696 -- # echo /dev/nvme0n1 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@697 -- # echo 1 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@699 -- # echo 10.0.0.1 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@700 -- # echo tcp 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@701 -- # echo 4420 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@702 -- # echo ipv4 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@705 -- # ln -s /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn /sys/kernel/config/nvmet/ports/1/subsystems/ 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@708 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 --hostid=5b23e107-7094-e311-b1cb-001e67a97d55 -a 10.0.0.1 -t tcp -s 4420 00:45:12.852 00:45:12.852 Discovery Log Number of Records 2, Generation counter 2 00:45:12.852 =====Discovery Log Entry 0====== 00:45:12.852 trtype: tcp 00:45:12.852 adrfam: ipv4 00:45:12.852 subtype: current discovery subsystem 00:45:12.852 treq: not specified, sq flow control disable supported 00:45:12.852 portid: 1 00:45:12.852 trsvcid: 4420 00:45:12.852 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:45:12.852 traddr: 10.0.0.1 00:45:12.852 eflags: none 00:45:12.852 sectype: none 00:45:12.852 =====Discovery Log Entry 1====== 00:45:12.852 trtype: tcp 00:45:12.852 adrfam: ipv4 00:45:12.852 subtype: nvme subsystem 00:45:12.852 treq: not specified, sq flow control disable supported 00:45:12.852 portid: 1 00:45:12.852 trsvcid: 4420 00:45:12.852 subnqn: nqn.2016-06.io.spdk:testnqn 00:45:12.852 traddr: 10.0.0.1 00:45:12.852 eflags: none 00:45:12.852 sectype: none 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@66 -- # rabort tcp IPv4 10.0.0.1 4420 nqn.2016-06.io.spdk:testnqn 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@17 -- # local trtype=tcp 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@18 -- # local adrfam=IPv4 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@19 -- # local traddr=10.0.0.1 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@20 -- # local trsvcid=4420 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@21 -- # local subnqn=nqn.2016-06.io.spdk:testnqn 00:45:12.852 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@23 -- # local qds qd 00:45:12.853 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@24 -- # local target r 00:45:12.853 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@26 -- # qds=(4 24 64) 00:45:12.853 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:45:12.853 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@29 -- # target=trtype:tcp 00:45:12.853 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:45:12.853 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4' 00:45:12.853 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:45:12.853 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.1' 00:45:12.853 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:45:12.853 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420' 00:45:12.853 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:45:12.853 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:45:12.853 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:45:12.853 18:00:54 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 4 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:45:16.145 Initializing NVMe Controllers 00:45:16.145 Attached to NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2016-06.io.spdk:testnqn 00:45:16.145 Associating TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:45:16.145 Initialization complete. Launching workers. 00:45:16.145 NS: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 55991, failed: 0 00:45:16.145 CTRLR: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 55991, failed to submit 0 00:45:16.145 success 0, unsuccessful 55991, failed 0 00:45:16.145 18:00:57 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:45:16.145 18:00:57 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 24 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:45:19.427 Initializing NVMe Controllers 00:45:19.427 Attached to NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2016-06.io.spdk:testnqn 00:45:19.427 Associating TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:45:19.427 Initialization complete. Launching workers. 00:45:19.427 NS: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 99510, failed: 0 00:45:19.427 CTRLR: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 25090, failed to submit 74420 00:45:19.427 success 0, unsuccessful 25090, failed 0 00:45:19.427 18:01:00 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:45:19.427 18:01:00 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 64 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:45:22.731 Initializing NVMe Controllers 00:45:22.731 Attached to NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2016-06.io.spdk:testnqn 00:45:22.731 Associating TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:45:22.731 Initialization complete. Launching workers. 00:45:22.731 NS: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 98546, failed: 0 00:45:22.731 CTRLR: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 24634, failed to submit 73912 00:45:22.731 success 0, unsuccessful 24634, failed 0 00:45:22.731 18:01:04 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@67 -- # clean_kernel_target 00:45:22.731 18:01:04 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@712 -- # [[ -e /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn ]] 00:45:22.731 18:01:04 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@714 -- # echo 0 00:45:22.731 18:01:04 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@716 -- # rm -f /sys/kernel/config/nvmet/ports/1/subsystems/nqn.2016-06.io.spdk:testnqn 00:45:22.731 18:01:04 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@717 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:45:22.731 18:01:04 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@718 -- # rmdir /sys/kernel/config/nvmet/ports/1 00:45:22.731 18:01:04 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@719 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:45:22.731 18:01:04 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@721 -- # modules=(/sys/module/nvmet/holders/*) 00:45:22.731 18:01:04 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@723 -- # modprobe -r nvmet_tcp nvmet 00:45:22.731 18:01:04 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@726 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:45:23.670 0000:00:04.7 (8086 0e27): ioatdma -> vfio-pci 00:45:23.670 0000:00:04.6 (8086 0e26): ioatdma -> vfio-pci 00:45:23.670 0000:00:04.5 (8086 0e25): ioatdma -> vfio-pci 00:45:23.670 0000:00:04.4 (8086 0e24): ioatdma -> vfio-pci 00:45:23.670 0000:00:04.3 (8086 0e23): ioatdma -> vfio-pci 00:45:23.670 0000:00:04.2 (8086 0e22): ioatdma -> vfio-pci 00:45:23.670 0000:00:04.1 (8086 0e21): ioatdma -> vfio-pci 00:45:23.670 0000:00:04.0 (8086 0e20): ioatdma -> vfio-pci 00:45:23.670 0000:80:04.7 (8086 0e27): ioatdma -> vfio-pci 00:45:23.670 0000:80:04.6 (8086 0e26): ioatdma -> vfio-pci 00:45:23.670 0000:80:04.5 (8086 0e25): ioatdma -> vfio-pci 00:45:23.670 0000:80:04.4 (8086 0e24): ioatdma -> vfio-pci 00:45:23.670 0000:80:04.3 (8086 0e23): ioatdma -> vfio-pci 00:45:23.670 0000:80:04.2 (8086 0e22): ioatdma -> vfio-pci 00:45:23.670 0000:80:04.1 (8086 0e21): ioatdma -> vfio-pci 00:45:23.927 0000:80:04.0 (8086 0e20): ioatdma -> vfio-pci 00:45:24.860 0000:88:00.0 (8086 0a54): nvme -> vfio-pci 00:45:24.860 00:45:24.860 real 0m14.585s 00:45:24.860 user 0m6.796s 00:45:24.860 sys 0m3.274s 00:45:24.860 18:01:06 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@1130 -- # xtrace_disable 00:45:24.860 18:01:06 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@10 -- # set +x 00:45:24.860 ************************************ 00:45:24.860 END TEST kernel_target_abort 00:45:24.860 ************************************ 00:45:24.860 18:01:06 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:45:24.860 18:01:06 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@84 -- # nvmftestfini 00:45:24.860 18:01:06 nvmf_abort_qd_sizes -- nvmf/common.sh@516 -- # nvmfcleanup 00:45:24.860 18:01:06 nvmf_abort_qd_sizes -- nvmf/common.sh@121 -- # sync 00:45:24.860 18:01:06 nvmf_abort_qd_sizes -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:45:24.860 18:01:06 nvmf_abort_qd_sizes -- nvmf/common.sh@124 -- # set +e 00:45:24.860 18:01:06 nvmf_abort_qd_sizes -- nvmf/common.sh@125 -- # for i in {1..20} 00:45:24.860 18:01:06 nvmf_abort_qd_sizes -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:45:24.860 rmmod nvme_tcp 00:45:24.860 rmmod nvme_fabrics 00:45:24.860 rmmod nvme_keyring 00:45:24.860 18:01:06 nvmf_abort_qd_sizes -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:45:24.860 18:01:06 nvmf_abort_qd_sizes -- nvmf/common.sh@128 -- # set -e 00:45:24.860 18:01:06 nvmf_abort_qd_sizes -- nvmf/common.sh@129 -- # return 0 00:45:24.860 18:01:06 nvmf_abort_qd_sizes -- nvmf/common.sh@517 -- # '[' -n 486822 ']' 00:45:24.860 18:01:06 nvmf_abort_qd_sizes -- nvmf/common.sh@518 -- # killprocess 486822 00:45:24.860 18:01:06 nvmf_abort_qd_sizes -- common/autotest_common.sh@954 -- # '[' -z 486822 ']' 00:45:24.860 18:01:06 nvmf_abort_qd_sizes -- common/autotest_common.sh@958 -- # kill -0 486822 00:45:24.860 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 958: kill: (486822) - No such process 00:45:24.860 18:01:06 nvmf_abort_qd_sizes -- common/autotest_common.sh@981 -- # echo 'Process with pid 486822 is not found' 00:45:24.860 Process with pid 486822 is not found 00:45:24.860 18:01:06 nvmf_abort_qd_sizes -- nvmf/common.sh@520 -- # '[' iso == iso ']' 00:45:24.860 18:01:06 nvmf_abort_qd_sizes -- nvmf/common.sh@521 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:45:26.277 Waiting for block devices as requested 00:45:26.277 0000:88:00.0 (8086 0a54): vfio-pci -> nvme 00:45:26.277 0000:00:04.7 (8086 0e27): vfio-pci -> ioatdma 00:45:26.277 0000:00:04.6 (8086 0e26): vfio-pci -> ioatdma 00:45:26.535 0000:00:04.5 (8086 0e25): vfio-pci -> ioatdma 00:45:26.535 0000:00:04.4 (8086 0e24): vfio-pci -> ioatdma 00:45:26.535 0000:00:04.3 (8086 0e23): vfio-pci -> ioatdma 00:45:26.535 0000:00:04.2 (8086 0e22): vfio-pci -> ioatdma 00:45:26.535 0000:00:04.1 (8086 0e21): vfio-pci -> ioatdma 00:45:26.794 0000:00:04.0 (8086 0e20): vfio-pci -> ioatdma 00:45:26.794 0000:80:04.7 (8086 0e27): vfio-pci -> ioatdma 00:45:26.794 0000:80:04.6 (8086 0e26): vfio-pci -> ioatdma 00:45:26.794 0000:80:04.5 (8086 0e25): vfio-pci -> ioatdma 00:45:27.053 0000:80:04.4 (8086 0e24): vfio-pci -> ioatdma 00:45:27.053 0000:80:04.3 (8086 0e23): vfio-pci -> ioatdma 00:45:27.053 0000:80:04.2 (8086 0e22): vfio-pci -> ioatdma 00:45:27.053 0000:80:04.1 (8086 0e21): vfio-pci -> ioatdma 00:45:27.311 0000:80:04.0 (8086 0e20): vfio-pci -> ioatdma 00:45:27.311 18:01:09 nvmf_abort_qd_sizes -- nvmf/common.sh@523 -- # [[ tcp == \t\c\p ]] 00:45:27.311 18:01:09 nvmf_abort_qd_sizes -- nvmf/common.sh@524 -- # nvmf_tcp_fini 00:45:27.311 18:01:09 nvmf_abort_qd_sizes -- nvmf/common.sh@297 -- # iptr 00:45:27.311 18:01:09 nvmf_abort_qd_sizes -- nvmf/common.sh@791 -- # iptables-save 00:45:27.311 18:01:09 nvmf_abort_qd_sizes -- nvmf/common.sh@791 -- # grep -v SPDK_NVMF 00:45:27.311 18:01:09 nvmf_abort_qd_sizes -- nvmf/common.sh@791 -- # iptables-restore 00:45:27.311 18:01:09 nvmf_abort_qd_sizes -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:45:27.311 18:01:09 nvmf_abort_qd_sizes -- nvmf/common.sh@302 -- # remove_spdk_ns 00:45:27.311 18:01:09 nvmf_abort_qd_sizes -- nvmf/common.sh@656 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:45:27.311 18:01:09 nvmf_abort_qd_sizes -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:45:27.311 18:01:09 nvmf_abort_qd_sizes -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:45:29.846 18:01:11 nvmf_abort_qd_sizes -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:45:29.846 00:45:29.846 real 0m38.810s 00:45:29.846 user 1m4.055s 00:45:29.846 sys 0m9.392s 00:45:29.846 18:01:11 nvmf_abort_qd_sizes -- common/autotest_common.sh@1130 -- # xtrace_disable 00:45:29.846 18:01:11 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:45:29.846 ************************************ 00:45:29.846 END TEST nvmf_abort_qd_sizes 00:45:29.847 ************************************ 00:45:29.847 18:01:11 -- spdk/autotest.sh@292 -- # run_test keyring_file /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/file.sh 00:45:29.847 18:01:11 -- common/autotest_common.sh@1105 -- # '[' 2 -le 1 ']' 00:45:29.847 18:01:11 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:45:29.847 18:01:11 -- common/autotest_common.sh@10 -- # set +x 00:45:29.847 ************************************ 00:45:29.847 START TEST keyring_file 00:45:29.847 ************************************ 00:45:29.847 18:01:11 keyring_file -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/file.sh 00:45:29.847 * Looking for test storage... 00:45:29.847 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring 00:45:29.847 18:01:11 keyring_file -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:45:29.847 18:01:11 keyring_file -- common/autotest_common.sh@1711 -- # lcov --version 00:45:29.847 18:01:11 keyring_file -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:45:29.847 18:01:11 keyring_file -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@333 -- # local ver1 ver1_l 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@334 -- # local ver2 ver2_l 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@336 -- # IFS=.-: 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@336 -- # read -ra ver1 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@337 -- # IFS=.-: 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@337 -- # read -ra ver2 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@338 -- # local 'op=<' 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@340 -- # ver1_l=2 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@341 -- # ver2_l=1 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@344 -- # case "$op" in 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@345 -- # : 1 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@364 -- # (( v = 0 )) 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@365 -- # decimal 1 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@353 -- # local d=1 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@355 -- # echo 1 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@365 -- # ver1[v]=1 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@366 -- # decimal 2 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@353 -- # local d=2 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@355 -- # echo 2 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@366 -- # ver2[v]=2 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@368 -- # return 0 00:45:29.847 18:01:11 keyring_file -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:45:29.847 18:01:11 keyring_file -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:45:29.847 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:45:29.847 --rc genhtml_branch_coverage=1 00:45:29.847 --rc genhtml_function_coverage=1 00:45:29.847 --rc genhtml_legend=1 00:45:29.847 --rc geninfo_all_blocks=1 00:45:29.847 --rc geninfo_unexecuted_blocks=1 00:45:29.847 00:45:29.847 ' 00:45:29.847 18:01:11 keyring_file -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:45:29.847 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:45:29.847 --rc genhtml_branch_coverage=1 00:45:29.847 --rc genhtml_function_coverage=1 00:45:29.847 --rc genhtml_legend=1 00:45:29.847 --rc geninfo_all_blocks=1 00:45:29.847 --rc geninfo_unexecuted_blocks=1 00:45:29.847 00:45:29.847 ' 00:45:29.847 18:01:11 keyring_file -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:45:29.847 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:45:29.847 --rc genhtml_branch_coverage=1 00:45:29.847 --rc genhtml_function_coverage=1 00:45:29.847 --rc genhtml_legend=1 00:45:29.847 --rc geninfo_all_blocks=1 00:45:29.847 --rc geninfo_unexecuted_blocks=1 00:45:29.847 00:45:29.847 ' 00:45:29.847 18:01:11 keyring_file -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:45:29.847 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:45:29.847 --rc genhtml_branch_coverage=1 00:45:29.847 --rc genhtml_function_coverage=1 00:45:29.847 --rc genhtml_legend=1 00:45:29.847 --rc geninfo_all_blocks=1 00:45:29.847 --rc geninfo_unexecuted_blocks=1 00:45:29.847 00:45:29.847 ' 00:45:29.847 18:01:11 keyring_file -- keyring/file.sh@11 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/common.sh 00:45:29.847 18:01:11 keyring_file -- keyring/common.sh@4 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:45:29.847 18:01:11 keyring_file -- nvmf/common.sh@7 -- # uname -s 00:45:29.847 18:01:11 keyring_file -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:45:29.847 18:01:11 keyring_file -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:45:29.847 18:01:11 keyring_file -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:45:29.847 18:01:11 keyring_file -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:45:29.847 18:01:11 keyring_file -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:45:29.847 18:01:11 keyring_file -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:45:29.847 18:01:11 keyring_file -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:45:29.847 18:01:11 keyring_file -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:45:29.847 18:01:11 keyring_file -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:45:29.847 18:01:11 keyring_file -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:45:29.847 18:01:11 keyring_file -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:45:29.847 18:01:11 keyring_file -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:45:29.847 18:01:11 keyring_file -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:45:29.847 18:01:11 keyring_file -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:45:29.847 18:01:11 keyring_file -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:45:29.847 18:01:11 keyring_file -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:45:29.847 18:01:11 keyring_file -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@15 -- # shopt -s extglob 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:45:29.847 18:01:11 keyring_file -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:45:29.847 18:01:11 keyring_file -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:45:29.847 18:01:11 keyring_file -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:45:29.847 18:01:11 keyring_file -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:45:29.847 18:01:11 keyring_file -- paths/export.sh@5 -- # export PATH 00:45:29.847 18:01:11 keyring_file -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:45:29.847 18:01:11 keyring_file -- nvmf/common.sh@51 -- # : 0 00:45:29.847 18:01:11 keyring_file -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:45:29.847 18:01:11 keyring_file -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:45:29.847 18:01:11 keyring_file -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:45:29.847 18:01:11 keyring_file -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:45:29.847 18:01:11 keyring_file -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:45:29.847 18:01:11 keyring_file -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:45:29.847 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:45:29.847 18:01:11 keyring_file -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:45:29.847 18:01:11 keyring_file -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:45:29.847 18:01:11 keyring_file -- nvmf/common.sh@55 -- # have_pci_nics=0 00:45:29.847 18:01:11 keyring_file -- keyring/common.sh@6 -- # bperfsock=/var/tmp/bperf.sock 00:45:29.847 18:01:11 keyring_file -- keyring/file.sh@13 -- # subnqn=nqn.2016-06.io.spdk:cnode0 00:45:29.847 18:01:11 keyring_file -- keyring/file.sh@14 -- # hostnqn=nqn.2016-06.io.spdk:host0 00:45:29.847 18:01:11 keyring_file -- keyring/file.sh@15 -- # key0=00112233445566778899aabbccddeeff 00:45:29.847 18:01:11 keyring_file -- keyring/file.sh@16 -- # key1=112233445566778899aabbccddeeff00 00:45:29.847 18:01:11 keyring_file -- keyring/file.sh@24 -- # trap cleanup EXIT 00:45:29.847 18:01:11 keyring_file -- keyring/file.sh@26 -- # prep_key key0 00112233445566778899aabbccddeeff 0 00:45:29.847 18:01:11 keyring_file -- keyring/common.sh@15 -- # local name key digest path 00:45:29.847 18:01:11 keyring_file -- keyring/common.sh@17 -- # name=key0 00:45:29.847 18:01:11 keyring_file -- keyring/common.sh@17 -- # key=00112233445566778899aabbccddeeff 00:45:29.847 18:01:11 keyring_file -- keyring/common.sh@17 -- # digest=0 00:45:29.847 18:01:11 keyring_file -- keyring/common.sh@18 -- # mktemp 00:45:29.847 18:01:11 keyring_file -- keyring/common.sh@18 -- # path=/tmp/tmp.L6jFXkAYod 00:45:29.847 18:01:11 keyring_file -- keyring/common.sh@20 -- # format_interchange_psk 00112233445566778899aabbccddeeff 0 00:45:29.848 18:01:11 keyring_file -- nvmf/common.sh@743 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff 0 00:45:29.848 18:01:11 keyring_file -- nvmf/common.sh@730 -- # local prefix key digest 00:45:29.848 18:01:11 keyring_file -- nvmf/common.sh@732 -- # prefix=NVMeTLSkey-1 00:45:29.848 18:01:11 keyring_file -- nvmf/common.sh@732 -- # key=00112233445566778899aabbccddeeff 00:45:29.848 18:01:11 keyring_file -- nvmf/common.sh@732 -- # digest=0 00:45:29.848 18:01:11 keyring_file -- nvmf/common.sh@733 -- # python - 00:45:29.848 18:01:11 keyring_file -- keyring/common.sh@21 -- # chmod 0600 /tmp/tmp.L6jFXkAYod 00:45:29.848 18:01:11 keyring_file -- keyring/common.sh@23 -- # echo /tmp/tmp.L6jFXkAYod 00:45:29.848 18:01:11 keyring_file -- keyring/file.sh@26 -- # key0path=/tmp/tmp.L6jFXkAYod 00:45:29.848 18:01:11 keyring_file -- keyring/file.sh@27 -- # prep_key key1 112233445566778899aabbccddeeff00 0 00:45:29.848 18:01:11 keyring_file -- keyring/common.sh@15 -- # local name key digest path 00:45:29.848 18:01:11 keyring_file -- keyring/common.sh@17 -- # name=key1 00:45:29.848 18:01:11 keyring_file -- keyring/common.sh@17 -- # key=112233445566778899aabbccddeeff00 00:45:29.848 18:01:11 keyring_file -- keyring/common.sh@17 -- # digest=0 00:45:29.848 18:01:11 keyring_file -- keyring/common.sh@18 -- # mktemp 00:45:29.848 18:01:11 keyring_file -- keyring/common.sh@18 -- # path=/tmp/tmp.CtFJyqdCdF 00:45:29.848 18:01:11 keyring_file -- keyring/common.sh@20 -- # format_interchange_psk 112233445566778899aabbccddeeff00 0 00:45:29.848 18:01:11 keyring_file -- nvmf/common.sh@743 -- # format_key NVMeTLSkey-1 112233445566778899aabbccddeeff00 0 00:45:29.848 18:01:11 keyring_file -- nvmf/common.sh@730 -- # local prefix key digest 00:45:29.848 18:01:11 keyring_file -- nvmf/common.sh@732 -- # prefix=NVMeTLSkey-1 00:45:29.848 18:01:11 keyring_file -- nvmf/common.sh@732 -- # key=112233445566778899aabbccddeeff00 00:45:29.848 18:01:11 keyring_file -- nvmf/common.sh@732 -- # digest=0 00:45:29.848 18:01:11 keyring_file -- nvmf/common.sh@733 -- # python - 00:45:29.848 18:01:11 keyring_file -- keyring/common.sh@21 -- # chmod 0600 /tmp/tmp.CtFJyqdCdF 00:45:29.848 18:01:11 keyring_file -- keyring/common.sh@23 -- # echo /tmp/tmp.CtFJyqdCdF 00:45:29.848 18:01:11 keyring_file -- keyring/file.sh@27 -- # key1path=/tmp/tmp.CtFJyqdCdF 00:45:29.848 18:01:11 keyring_file -- keyring/file.sh@30 -- # tgtpid=492595 00:45:29.848 18:01:11 keyring_file -- keyring/file.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:45:29.848 18:01:11 keyring_file -- keyring/file.sh@32 -- # waitforlisten 492595 00:45:29.848 18:01:11 keyring_file -- common/autotest_common.sh@835 -- # '[' -z 492595 ']' 00:45:29.848 18:01:11 keyring_file -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:45:29.848 18:01:11 keyring_file -- common/autotest_common.sh@840 -- # local max_retries=100 00:45:29.848 18:01:11 keyring_file -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:45:29.848 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:45:29.848 18:01:11 keyring_file -- common/autotest_common.sh@844 -- # xtrace_disable 00:45:29.848 18:01:11 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:45:29.848 [2024-12-06 18:01:11.430416] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:45:29.848 [2024-12-06 18:01:11.430500] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid492595 ] 00:45:29.848 [2024-12-06 18:01:11.498041] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:45:29.848 [2024-12-06 18:01:11.544752] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:45:30.106 18:01:11 keyring_file -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:45:30.106 18:01:11 keyring_file -- common/autotest_common.sh@868 -- # return 0 00:45:30.106 18:01:11 keyring_file -- keyring/file.sh@33 -- # rpc_cmd 00:45:30.106 18:01:11 keyring_file -- common/autotest_common.sh@563 -- # xtrace_disable 00:45:30.106 18:01:11 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:45:30.106 [2024-12-06 18:01:11.783140] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:45:30.106 null0 00:45:30.106 [2024-12-06 18:01:11.815176] tcp.c:1049:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:45:30.106 [2024-12-06 18:01:11.815638] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4420 *** 00:45:30.106 18:01:11 keyring_file -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:45:30.106 18:01:11 keyring_file -- keyring/file.sh@44 -- # NOT rpc_cmd nvmf_subsystem_add_listener -t tcp -a 127.0.0.1 -s 4420 nqn.2016-06.io.spdk:cnode0 00:45:30.106 18:01:11 keyring_file -- common/autotest_common.sh@652 -- # local es=0 00:45:30.106 18:01:11 keyring_file -- common/autotest_common.sh@654 -- # valid_exec_arg rpc_cmd nvmf_subsystem_add_listener -t tcp -a 127.0.0.1 -s 4420 nqn.2016-06.io.spdk:cnode0 00:45:30.106 18:01:11 keyring_file -- common/autotest_common.sh@640 -- # local arg=rpc_cmd 00:45:30.106 18:01:11 keyring_file -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:45:30.106 18:01:11 keyring_file -- common/autotest_common.sh@644 -- # type -t rpc_cmd 00:45:30.106 18:01:11 keyring_file -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:45:30.106 18:01:11 keyring_file -- common/autotest_common.sh@655 -- # rpc_cmd nvmf_subsystem_add_listener -t tcp -a 127.0.0.1 -s 4420 nqn.2016-06.io.spdk:cnode0 00:45:30.106 18:01:11 keyring_file -- common/autotest_common.sh@563 -- # xtrace_disable 00:45:30.106 18:01:11 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:45:30.106 [2024-12-06 18:01:11.843237] nvmf_rpc.c: 762:nvmf_rpc_listen_paused: *ERROR*: Listener already exists 00:45:30.106 request: 00:45:30.106 { 00:45:30.106 "nqn": "nqn.2016-06.io.spdk:cnode0", 00:45:30.106 "secure_channel": false, 00:45:30.106 "listen_address": { 00:45:30.106 "trtype": "tcp", 00:45:30.106 "traddr": "127.0.0.1", 00:45:30.106 "trsvcid": "4420" 00:45:30.106 }, 00:45:30.106 "method": "nvmf_subsystem_add_listener", 00:45:30.106 "req_id": 1 00:45:30.106 } 00:45:30.106 Got JSON-RPC error response 00:45:30.106 response: 00:45:30.106 { 00:45:30.106 "code": -32602, 00:45:30.106 "message": "Invalid parameters" 00:45:30.106 } 00:45:30.106 18:01:11 keyring_file -- common/autotest_common.sh@591 -- # [[ 1 == 0 ]] 00:45:30.106 18:01:11 keyring_file -- common/autotest_common.sh@655 -- # es=1 00:45:30.106 18:01:11 keyring_file -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:45:30.106 18:01:11 keyring_file -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:45:30.106 18:01:11 keyring_file -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:45:30.106 18:01:11 keyring_file -- keyring/file.sh@47 -- # bperfpid=492604 00:45:30.106 18:01:11 keyring_file -- keyring/file.sh@49 -- # waitforlisten 492604 /var/tmp/bperf.sock 00:45:30.106 18:01:11 keyring_file -- keyring/file.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -q 128 -o 4k -w randrw -M 50 -t 1 -m 2 -r /var/tmp/bperf.sock -z 00:45:30.106 18:01:11 keyring_file -- common/autotest_common.sh@835 -- # '[' -z 492604 ']' 00:45:30.106 18:01:11 keyring_file -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bperf.sock 00:45:30.106 18:01:11 keyring_file -- common/autotest_common.sh@840 -- # local max_retries=100 00:45:30.106 18:01:11 keyring_file -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:45:30.107 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:45:30.107 18:01:11 keyring_file -- common/autotest_common.sh@844 -- # xtrace_disable 00:45:30.107 18:01:11 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:45:30.107 [2024-12-06 18:01:11.893803] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:45:30.107 [2024-12-06 18:01:11.893873] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid492604 ] 00:45:30.365 [2024-12-06 18:01:11.960613] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:45:30.365 [2024-12-06 18:01:12.006201] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:45:30.365 18:01:12 keyring_file -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:45:30.365 18:01:12 keyring_file -- common/autotest_common.sh@868 -- # return 0 00:45:30.365 18:01:12 keyring_file -- keyring/file.sh@50 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.L6jFXkAYod 00:45:30.365 18:01:12 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.L6jFXkAYod 00:45:30.623 18:01:12 keyring_file -- keyring/file.sh@51 -- # bperf_cmd keyring_file_add_key key1 /tmp/tmp.CtFJyqdCdF 00:45:30.623 18:01:12 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key1 /tmp/tmp.CtFJyqdCdF 00:45:30.881 18:01:12 keyring_file -- keyring/file.sh@52 -- # get_key key0 00:45:30.881 18:01:12 keyring_file -- keyring/file.sh@52 -- # jq -r .path 00:45:30.881 18:01:12 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:45:30.881 18:01:12 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:45:30.881 18:01:12 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:45:31.139 18:01:12 keyring_file -- keyring/file.sh@52 -- # [[ /tmp/tmp.L6jFXkAYod == \/\t\m\p\/\t\m\p\.\L\6\j\F\X\k\A\Y\o\d ]] 00:45:31.139 18:01:12 keyring_file -- keyring/file.sh@53 -- # get_key key1 00:45:31.139 18:01:12 keyring_file -- keyring/file.sh@53 -- # jq -r .path 00:45:31.139 18:01:12 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:45:31.139 18:01:12 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:45:31.139 18:01:12 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:45:31.398 18:01:13 keyring_file -- keyring/file.sh@53 -- # [[ /tmp/tmp.CtFJyqdCdF == \/\t\m\p\/\t\m\p\.\C\t\F\J\y\q\d\C\d\F ]] 00:45:31.398 18:01:13 keyring_file -- keyring/file.sh@54 -- # get_refcnt key0 00:45:31.398 18:01:13 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:45:31.398 18:01:13 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:45:31.398 18:01:13 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:45:31.398 18:01:13 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:45:31.398 18:01:13 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:45:31.655 18:01:13 keyring_file -- keyring/file.sh@54 -- # (( 1 == 1 )) 00:45:31.913 18:01:13 keyring_file -- keyring/file.sh@55 -- # get_refcnt key1 00:45:31.913 18:01:13 keyring_file -- keyring/common.sh@12 -- # get_key key1 00:45:31.913 18:01:13 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:45:31.913 18:01:13 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:45:31.913 18:01:13 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:45:31.913 18:01:13 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:45:32.172 18:01:13 keyring_file -- keyring/file.sh@55 -- # (( 1 == 1 )) 00:45:32.172 18:01:13 keyring_file -- keyring/file.sh@58 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:45:32.172 18:01:13 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:45:32.430 [2024-12-06 18:01:14.022821] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:45:32.430 nvme0n1 00:45:32.430 18:01:14 keyring_file -- keyring/file.sh@60 -- # get_refcnt key0 00:45:32.430 18:01:14 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:45:32.430 18:01:14 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:45:32.430 18:01:14 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:45:32.431 18:01:14 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:45:32.431 18:01:14 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:45:32.689 18:01:14 keyring_file -- keyring/file.sh@60 -- # (( 2 == 2 )) 00:45:32.689 18:01:14 keyring_file -- keyring/file.sh@61 -- # get_refcnt key1 00:45:32.689 18:01:14 keyring_file -- keyring/common.sh@12 -- # get_key key1 00:45:32.689 18:01:14 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:45:32.689 18:01:14 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:45:32.689 18:01:14 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:45:32.689 18:01:14 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:45:32.947 18:01:14 keyring_file -- keyring/file.sh@61 -- # (( 1 == 1 )) 00:45:32.947 18:01:14 keyring_file -- keyring/file.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:45:32.947 Running I/O for 1 seconds... 00:45:34.319 10364.00 IOPS, 40.48 MiB/s 00:45:34.319 Latency(us) 00:45:34.319 [2024-12-06T17:01:16.158Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:45:34.319 Job: nvme0n1 (Core Mask 0x2, workload: randrw, percentage: 50, depth: 128, IO size: 4096) 00:45:34.319 nvme0n1 : 1.01 10412.31 40.67 0.00 0.00 12252.45 4466.16 18447.17 00:45:34.319 [2024-12-06T17:01:16.158Z] =================================================================================================================== 00:45:34.319 [2024-12-06T17:01:16.158Z] Total : 10412.31 40.67 0.00 0.00 12252.45 4466.16 18447.17 00:45:34.319 { 00:45:34.319 "results": [ 00:45:34.319 { 00:45:34.319 "job": "nvme0n1", 00:45:34.319 "core_mask": "0x2", 00:45:34.319 "workload": "randrw", 00:45:34.319 "percentage": 50, 00:45:34.319 "status": "finished", 00:45:34.319 "queue_depth": 128, 00:45:34.319 "io_size": 4096, 00:45:34.319 "runtime": 1.007749, 00:45:34.319 "iops": 10412.314971287493, 00:45:34.319 "mibps": 40.67310535659177, 00:45:34.319 "io_failed": 0, 00:45:34.319 "io_timeout": 0, 00:45:34.319 "avg_latency_us": 12252.446423329839, 00:45:34.319 "min_latency_us": 4466.157037037037, 00:45:34.319 "max_latency_us": 18447.17037037037 00:45:34.319 } 00:45:34.319 ], 00:45:34.319 "core_count": 1 00:45:34.319 } 00:45:34.319 18:01:15 keyring_file -- keyring/file.sh@65 -- # bperf_cmd bdev_nvme_detach_controller nvme0 00:45:34.319 18:01:15 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_detach_controller nvme0 00:45:34.319 18:01:16 keyring_file -- keyring/file.sh@66 -- # get_refcnt key0 00:45:34.319 18:01:16 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:45:34.319 18:01:16 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:45:34.319 18:01:16 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:45:34.319 18:01:16 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:45:34.319 18:01:16 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:45:34.577 18:01:16 keyring_file -- keyring/file.sh@66 -- # (( 1 == 1 )) 00:45:34.577 18:01:16 keyring_file -- keyring/file.sh@67 -- # get_refcnt key1 00:45:34.577 18:01:16 keyring_file -- keyring/common.sh@12 -- # get_key key1 00:45:34.577 18:01:16 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:45:34.577 18:01:16 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:45:34.577 18:01:16 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:45:34.577 18:01:16 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:45:34.834 18:01:16 keyring_file -- keyring/file.sh@67 -- # (( 1 == 1 )) 00:45:34.834 18:01:16 keyring_file -- keyring/file.sh@70 -- # NOT bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key1 00:45:34.834 18:01:16 keyring_file -- common/autotest_common.sh@652 -- # local es=0 00:45:34.834 18:01:16 keyring_file -- common/autotest_common.sh@654 -- # valid_exec_arg bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key1 00:45:34.834 18:01:16 keyring_file -- common/autotest_common.sh@640 -- # local arg=bperf_cmd 00:45:34.834 18:01:16 keyring_file -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:45:34.834 18:01:16 keyring_file -- common/autotest_common.sh@644 -- # type -t bperf_cmd 00:45:34.834 18:01:16 keyring_file -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:45:34.834 18:01:16 keyring_file -- common/autotest_common.sh@655 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key1 00:45:34.834 18:01:16 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key1 00:45:35.092 [2024-12-06 18:01:16.873552] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 421:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:45:35.092 [2024-12-06 18:01:16.874008] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1407c20 (107): Transport endpoint is not connected 00:45:35.092 [2024-12-06 18:01:16.875000] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1407c20 (9): Bad file descriptor 00:45:35.092 [2024-12-06 18:01:16.875998] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 0] Ctrlr is in error state 00:45:35.092 [2024-12-06 18:01:16.876017] nvme.c: 709:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 127.0.0.1 00:45:35.092 [2024-12-06 18:01:16.876045] nvme.c: 895:nvme_dummy_attach_fail_cb: *ERROR*: Failed to attach nvme ctrlr: trtype=TCP adrfam=IPv4 traddr=127.0.0.1 trsvcid=4420 subnqn=nqn.2016-06.io.spdk:cnode0, Operation not permitted 00:45:35.092 [2024-12-06 18:01:16.876059] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 0] in failed state. 00:45:35.092 request: 00:45:35.092 { 00:45:35.092 "name": "nvme0", 00:45:35.092 "trtype": "tcp", 00:45:35.092 "traddr": "127.0.0.1", 00:45:35.092 "adrfam": "ipv4", 00:45:35.092 "trsvcid": "4420", 00:45:35.092 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:45:35.092 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:45:35.092 "prchk_reftag": false, 00:45:35.092 "prchk_guard": false, 00:45:35.092 "hdgst": false, 00:45:35.092 "ddgst": false, 00:45:35.092 "psk": "key1", 00:45:35.092 "allow_unrecognized_csi": false, 00:45:35.092 "method": "bdev_nvme_attach_controller", 00:45:35.092 "req_id": 1 00:45:35.092 } 00:45:35.092 Got JSON-RPC error response 00:45:35.092 response: 00:45:35.092 { 00:45:35.092 "code": -5, 00:45:35.092 "message": "Input/output error" 00:45:35.092 } 00:45:35.092 18:01:16 keyring_file -- common/autotest_common.sh@655 -- # es=1 00:45:35.092 18:01:16 keyring_file -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:45:35.092 18:01:16 keyring_file -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:45:35.092 18:01:16 keyring_file -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:45:35.092 18:01:16 keyring_file -- keyring/file.sh@72 -- # get_refcnt key0 00:45:35.092 18:01:16 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:45:35.092 18:01:16 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:45:35.092 18:01:16 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:45:35.092 18:01:16 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:45:35.092 18:01:16 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:45:35.348 18:01:17 keyring_file -- keyring/file.sh@72 -- # (( 1 == 1 )) 00:45:35.348 18:01:17 keyring_file -- keyring/file.sh@73 -- # get_refcnt key1 00:45:35.348 18:01:17 keyring_file -- keyring/common.sh@12 -- # get_key key1 00:45:35.348 18:01:17 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:45:35.348 18:01:17 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:45:35.348 18:01:17 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:45:35.349 18:01:17 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:45:35.606 18:01:17 keyring_file -- keyring/file.sh@73 -- # (( 1 == 1 )) 00:45:35.606 18:01:17 keyring_file -- keyring/file.sh@76 -- # bperf_cmd keyring_file_remove_key key0 00:45:35.606 18:01:17 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_remove_key key0 00:45:35.863 18:01:17 keyring_file -- keyring/file.sh@77 -- # bperf_cmd keyring_file_remove_key key1 00:45:35.863 18:01:17 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_remove_key key1 00:45:36.429 18:01:17 keyring_file -- keyring/file.sh@78 -- # bperf_cmd keyring_get_keys 00:45:36.429 18:01:17 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:45:36.429 18:01:17 keyring_file -- keyring/file.sh@78 -- # jq length 00:45:36.429 18:01:18 keyring_file -- keyring/file.sh@78 -- # (( 0 == 0 )) 00:45:36.429 18:01:18 keyring_file -- keyring/file.sh@81 -- # chmod 0660 /tmp/tmp.L6jFXkAYod 00:45:36.429 18:01:18 keyring_file -- keyring/file.sh@82 -- # NOT bperf_cmd keyring_file_add_key key0 /tmp/tmp.L6jFXkAYod 00:45:36.429 18:01:18 keyring_file -- common/autotest_common.sh@652 -- # local es=0 00:45:36.429 18:01:18 keyring_file -- common/autotest_common.sh@654 -- # valid_exec_arg bperf_cmd keyring_file_add_key key0 /tmp/tmp.L6jFXkAYod 00:45:36.429 18:01:18 keyring_file -- common/autotest_common.sh@640 -- # local arg=bperf_cmd 00:45:36.429 18:01:18 keyring_file -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:45:36.429 18:01:18 keyring_file -- common/autotest_common.sh@644 -- # type -t bperf_cmd 00:45:36.429 18:01:18 keyring_file -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:45:36.429 18:01:18 keyring_file -- common/autotest_common.sh@655 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.L6jFXkAYod 00:45:36.429 18:01:18 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.L6jFXkAYod 00:45:36.996 [2024-12-06 18:01:18.529109] keyring.c: 36:keyring_file_check_path: *ERROR*: Invalid permissions for key file '/tmp/tmp.L6jFXkAYod': 0100660 00:45:36.996 [2024-12-06 18:01:18.529152] keyring.c: 126:spdk_keyring_add_key: *ERROR*: Failed to add key 'key0' to the keyring 00:45:36.996 request: 00:45:36.996 { 00:45:36.996 "name": "key0", 00:45:36.996 "path": "/tmp/tmp.L6jFXkAYod", 00:45:36.996 "method": "keyring_file_add_key", 00:45:36.996 "req_id": 1 00:45:36.996 } 00:45:36.996 Got JSON-RPC error response 00:45:36.996 response: 00:45:36.996 { 00:45:36.996 "code": -1, 00:45:36.996 "message": "Operation not permitted" 00:45:36.996 } 00:45:36.996 18:01:18 keyring_file -- common/autotest_common.sh@655 -- # es=1 00:45:36.996 18:01:18 keyring_file -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:45:36.996 18:01:18 keyring_file -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:45:36.996 18:01:18 keyring_file -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:45:36.996 18:01:18 keyring_file -- keyring/file.sh@85 -- # chmod 0600 /tmp/tmp.L6jFXkAYod 00:45:36.996 18:01:18 keyring_file -- keyring/file.sh@86 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.L6jFXkAYod 00:45:36.996 18:01:18 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.L6jFXkAYod 00:45:36.996 18:01:18 keyring_file -- keyring/file.sh@87 -- # rm -f /tmp/tmp.L6jFXkAYod 00:45:36.996 18:01:18 keyring_file -- keyring/file.sh@89 -- # get_refcnt key0 00:45:36.996 18:01:18 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:45:36.996 18:01:18 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:45:36.996 18:01:18 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:45:36.996 18:01:18 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:45:36.996 18:01:18 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:45:37.562 18:01:19 keyring_file -- keyring/file.sh@89 -- # (( 1 == 1 )) 00:45:37.562 18:01:19 keyring_file -- keyring/file.sh@91 -- # NOT bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:45:37.562 18:01:19 keyring_file -- common/autotest_common.sh@652 -- # local es=0 00:45:37.562 18:01:19 keyring_file -- common/autotest_common.sh@654 -- # valid_exec_arg bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:45:37.562 18:01:19 keyring_file -- common/autotest_common.sh@640 -- # local arg=bperf_cmd 00:45:37.562 18:01:19 keyring_file -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:45:37.562 18:01:19 keyring_file -- common/autotest_common.sh@644 -- # type -t bperf_cmd 00:45:37.562 18:01:19 keyring_file -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:45:37.562 18:01:19 keyring_file -- common/autotest_common.sh@655 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:45:37.562 18:01:19 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:45:37.562 [2024-12-06 18:01:19.339299] keyring.c: 31:keyring_file_check_path: *ERROR*: Could not stat key file '/tmp/tmp.L6jFXkAYod': No such file or directory 00:45:37.562 [2024-12-06 18:01:19.339338] nvme_tcp.c:2498:nvme_tcp_generate_tls_credentials: *ERROR*: Failed to obtain key 'key0': No such file or directory 00:45:37.562 [2024-12-06 18:01:19.339379] nvme.c: 682:nvme_ctrlr_probe: *ERROR*: Failed to construct NVMe controller for SSD: 127.0.0.1 00:45:37.562 [2024-12-06 18:01:19.339392] nvme.c: 895:nvme_dummy_attach_fail_cb: *ERROR*: Failed to attach nvme ctrlr: trtype=TCP adrfam=IPv4 traddr=127.0.0.1 trsvcid=4420 subnqn=nqn.2016-06.io.spdk:cnode0, No such device 00:45:37.562 [2024-12-06 18:01:19.339404] nvme.c: 842:nvme_probe_internal: *ERROR*: NVMe ctrlr scan failed 00:45:37.562 [2024-12-06 18:01:19.339415] bdev_nvme.c:6796:spdk_bdev_nvme_create: *ERROR*: No controller was found with provided trid (traddr: 127.0.0.1) 00:45:37.562 request: 00:45:37.562 { 00:45:37.562 "name": "nvme0", 00:45:37.562 "trtype": "tcp", 00:45:37.562 "traddr": "127.0.0.1", 00:45:37.562 "adrfam": "ipv4", 00:45:37.562 "trsvcid": "4420", 00:45:37.562 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:45:37.562 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:45:37.562 "prchk_reftag": false, 00:45:37.562 "prchk_guard": false, 00:45:37.562 "hdgst": false, 00:45:37.562 "ddgst": false, 00:45:37.562 "psk": "key0", 00:45:37.562 "allow_unrecognized_csi": false, 00:45:37.562 "method": "bdev_nvme_attach_controller", 00:45:37.562 "req_id": 1 00:45:37.562 } 00:45:37.562 Got JSON-RPC error response 00:45:37.562 response: 00:45:37.562 { 00:45:37.562 "code": -19, 00:45:37.562 "message": "No such device" 00:45:37.562 } 00:45:37.562 18:01:19 keyring_file -- common/autotest_common.sh@655 -- # es=1 00:45:37.562 18:01:19 keyring_file -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:45:37.562 18:01:19 keyring_file -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:45:37.562 18:01:19 keyring_file -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:45:37.562 18:01:19 keyring_file -- keyring/file.sh@93 -- # bperf_cmd keyring_file_remove_key key0 00:45:37.562 18:01:19 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_remove_key key0 00:45:37.821 18:01:19 keyring_file -- keyring/file.sh@96 -- # prep_key key0 00112233445566778899aabbccddeeff 0 00:45:37.821 18:01:19 keyring_file -- keyring/common.sh@15 -- # local name key digest path 00:45:37.821 18:01:19 keyring_file -- keyring/common.sh@17 -- # name=key0 00:45:37.821 18:01:19 keyring_file -- keyring/common.sh@17 -- # key=00112233445566778899aabbccddeeff 00:45:37.821 18:01:19 keyring_file -- keyring/common.sh@17 -- # digest=0 00:45:37.821 18:01:19 keyring_file -- keyring/common.sh@18 -- # mktemp 00:45:37.821 18:01:19 keyring_file -- keyring/common.sh@18 -- # path=/tmp/tmp.zvWtEHZmQ0 00:45:37.821 18:01:19 keyring_file -- keyring/common.sh@20 -- # format_interchange_psk 00112233445566778899aabbccddeeff 0 00:45:37.821 18:01:19 keyring_file -- nvmf/common.sh@743 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff 0 00:45:37.821 18:01:19 keyring_file -- nvmf/common.sh@730 -- # local prefix key digest 00:45:37.821 18:01:19 keyring_file -- nvmf/common.sh@732 -- # prefix=NVMeTLSkey-1 00:45:37.821 18:01:19 keyring_file -- nvmf/common.sh@732 -- # key=00112233445566778899aabbccddeeff 00:45:37.821 18:01:19 keyring_file -- nvmf/common.sh@732 -- # digest=0 00:45:37.821 18:01:19 keyring_file -- nvmf/common.sh@733 -- # python - 00:45:38.079 18:01:19 keyring_file -- keyring/common.sh@21 -- # chmod 0600 /tmp/tmp.zvWtEHZmQ0 00:45:38.079 18:01:19 keyring_file -- keyring/common.sh@23 -- # echo /tmp/tmp.zvWtEHZmQ0 00:45:38.079 18:01:19 keyring_file -- keyring/file.sh@96 -- # key0path=/tmp/tmp.zvWtEHZmQ0 00:45:38.079 18:01:19 keyring_file -- keyring/file.sh@97 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.zvWtEHZmQ0 00:45:38.079 18:01:19 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.zvWtEHZmQ0 00:45:38.337 18:01:19 keyring_file -- keyring/file.sh@98 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:45:38.337 18:01:19 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:45:38.596 nvme0n1 00:45:38.596 18:01:20 keyring_file -- keyring/file.sh@100 -- # get_refcnt key0 00:45:38.596 18:01:20 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:45:38.596 18:01:20 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:45:38.596 18:01:20 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:45:38.596 18:01:20 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:45:38.596 18:01:20 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:45:38.854 18:01:20 keyring_file -- keyring/file.sh@100 -- # (( 2 == 2 )) 00:45:38.854 18:01:20 keyring_file -- keyring/file.sh@101 -- # bperf_cmd keyring_file_remove_key key0 00:45:38.854 18:01:20 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_remove_key key0 00:45:39.113 18:01:20 keyring_file -- keyring/file.sh@102 -- # get_key key0 00:45:39.113 18:01:20 keyring_file -- keyring/file.sh@102 -- # jq -r .removed 00:45:39.113 18:01:20 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:45:39.113 18:01:20 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:45:39.113 18:01:20 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:45:39.371 18:01:21 keyring_file -- keyring/file.sh@102 -- # [[ true == \t\r\u\e ]] 00:45:39.371 18:01:21 keyring_file -- keyring/file.sh@103 -- # get_refcnt key0 00:45:39.371 18:01:21 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:45:39.371 18:01:21 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:45:39.371 18:01:21 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:45:39.371 18:01:21 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:45:39.371 18:01:21 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:45:39.629 18:01:21 keyring_file -- keyring/file.sh@103 -- # (( 1 == 1 )) 00:45:39.629 18:01:21 keyring_file -- keyring/file.sh@104 -- # bperf_cmd bdev_nvme_detach_controller nvme0 00:45:39.629 18:01:21 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_detach_controller nvme0 00:45:39.887 18:01:21 keyring_file -- keyring/file.sh@105 -- # bperf_cmd keyring_get_keys 00:45:39.887 18:01:21 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:45:39.887 18:01:21 keyring_file -- keyring/file.sh@105 -- # jq length 00:45:40.145 18:01:21 keyring_file -- keyring/file.sh@105 -- # (( 0 == 0 )) 00:45:40.145 18:01:21 keyring_file -- keyring/file.sh@108 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.zvWtEHZmQ0 00:45:40.145 18:01:21 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.zvWtEHZmQ0 00:45:40.403 18:01:22 keyring_file -- keyring/file.sh@109 -- # bperf_cmd keyring_file_add_key key1 /tmp/tmp.CtFJyqdCdF 00:45:40.403 18:01:22 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key1 /tmp/tmp.CtFJyqdCdF 00:45:40.969 18:01:22 keyring_file -- keyring/file.sh@110 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:45:40.969 18:01:22 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:45:41.227 nvme0n1 00:45:41.227 18:01:22 keyring_file -- keyring/file.sh@113 -- # bperf_cmd save_config 00:45:41.227 18:01:22 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock save_config 00:45:41.485 18:01:23 keyring_file -- keyring/file.sh@113 -- # config='{ 00:45:41.485 "subsystems": [ 00:45:41.485 { 00:45:41.485 "subsystem": "keyring", 00:45:41.485 "config": [ 00:45:41.485 { 00:45:41.485 "method": "keyring_file_add_key", 00:45:41.485 "params": { 00:45:41.485 "name": "key0", 00:45:41.485 "path": "/tmp/tmp.zvWtEHZmQ0" 00:45:41.485 } 00:45:41.485 }, 00:45:41.485 { 00:45:41.485 "method": "keyring_file_add_key", 00:45:41.485 "params": { 00:45:41.485 "name": "key1", 00:45:41.485 "path": "/tmp/tmp.CtFJyqdCdF" 00:45:41.485 } 00:45:41.485 } 00:45:41.485 ] 00:45:41.485 }, 00:45:41.485 { 00:45:41.485 "subsystem": "iobuf", 00:45:41.485 "config": [ 00:45:41.485 { 00:45:41.485 "method": "iobuf_set_options", 00:45:41.485 "params": { 00:45:41.485 "small_pool_count": 8192, 00:45:41.485 "large_pool_count": 1024, 00:45:41.485 "small_bufsize": 8192, 00:45:41.485 "large_bufsize": 135168, 00:45:41.485 "enable_numa": false 00:45:41.485 } 00:45:41.485 } 00:45:41.485 ] 00:45:41.485 }, 00:45:41.485 { 00:45:41.485 "subsystem": "sock", 00:45:41.485 "config": [ 00:45:41.485 { 00:45:41.485 "method": "sock_set_default_impl", 00:45:41.485 "params": { 00:45:41.485 "impl_name": "posix" 00:45:41.485 } 00:45:41.485 }, 00:45:41.486 { 00:45:41.486 "method": "sock_impl_set_options", 00:45:41.486 "params": { 00:45:41.486 "impl_name": "ssl", 00:45:41.486 "recv_buf_size": 4096, 00:45:41.486 "send_buf_size": 4096, 00:45:41.486 "enable_recv_pipe": true, 00:45:41.486 "enable_quickack": false, 00:45:41.486 "enable_placement_id": 0, 00:45:41.486 "enable_zerocopy_send_server": true, 00:45:41.486 "enable_zerocopy_send_client": false, 00:45:41.486 "zerocopy_threshold": 0, 00:45:41.486 "tls_version": 0, 00:45:41.486 "enable_ktls": false 00:45:41.486 } 00:45:41.486 }, 00:45:41.486 { 00:45:41.486 "method": "sock_impl_set_options", 00:45:41.486 "params": { 00:45:41.486 "impl_name": "posix", 00:45:41.486 "recv_buf_size": 2097152, 00:45:41.486 "send_buf_size": 2097152, 00:45:41.486 "enable_recv_pipe": true, 00:45:41.486 "enable_quickack": false, 00:45:41.486 "enable_placement_id": 0, 00:45:41.486 "enable_zerocopy_send_server": true, 00:45:41.486 "enable_zerocopy_send_client": false, 00:45:41.486 "zerocopy_threshold": 0, 00:45:41.486 "tls_version": 0, 00:45:41.486 "enable_ktls": false 00:45:41.486 } 00:45:41.486 } 00:45:41.486 ] 00:45:41.486 }, 00:45:41.486 { 00:45:41.486 "subsystem": "vmd", 00:45:41.486 "config": [] 00:45:41.486 }, 00:45:41.486 { 00:45:41.486 "subsystem": "accel", 00:45:41.486 "config": [ 00:45:41.486 { 00:45:41.486 "method": "accel_set_options", 00:45:41.486 "params": { 00:45:41.486 "small_cache_size": 128, 00:45:41.486 "large_cache_size": 16, 00:45:41.486 "task_count": 2048, 00:45:41.486 "sequence_count": 2048, 00:45:41.486 "buf_count": 2048 00:45:41.486 } 00:45:41.486 } 00:45:41.486 ] 00:45:41.486 }, 00:45:41.486 { 00:45:41.486 "subsystem": "bdev", 00:45:41.486 "config": [ 00:45:41.486 { 00:45:41.486 "method": "bdev_set_options", 00:45:41.486 "params": { 00:45:41.486 "bdev_io_pool_size": 65535, 00:45:41.486 "bdev_io_cache_size": 256, 00:45:41.486 "bdev_auto_examine": true, 00:45:41.486 "iobuf_small_cache_size": 128, 00:45:41.486 "iobuf_large_cache_size": 16 00:45:41.486 } 00:45:41.486 }, 00:45:41.486 { 00:45:41.486 "method": "bdev_raid_set_options", 00:45:41.486 "params": { 00:45:41.486 "process_window_size_kb": 1024, 00:45:41.486 "process_max_bandwidth_mb_sec": 0 00:45:41.486 } 00:45:41.486 }, 00:45:41.486 { 00:45:41.486 "method": "bdev_iscsi_set_options", 00:45:41.486 "params": { 00:45:41.486 "timeout_sec": 30 00:45:41.486 } 00:45:41.486 }, 00:45:41.486 { 00:45:41.486 "method": "bdev_nvme_set_options", 00:45:41.486 "params": { 00:45:41.486 "action_on_timeout": "none", 00:45:41.486 "timeout_us": 0, 00:45:41.486 "timeout_admin_us": 0, 00:45:41.486 "keep_alive_timeout_ms": 10000, 00:45:41.486 "arbitration_burst": 0, 00:45:41.486 "low_priority_weight": 0, 00:45:41.486 "medium_priority_weight": 0, 00:45:41.486 "high_priority_weight": 0, 00:45:41.486 "nvme_adminq_poll_period_us": 10000, 00:45:41.486 "nvme_ioq_poll_period_us": 0, 00:45:41.486 "io_queue_requests": 512, 00:45:41.486 "delay_cmd_submit": true, 00:45:41.486 "transport_retry_count": 4, 00:45:41.486 "bdev_retry_count": 3, 00:45:41.486 "transport_ack_timeout": 0, 00:45:41.486 "ctrlr_loss_timeout_sec": 0, 00:45:41.486 "reconnect_delay_sec": 0, 00:45:41.486 "fast_io_fail_timeout_sec": 0, 00:45:41.486 "disable_auto_failback": false, 00:45:41.486 "generate_uuids": false, 00:45:41.486 "transport_tos": 0, 00:45:41.486 "nvme_error_stat": false, 00:45:41.486 "rdma_srq_size": 0, 00:45:41.486 "io_path_stat": false, 00:45:41.486 "allow_accel_sequence": false, 00:45:41.486 "rdma_max_cq_size": 0, 00:45:41.486 "rdma_cm_event_timeout_ms": 0, 00:45:41.486 "dhchap_digests": [ 00:45:41.486 "sha256", 00:45:41.486 "sha384", 00:45:41.486 "sha512" 00:45:41.486 ], 00:45:41.486 "dhchap_dhgroups": [ 00:45:41.486 "null", 00:45:41.486 "ffdhe2048", 00:45:41.486 "ffdhe3072", 00:45:41.486 "ffdhe4096", 00:45:41.486 "ffdhe6144", 00:45:41.486 "ffdhe8192" 00:45:41.486 ] 00:45:41.486 } 00:45:41.486 }, 00:45:41.486 { 00:45:41.486 "method": "bdev_nvme_attach_controller", 00:45:41.486 "params": { 00:45:41.486 "name": "nvme0", 00:45:41.486 "trtype": "TCP", 00:45:41.486 "adrfam": "IPv4", 00:45:41.486 "traddr": "127.0.0.1", 00:45:41.486 "trsvcid": "4420", 00:45:41.486 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:45:41.486 "prchk_reftag": false, 00:45:41.486 "prchk_guard": false, 00:45:41.486 "ctrlr_loss_timeout_sec": 0, 00:45:41.486 "reconnect_delay_sec": 0, 00:45:41.486 "fast_io_fail_timeout_sec": 0, 00:45:41.486 "psk": "key0", 00:45:41.486 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:45:41.486 "hdgst": false, 00:45:41.486 "ddgst": false, 00:45:41.486 "multipath": "multipath" 00:45:41.486 } 00:45:41.486 }, 00:45:41.486 { 00:45:41.486 "method": "bdev_nvme_set_hotplug", 00:45:41.486 "params": { 00:45:41.486 "period_us": 100000, 00:45:41.486 "enable": false 00:45:41.486 } 00:45:41.486 }, 00:45:41.486 { 00:45:41.486 "method": "bdev_wait_for_examine" 00:45:41.486 } 00:45:41.486 ] 00:45:41.486 }, 00:45:41.486 { 00:45:41.486 "subsystem": "nbd", 00:45:41.486 "config": [] 00:45:41.486 } 00:45:41.486 ] 00:45:41.486 }' 00:45:41.486 18:01:23 keyring_file -- keyring/file.sh@115 -- # killprocess 492604 00:45:41.486 18:01:23 keyring_file -- common/autotest_common.sh@954 -- # '[' -z 492604 ']' 00:45:41.486 18:01:23 keyring_file -- common/autotest_common.sh@958 -- # kill -0 492604 00:45:41.486 18:01:23 keyring_file -- common/autotest_common.sh@959 -- # uname 00:45:41.486 18:01:23 keyring_file -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:45:41.486 18:01:23 keyring_file -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 492604 00:45:41.486 18:01:23 keyring_file -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:45:41.486 18:01:23 keyring_file -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:45:41.486 18:01:23 keyring_file -- common/autotest_common.sh@972 -- # echo 'killing process with pid 492604' 00:45:41.486 killing process with pid 492604 00:45:41.486 18:01:23 keyring_file -- common/autotest_common.sh@973 -- # kill 492604 00:45:41.486 Received shutdown signal, test time was about 1.000000 seconds 00:45:41.486 00:45:41.486 Latency(us) 00:45:41.486 [2024-12-06T17:01:23.325Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:45:41.486 [2024-12-06T17:01:23.325Z] =================================================================================================================== 00:45:41.486 [2024-12-06T17:01:23.325Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:45:41.486 18:01:23 keyring_file -- common/autotest_common.sh@978 -- # wait 492604 00:45:41.745 18:01:23 keyring_file -- keyring/file.sh@118 -- # bperfpid=494070 00:45:41.745 18:01:23 keyring_file -- keyring/file.sh@120 -- # waitforlisten 494070 /var/tmp/bperf.sock 00:45:41.745 18:01:23 keyring_file -- common/autotest_common.sh@835 -- # '[' -z 494070 ']' 00:45:41.745 18:01:23 keyring_file -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bperf.sock 00:45:41.745 18:01:23 keyring_file -- keyring/file.sh@116 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -q 128 -o 4k -w randrw -M 50 -t 1 -m 2 -r /var/tmp/bperf.sock -z -c /dev/fd/63 00:45:41.745 18:01:23 keyring_file -- common/autotest_common.sh@840 -- # local max_retries=100 00:45:41.745 18:01:23 keyring_file -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:45:41.745 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:45:41.746 18:01:23 keyring_file -- common/autotest_common.sh@844 -- # xtrace_disable 00:45:41.746 18:01:23 keyring_file -- keyring/file.sh@116 -- # echo '{ 00:45:41.746 "subsystems": [ 00:45:41.746 { 00:45:41.746 "subsystem": "keyring", 00:45:41.746 "config": [ 00:45:41.746 { 00:45:41.746 "method": "keyring_file_add_key", 00:45:41.746 "params": { 00:45:41.746 "name": "key0", 00:45:41.746 "path": "/tmp/tmp.zvWtEHZmQ0" 00:45:41.746 } 00:45:41.746 }, 00:45:41.746 { 00:45:41.746 "method": "keyring_file_add_key", 00:45:41.746 "params": { 00:45:41.746 "name": "key1", 00:45:41.746 "path": "/tmp/tmp.CtFJyqdCdF" 00:45:41.746 } 00:45:41.746 } 00:45:41.746 ] 00:45:41.746 }, 00:45:41.746 { 00:45:41.746 "subsystem": "iobuf", 00:45:41.746 "config": [ 00:45:41.746 { 00:45:41.746 "method": "iobuf_set_options", 00:45:41.746 "params": { 00:45:41.746 "small_pool_count": 8192, 00:45:41.746 "large_pool_count": 1024, 00:45:41.746 "small_bufsize": 8192, 00:45:41.746 "large_bufsize": 135168, 00:45:41.746 "enable_numa": false 00:45:41.746 } 00:45:41.746 } 00:45:41.746 ] 00:45:41.746 }, 00:45:41.746 { 00:45:41.746 "subsystem": "sock", 00:45:41.746 "config": [ 00:45:41.746 { 00:45:41.746 "method": "sock_set_default_impl", 00:45:41.746 "params": { 00:45:41.746 "impl_name": "posix" 00:45:41.746 } 00:45:41.746 }, 00:45:41.746 { 00:45:41.746 "method": "sock_impl_set_options", 00:45:41.746 "params": { 00:45:41.746 "impl_name": "ssl", 00:45:41.746 "recv_buf_size": 4096, 00:45:41.746 "send_buf_size": 4096, 00:45:41.746 "enable_recv_pipe": true, 00:45:41.746 "enable_quickack": false, 00:45:41.746 "enable_placement_id": 0, 00:45:41.746 "enable_zerocopy_send_server": true, 00:45:41.746 "enable_zerocopy_send_client": false, 00:45:41.746 "zerocopy_threshold": 0, 00:45:41.746 "tls_version": 0, 00:45:41.746 "enable_ktls": false 00:45:41.746 } 00:45:41.746 }, 00:45:41.746 { 00:45:41.746 "method": "sock_impl_set_options", 00:45:41.746 "params": { 00:45:41.746 "impl_name": "posix", 00:45:41.746 "recv_buf_size": 2097152, 00:45:41.746 "send_buf_size": 2097152, 00:45:41.746 "enable_recv_pipe": true, 00:45:41.746 "enable_quickack": false, 00:45:41.746 "enable_placement_id": 0, 00:45:41.746 "enable_zerocopy_send_server": true, 00:45:41.746 "enable_zerocopy_send_client": false, 00:45:41.746 "zerocopy_threshold": 0, 00:45:41.746 "tls_version": 0, 00:45:41.746 "enable_ktls": false 00:45:41.746 } 00:45:41.746 } 00:45:41.746 ] 00:45:41.746 }, 00:45:41.746 { 00:45:41.746 "subsystem": "vmd", 00:45:41.746 "config": [] 00:45:41.746 }, 00:45:41.746 { 00:45:41.746 "subsystem": "accel", 00:45:41.746 "config": [ 00:45:41.746 { 00:45:41.746 "method": "accel_set_options", 00:45:41.746 "params": { 00:45:41.746 "small_cache_size": 128, 00:45:41.746 "large_cache_size": 16, 00:45:41.746 "task_count": 2048, 00:45:41.746 "sequence_count": 2048, 00:45:41.746 "buf_count": 2048 00:45:41.746 } 00:45:41.746 } 00:45:41.746 ] 00:45:41.746 }, 00:45:41.746 { 00:45:41.746 "subsystem": "bdev", 00:45:41.746 "config": [ 00:45:41.746 { 00:45:41.746 "method": "bdev_set_options", 00:45:41.746 "params": { 00:45:41.746 "bdev_io_pool_size": 65535, 00:45:41.746 "bdev_io_cache_size": 256, 00:45:41.746 "bdev_auto_examine": true, 00:45:41.746 "iobuf_small_cache_size": 128, 00:45:41.746 "iobuf_large_cache_size": 16 00:45:41.746 } 00:45:41.746 }, 00:45:41.746 { 00:45:41.746 "method": "bdev_raid_set_options", 00:45:41.746 "params": { 00:45:41.746 "process_window_size_kb": 1024, 00:45:41.746 "process_max_bandwidth_mb_sec": 0 00:45:41.746 } 00:45:41.746 }, 00:45:41.746 { 00:45:41.746 "method": "bdev_iscsi_set_options", 00:45:41.746 "params": { 00:45:41.746 "timeout_sec": 30 00:45:41.746 } 00:45:41.746 }, 00:45:41.746 { 00:45:41.746 "method": "bdev_nvme_set_options", 00:45:41.746 "params": { 00:45:41.746 "action_on_timeout": "none", 00:45:41.746 "timeout_us": 0, 00:45:41.746 "timeout_admin_us": 0, 00:45:41.746 "keep_alive_timeout_ms": 10000, 00:45:41.746 "arbitration_burst": 0, 00:45:41.746 "low_priority_weight": 0, 00:45:41.746 "medium_priority_weight": 0, 00:45:41.746 "high_priority_weight": 0, 00:45:41.746 "nvme_adminq_poll_period_us": 10000, 00:45:41.746 "nvme_ioq_poll_period_us": 0, 00:45:41.746 "io_queue_requests": 512, 00:45:41.746 "delay_cmd_submit": true, 00:45:41.746 "transport_retry_count": 4, 00:45:41.746 "bdev_retry_count": 3, 00:45:41.746 "transport_ack_timeout": 0, 00:45:41.746 "ctrlr_loss_timeout_sec": 0, 00:45:41.746 "reconnect_delay_sec": 0, 00:45:41.746 "fast_io_fail_timeout_sec": 0, 00:45:41.746 "disable_auto_failback": false, 00:45:41.746 "generate_uuids": false, 00:45:41.746 "transport_tos": 0, 00:45:41.746 "nvme_error_stat": false, 00:45:41.746 "rdma_srq_size": 0, 00:45:41.746 18:01:23 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:45:41.746 "io_path_stat": false, 00:45:41.746 "allow_accel_sequence": false, 00:45:41.746 "rdma_max_cq_size": 0, 00:45:41.746 "rdma_cm_event_timeout_ms": 0, 00:45:41.746 "dhchap_digests": [ 00:45:41.746 "sha256", 00:45:41.746 "sha384", 00:45:41.746 "sha512" 00:45:41.746 ], 00:45:41.746 "dhchap_dhgroups": [ 00:45:41.746 "null", 00:45:41.746 "ffdhe2048", 00:45:41.746 "ffdhe3072", 00:45:41.746 "ffdhe4096", 00:45:41.747 "ffdhe6144", 00:45:41.747 "ffdhe8192" 00:45:41.747 ] 00:45:41.747 } 00:45:41.747 }, 00:45:41.747 { 00:45:41.747 "method": "bdev_nvme_attach_controller", 00:45:41.747 "params": { 00:45:41.747 "name": "nvme0", 00:45:41.747 "trtype": "TCP", 00:45:41.747 "adrfam": "IPv4", 00:45:41.747 "traddr": "127.0.0.1", 00:45:41.747 "trsvcid": "4420", 00:45:41.747 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:45:41.747 "prchk_reftag": false, 00:45:41.747 "prchk_guard": false, 00:45:41.747 "ctrlr_loss_timeout_sec": 0, 00:45:41.747 "reconnect_delay_sec": 0, 00:45:41.747 "fast_io_fail_timeout_sec": 0, 00:45:41.747 "psk": "key0", 00:45:41.747 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:45:41.747 "hdgst": false, 00:45:41.747 "ddgst": false, 00:45:41.747 "multipath": "multipath" 00:45:41.747 } 00:45:41.747 }, 00:45:41.747 { 00:45:41.747 "method": "bdev_nvme_set_hotplug", 00:45:41.747 "params": { 00:45:41.747 "period_us": 100000, 00:45:41.747 "enable": false 00:45:41.747 } 00:45:41.747 }, 00:45:41.747 { 00:45:41.747 "method": "bdev_wait_for_examine" 00:45:41.747 } 00:45:41.747 ] 00:45:41.747 }, 00:45:41.747 { 00:45:41.747 "subsystem": "nbd", 00:45:41.747 "config": [] 00:45:41.747 } 00:45:41.747 ] 00:45:41.747 }' 00:45:41.747 [2024-12-06 18:01:23.429226] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:45:41.747 [2024-12-06 18:01:23.429307] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid494070 ] 00:45:41.747 [2024-12-06 18:01:23.494639] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:45:41.747 [2024-12-06 18:01:23.544137] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:45:42.005 [2024-12-06 18:01:23.722852] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:45:42.005 18:01:23 keyring_file -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:45:42.005 18:01:23 keyring_file -- common/autotest_common.sh@868 -- # return 0 00:45:42.005 18:01:23 keyring_file -- keyring/file.sh@121 -- # bperf_cmd keyring_get_keys 00:45:42.005 18:01:23 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:45:42.005 18:01:23 keyring_file -- keyring/file.sh@121 -- # jq length 00:45:42.572 18:01:24 keyring_file -- keyring/file.sh@121 -- # (( 2 == 2 )) 00:45:42.572 18:01:24 keyring_file -- keyring/file.sh@122 -- # get_refcnt key0 00:45:42.572 18:01:24 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:45:42.572 18:01:24 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:45:42.572 18:01:24 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:45:42.572 18:01:24 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:45:42.572 18:01:24 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:45:42.572 18:01:24 keyring_file -- keyring/file.sh@122 -- # (( 2 == 2 )) 00:45:42.572 18:01:24 keyring_file -- keyring/file.sh@123 -- # get_refcnt key1 00:45:42.572 18:01:24 keyring_file -- keyring/common.sh@12 -- # get_key key1 00:45:42.572 18:01:24 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:45:42.572 18:01:24 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:45:42.572 18:01:24 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:45:42.572 18:01:24 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:45:42.830 18:01:24 keyring_file -- keyring/file.sh@123 -- # (( 1 == 1 )) 00:45:42.830 18:01:24 keyring_file -- keyring/file.sh@124 -- # bperf_cmd bdev_nvme_get_controllers 00:45:42.830 18:01:24 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_get_controllers 00:45:42.830 18:01:24 keyring_file -- keyring/file.sh@124 -- # jq -r '.[].name' 00:45:43.088 18:01:24 keyring_file -- keyring/file.sh@124 -- # [[ nvme0 == nvme0 ]] 00:45:43.088 18:01:24 keyring_file -- keyring/file.sh@1 -- # cleanup 00:45:43.088 18:01:24 keyring_file -- keyring/file.sh@19 -- # rm -f /tmp/tmp.zvWtEHZmQ0 /tmp/tmp.CtFJyqdCdF 00:45:43.088 18:01:24 keyring_file -- keyring/file.sh@20 -- # killprocess 494070 00:45:43.088 18:01:24 keyring_file -- common/autotest_common.sh@954 -- # '[' -z 494070 ']' 00:45:43.088 18:01:24 keyring_file -- common/autotest_common.sh@958 -- # kill -0 494070 00:45:43.088 18:01:24 keyring_file -- common/autotest_common.sh@959 -- # uname 00:45:43.088 18:01:24 keyring_file -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:45:43.088 18:01:24 keyring_file -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 494070 00:45:43.345 18:01:24 keyring_file -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:45:43.345 18:01:24 keyring_file -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:45:43.345 18:01:24 keyring_file -- common/autotest_common.sh@972 -- # echo 'killing process with pid 494070' 00:45:43.345 killing process with pid 494070 00:45:43.345 18:01:24 keyring_file -- common/autotest_common.sh@973 -- # kill 494070 00:45:43.345 Received shutdown signal, test time was about 1.000000 seconds 00:45:43.345 00:45:43.345 Latency(us) 00:45:43.345 [2024-12-06T17:01:25.184Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:45:43.345 [2024-12-06T17:01:25.184Z] =================================================================================================================== 00:45:43.345 [2024-12-06T17:01:25.184Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:45:43.345 18:01:24 keyring_file -- common/autotest_common.sh@978 -- # wait 494070 00:45:43.345 18:01:25 keyring_file -- keyring/file.sh@21 -- # killprocess 492595 00:45:43.345 18:01:25 keyring_file -- common/autotest_common.sh@954 -- # '[' -z 492595 ']' 00:45:43.346 18:01:25 keyring_file -- common/autotest_common.sh@958 -- # kill -0 492595 00:45:43.346 18:01:25 keyring_file -- common/autotest_common.sh@959 -- # uname 00:45:43.346 18:01:25 keyring_file -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:45:43.346 18:01:25 keyring_file -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 492595 00:45:43.346 18:01:25 keyring_file -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:45:43.346 18:01:25 keyring_file -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:45:43.346 18:01:25 keyring_file -- common/autotest_common.sh@972 -- # echo 'killing process with pid 492595' 00:45:43.346 killing process with pid 492595 00:45:43.346 18:01:25 keyring_file -- common/autotest_common.sh@973 -- # kill 492595 00:45:43.346 18:01:25 keyring_file -- common/autotest_common.sh@978 -- # wait 492595 00:45:43.909 00:45:43.909 real 0m14.373s 00:45:43.909 user 0m36.852s 00:45:43.909 sys 0m3.215s 00:45:43.909 18:01:25 keyring_file -- common/autotest_common.sh@1130 -- # xtrace_disable 00:45:43.909 18:01:25 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:45:43.909 ************************************ 00:45:43.909 END TEST keyring_file 00:45:43.909 ************************************ 00:45:43.909 18:01:25 -- spdk/autotest.sh@293 -- # [[ y == y ]] 00:45:43.909 18:01:25 -- spdk/autotest.sh@294 -- # run_test keyring_linux /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/keyctl-session-wrapper /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/linux.sh 00:45:43.909 18:01:25 -- common/autotest_common.sh@1105 -- # '[' 3 -le 1 ']' 00:45:43.909 18:01:25 -- common/autotest_common.sh@1111 -- # xtrace_disable 00:45:43.909 18:01:25 -- common/autotest_common.sh@10 -- # set +x 00:45:43.909 ************************************ 00:45:43.909 START TEST keyring_linux 00:45:43.909 ************************************ 00:45:43.909 18:01:25 keyring_linux -- common/autotest_common.sh@1129 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/keyctl-session-wrapper /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/linux.sh 00:45:43.909 Joined session keyring: 610903497 00:45:43.909 * Looking for test storage... 00:45:43.909 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring 00:45:43.909 18:01:25 keyring_linux -- common/autotest_common.sh@1710 -- # [[ y == y ]] 00:45:43.909 18:01:25 keyring_linux -- common/autotest_common.sh@1711 -- # lcov --version 00:45:43.909 18:01:25 keyring_linux -- common/autotest_common.sh@1711 -- # awk '{print $NF}' 00:45:43.909 18:01:25 keyring_linux -- common/autotest_common.sh@1711 -- # lt 1.15 2 00:45:43.909 18:01:25 keyring_linux -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:45:43.909 18:01:25 keyring_linux -- scripts/common.sh@333 -- # local ver1 ver1_l 00:45:43.909 18:01:25 keyring_linux -- scripts/common.sh@334 -- # local ver2 ver2_l 00:45:43.909 18:01:25 keyring_linux -- scripts/common.sh@336 -- # IFS=.-: 00:45:43.909 18:01:25 keyring_linux -- scripts/common.sh@336 -- # read -ra ver1 00:45:43.909 18:01:25 keyring_linux -- scripts/common.sh@337 -- # IFS=.-: 00:45:43.909 18:01:25 keyring_linux -- scripts/common.sh@337 -- # read -ra ver2 00:45:43.909 18:01:25 keyring_linux -- scripts/common.sh@338 -- # local 'op=<' 00:45:43.909 18:01:25 keyring_linux -- scripts/common.sh@340 -- # ver1_l=2 00:45:43.909 18:01:25 keyring_linux -- scripts/common.sh@341 -- # ver2_l=1 00:45:43.909 18:01:25 keyring_linux -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:45:43.909 18:01:25 keyring_linux -- scripts/common.sh@344 -- # case "$op" in 00:45:43.909 18:01:25 keyring_linux -- scripts/common.sh@345 -- # : 1 00:45:43.909 18:01:25 keyring_linux -- scripts/common.sh@364 -- # (( v = 0 )) 00:45:43.909 18:01:25 keyring_linux -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:45:43.909 18:01:25 keyring_linux -- scripts/common.sh@365 -- # decimal 1 00:45:43.909 18:01:25 keyring_linux -- scripts/common.sh@353 -- # local d=1 00:45:43.909 18:01:25 keyring_linux -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:45:43.909 18:01:25 keyring_linux -- scripts/common.sh@355 -- # echo 1 00:45:43.909 18:01:25 keyring_linux -- scripts/common.sh@365 -- # ver1[v]=1 00:45:43.909 18:01:25 keyring_linux -- scripts/common.sh@366 -- # decimal 2 00:45:43.909 18:01:25 keyring_linux -- scripts/common.sh@353 -- # local d=2 00:45:43.909 18:01:25 keyring_linux -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:45:43.909 18:01:25 keyring_linux -- scripts/common.sh@355 -- # echo 2 00:45:43.909 18:01:25 keyring_linux -- scripts/common.sh@366 -- # ver2[v]=2 00:45:43.909 18:01:25 keyring_linux -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:45:43.909 18:01:25 keyring_linux -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:45:43.910 18:01:25 keyring_linux -- scripts/common.sh@368 -- # return 0 00:45:43.910 18:01:25 keyring_linux -- common/autotest_common.sh@1712 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:45:43.910 18:01:25 keyring_linux -- common/autotest_common.sh@1724 -- # export 'LCOV_OPTS= 00:45:43.910 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:45:43.910 --rc genhtml_branch_coverage=1 00:45:43.910 --rc genhtml_function_coverage=1 00:45:43.910 --rc genhtml_legend=1 00:45:43.910 --rc geninfo_all_blocks=1 00:45:43.910 --rc geninfo_unexecuted_blocks=1 00:45:43.910 00:45:43.910 ' 00:45:43.910 18:01:25 keyring_linux -- common/autotest_common.sh@1724 -- # LCOV_OPTS=' 00:45:43.910 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:45:43.910 --rc genhtml_branch_coverage=1 00:45:43.910 --rc genhtml_function_coverage=1 00:45:43.910 --rc genhtml_legend=1 00:45:43.910 --rc geninfo_all_blocks=1 00:45:43.910 --rc geninfo_unexecuted_blocks=1 00:45:43.910 00:45:43.910 ' 00:45:43.910 18:01:25 keyring_linux -- common/autotest_common.sh@1725 -- # export 'LCOV=lcov 00:45:43.910 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:45:43.910 --rc genhtml_branch_coverage=1 00:45:43.910 --rc genhtml_function_coverage=1 00:45:43.910 --rc genhtml_legend=1 00:45:43.910 --rc geninfo_all_blocks=1 00:45:43.910 --rc geninfo_unexecuted_blocks=1 00:45:43.910 00:45:43.910 ' 00:45:43.910 18:01:25 keyring_linux -- common/autotest_common.sh@1725 -- # LCOV='lcov 00:45:43.910 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:45:43.910 --rc genhtml_branch_coverage=1 00:45:43.910 --rc genhtml_function_coverage=1 00:45:43.910 --rc genhtml_legend=1 00:45:43.910 --rc geninfo_all_blocks=1 00:45:43.910 --rc geninfo_unexecuted_blocks=1 00:45:43.910 00:45:43.910 ' 00:45:43.910 18:01:25 keyring_linux -- keyring/linux.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/common.sh 00:45:43.910 18:01:25 keyring_linux -- keyring/common.sh@4 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@7 -- # uname -s 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:5b23e107-7094-e311-b1cb-001e67a97d55 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@18 -- # NVME_HOSTID=5b23e107-7094-e311-b1cb-001e67a97d55 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:45:43.910 18:01:25 keyring_linux -- scripts/common.sh@15 -- # shopt -s extglob 00:45:43.910 18:01:25 keyring_linux -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:45:43.910 18:01:25 keyring_linux -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:45:43.910 18:01:25 keyring_linux -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:45:43.910 18:01:25 keyring_linux -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:45:43.910 18:01:25 keyring_linux -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:45:43.910 18:01:25 keyring_linux -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:45:43.910 18:01:25 keyring_linux -- paths/export.sh@5 -- # export PATH 00:45:43.910 18:01:25 keyring_linux -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@51 -- # : 0 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:45:43.910 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@55 -- # have_pci_nics=0 00:45:43.910 18:01:25 keyring_linux -- keyring/common.sh@6 -- # bperfsock=/var/tmp/bperf.sock 00:45:43.910 18:01:25 keyring_linux -- keyring/linux.sh@11 -- # subnqn=nqn.2016-06.io.spdk:cnode0 00:45:43.910 18:01:25 keyring_linux -- keyring/linux.sh@12 -- # hostnqn=nqn.2016-06.io.spdk:host0 00:45:43.910 18:01:25 keyring_linux -- keyring/linux.sh@13 -- # key0=00112233445566778899aabbccddeeff 00:45:43.910 18:01:25 keyring_linux -- keyring/linux.sh@14 -- # key1=112233445566778899aabbccddeeff00 00:45:43.910 18:01:25 keyring_linux -- keyring/linux.sh@45 -- # trap cleanup EXIT 00:45:43.910 18:01:25 keyring_linux -- keyring/linux.sh@47 -- # prep_key key0 00112233445566778899aabbccddeeff 0 /tmp/:spdk-test:key0 00:45:43.910 18:01:25 keyring_linux -- keyring/common.sh@15 -- # local name key digest path 00:45:43.910 18:01:25 keyring_linux -- keyring/common.sh@17 -- # name=key0 00:45:43.910 18:01:25 keyring_linux -- keyring/common.sh@17 -- # key=00112233445566778899aabbccddeeff 00:45:43.910 18:01:25 keyring_linux -- keyring/common.sh@17 -- # digest=0 00:45:43.910 18:01:25 keyring_linux -- keyring/common.sh@18 -- # path=/tmp/:spdk-test:key0 00:45:43.910 18:01:25 keyring_linux -- keyring/common.sh@20 -- # format_interchange_psk 00112233445566778899aabbccddeeff 0 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@743 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff 0 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@730 -- # local prefix key digest 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@732 -- # prefix=NVMeTLSkey-1 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@732 -- # key=00112233445566778899aabbccddeeff 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@732 -- # digest=0 00:45:43.910 18:01:25 keyring_linux -- nvmf/common.sh@733 -- # python - 00:45:44.167 18:01:25 keyring_linux -- keyring/common.sh@21 -- # chmod 0600 /tmp/:spdk-test:key0 00:45:44.167 18:01:25 keyring_linux -- keyring/common.sh@23 -- # echo /tmp/:spdk-test:key0 00:45:44.167 /tmp/:spdk-test:key0 00:45:44.167 18:01:25 keyring_linux -- keyring/linux.sh@48 -- # prep_key key1 112233445566778899aabbccddeeff00 0 /tmp/:spdk-test:key1 00:45:44.167 18:01:25 keyring_linux -- keyring/common.sh@15 -- # local name key digest path 00:45:44.167 18:01:25 keyring_linux -- keyring/common.sh@17 -- # name=key1 00:45:44.168 18:01:25 keyring_linux -- keyring/common.sh@17 -- # key=112233445566778899aabbccddeeff00 00:45:44.168 18:01:25 keyring_linux -- keyring/common.sh@17 -- # digest=0 00:45:44.168 18:01:25 keyring_linux -- keyring/common.sh@18 -- # path=/tmp/:spdk-test:key1 00:45:44.168 18:01:25 keyring_linux -- keyring/common.sh@20 -- # format_interchange_psk 112233445566778899aabbccddeeff00 0 00:45:44.168 18:01:25 keyring_linux -- nvmf/common.sh@743 -- # format_key NVMeTLSkey-1 112233445566778899aabbccddeeff00 0 00:45:44.168 18:01:25 keyring_linux -- nvmf/common.sh@730 -- # local prefix key digest 00:45:44.168 18:01:25 keyring_linux -- nvmf/common.sh@732 -- # prefix=NVMeTLSkey-1 00:45:44.168 18:01:25 keyring_linux -- nvmf/common.sh@732 -- # key=112233445566778899aabbccddeeff00 00:45:44.168 18:01:25 keyring_linux -- nvmf/common.sh@732 -- # digest=0 00:45:44.168 18:01:25 keyring_linux -- nvmf/common.sh@733 -- # python - 00:45:44.168 18:01:25 keyring_linux -- keyring/common.sh@21 -- # chmod 0600 /tmp/:spdk-test:key1 00:45:44.168 18:01:25 keyring_linux -- keyring/common.sh@23 -- # echo /tmp/:spdk-test:key1 00:45:44.168 /tmp/:spdk-test:key1 00:45:44.168 18:01:25 keyring_linux -- keyring/linux.sh@51 -- # tgtpid=494465 00:45:44.168 18:01:25 keyring_linux -- keyring/linux.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:45:44.168 18:01:25 keyring_linux -- keyring/linux.sh@53 -- # waitforlisten 494465 00:45:44.168 18:01:25 keyring_linux -- common/autotest_common.sh@835 -- # '[' -z 494465 ']' 00:45:44.168 18:01:25 keyring_linux -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/spdk.sock 00:45:44.168 18:01:25 keyring_linux -- common/autotest_common.sh@840 -- # local max_retries=100 00:45:44.168 18:01:25 keyring_linux -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:45:44.168 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:45:44.168 18:01:25 keyring_linux -- common/autotest_common.sh@844 -- # xtrace_disable 00:45:44.168 18:01:25 keyring_linux -- common/autotest_common.sh@10 -- # set +x 00:45:44.168 [2024-12-06 18:01:25.857936] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:45:44.168 [2024-12-06 18:01:25.858060] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid494465 ] 00:45:44.168 [2024-12-06 18:01:25.927431] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:45:44.168 [2024-12-06 18:01:25.971489] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:45:44.426 18:01:26 keyring_linux -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:45:44.426 18:01:26 keyring_linux -- common/autotest_common.sh@868 -- # return 0 00:45:44.426 18:01:26 keyring_linux -- keyring/linux.sh@54 -- # rpc_cmd 00:45:44.426 18:01:26 keyring_linux -- common/autotest_common.sh@563 -- # xtrace_disable 00:45:44.426 18:01:26 keyring_linux -- common/autotest_common.sh@10 -- # set +x 00:45:44.426 [2024-12-06 18:01:26.228064] tcp.c: 756:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:45:44.426 null0 00:45:44.426 [2024-12-06 18:01:26.260147] tcp.c:1049:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:45:44.426 [2024-12-06 18:01:26.260720] tcp.c:1099:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4420 *** 00:45:44.685 18:01:26 keyring_linux -- common/autotest_common.sh@591 -- # [[ 0 == 0 ]] 00:45:44.685 18:01:26 keyring_linux -- keyring/linux.sh@66 -- # keyctl add user :spdk-test:key0 NVMeTLSkey-1:00:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmZwJEiQ: @s 00:45:44.685 608868883 00:45:44.685 18:01:26 keyring_linux -- keyring/linux.sh@67 -- # keyctl add user :spdk-test:key1 NVMeTLSkey-1:00:MTEyMjMzNDQ1NTY2Nzc4ODk5YWFiYmNjZGRlZWZmMDA6CPcs: @s 00:45:44.685 496209784 00:45:44.685 18:01:26 keyring_linux -- keyring/linux.sh@70 -- # bperfpid=494563 00:45:44.685 18:01:26 keyring_linux -- keyring/linux.sh@72 -- # waitforlisten 494563 /var/tmp/bperf.sock 00:45:44.685 18:01:26 keyring_linux -- keyring/linux.sh@68 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -q 128 -o 4k -w randread -t 1 -m 2 -r /var/tmp/bperf.sock -z --wait-for-rpc 00:45:44.685 18:01:26 keyring_linux -- common/autotest_common.sh@835 -- # '[' -z 494563 ']' 00:45:44.685 18:01:26 keyring_linux -- common/autotest_common.sh@839 -- # local rpc_addr=/var/tmp/bperf.sock 00:45:44.685 18:01:26 keyring_linux -- common/autotest_common.sh@840 -- # local max_retries=100 00:45:44.685 18:01:26 keyring_linux -- common/autotest_common.sh@842 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:45:44.685 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:45:44.685 18:01:26 keyring_linux -- common/autotest_common.sh@844 -- # xtrace_disable 00:45:44.685 18:01:26 keyring_linux -- common/autotest_common.sh@10 -- # set +x 00:45:44.685 [2024-12-06 18:01:26.329044] Starting SPDK v25.01-pre git sha1 a5e6ecf28 / DPDK 23.11.0 initialization... 00:45:44.685 [2024-12-06 18:01:26.329123] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid494563 ] 00:45:44.685 [2024-12-06 18:01:26.395095] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:45:44.685 [2024-12-06 18:01:26.439809] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:45:44.969 18:01:26 keyring_linux -- common/autotest_common.sh@864 -- # (( i == 0 )) 00:45:44.969 18:01:26 keyring_linux -- common/autotest_common.sh@868 -- # return 0 00:45:44.969 18:01:26 keyring_linux -- keyring/linux.sh@73 -- # bperf_cmd keyring_linux_set_options --enable 00:45:44.969 18:01:26 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_linux_set_options --enable 00:45:45.228 18:01:26 keyring_linux -- keyring/linux.sh@74 -- # bperf_cmd framework_start_init 00:45:45.228 18:01:26 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:45:45.485 18:01:27 keyring_linux -- keyring/linux.sh@75 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key0 00:45:45.485 18:01:27 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key0 00:45:45.743 [2024-12-06 18:01:27.436741] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:45:45.743 nvme0n1 00:45:45.743 18:01:27 keyring_linux -- keyring/linux.sh@77 -- # check_keys 1 :spdk-test:key0 00:45:45.743 18:01:27 keyring_linux -- keyring/linux.sh@19 -- # local count=1 name=:spdk-test:key0 00:45:45.743 18:01:27 keyring_linux -- keyring/linux.sh@20 -- # local sn 00:45:45.743 18:01:27 keyring_linux -- keyring/linux.sh@22 -- # bperf_cmd keyring_get_keys 00:45:45.743 18:01:27 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:45:45.743 18:01:27 keyring_linux -- keyring/linux.sh@22 -- # jq length 00:45:46.000 18:01:27 keyring_linux -- keyring/linux.sh@22 -- # (( 1 == count )) 00:45:46.000 18:01:27 keyring_linux -- keyring/linux.sh@23 -- # (( count == 0 )) 00:45:46.000 18:01:27 keyring_linux -- keyring/linux.sh@25 -- # get_key :spdk-test:key0 00:45:46.000 18:01:27 keyring_linux -- keyring/linux.sh@25 -- # jq -r .sn 00:45:46.000 18:01:27 keyring_linux -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:45:46.000 18:01:27 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:45:46.000 18:01:27 keyring_linux -- keyring/common.sh@10 -- # jq '.[] | select(.name == ":spdk-test:key0")' 00:45:46.258 18:01:28 keyring_linux -- keyring/linux.sh@25 -- # sn=608868883 00:45:46.258 18:01:28 keyring_linux -- keyring/linux.sh@26 -- # get_keysn :spdk-test:key0 00:45:46.258 18:01:28 keyring_linux -- keyring/linux.sh@16 -- # keyctl search @s user :spdk-test:key0 00:45:46.258 18:01:28 keyring_linux -- keyring/linux.sh@26 -- # [[ 608868883 == \6\0\8\8\6\8\8\8\3 ]] 00:45:46.258 18:01:28 keyring_linux -- keyring/linux.sh@27 -- # keyctl print 608868883 00:45:46.258 18:01:28 keyring_linux -- keyring/linux.sh@27 -- # [[ NVMeTLSkey-1:00:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmZwJEiQ: == \N\V\M\e\T\L\S\k\e\y\-\1\:\0\0\:\M\D\A\x\M\T\I\y\M\z\M\0\N\D\U\1\N\j\Y\3\N\z\g\4\O\T\l\h\Y\W\J\i\Y\2\N\k\Z\G\V\l\Z\m\Z\w\J\E\i\Q\: ]] 00:45:46.258 18:01:28 keyring_linux -- keyring/linux.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:45:46.515 Running I/O for 1 seconds... 00:45:47.450 11257.00 IOPS, 43.97 MiB/s 00:45:47.450 Latency(us) 00:45:47.450 [2024-12-06T17:01:29.289Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:45:47.450 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 128, IO size: 4096) 00:45:47.450 nvme0n1 : 1.01 11265.13 44.00 0.00 0.00 11294.24 5509.88 16990.81 00:45:47.450 [2024-12-06T17:01:29.289Z] =================================================================================================================== 00:45:47.450 [2024-12-06T17:01:29.289Z] Total : 11265.13 44.00 0.00 0.00 11294.24 5509.88 16990.81 00:45:47.450 { 00:45:47.450 "results": [ 00:45:47.450 { 00:45:47.450 "job": "nvme0n1", 00:45:47.450 "core_mask": "0x2", 00:45:47.450 "workload": "randread", 00:45:47.450 "status": "finished", 00:45:47.450 "queue_depth": 128, 00:45:47.450 "io_size": 4096, 00:45:47.450 "runtime": 1.01073, 00:45:47.450 "iops": 11265.125206533892, 00:45:47.450 "mibps": 44.004395338023016, 00:45:47.450 "io_failed": 0, 00:45:47.450 "io_timeout": 0, 00:45:47.450 "avg_latency_us": 11294.242411798767, 00:45:47.450 "min_latency_us": 5509.878518518519, 00:45:47.450 "max_latency_us": 16990.814814814814 00:45:47.450 } 00:45:47.450 ], 00:45:47.450 "core_count": 1 00:45:47.450 } 00:45:47.450 18:01:29 keyring_linux -- keyring/linux.sh@80 -- # bperf_cmd bdev_nvme_detach_controller nvme0 00:45:47.450 18:01:29 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_detach_controller nvme0 00:45:47.708 18:01:29 keyring_linux -- keyring/linux.sh@81 -- # check_keys 0 00:45:47.708 18:01:29 keyring_linux -- keyring/linux.sh@19 -- # local count=0 name= 00:45:47.708 18:01:29 keyring_linux -- keyring/linux.sh@20 -- # local sn 00:45:47.708 18:01:29 keyring_linux -- keyring/linux.sh@22 -- # bperf_cmd keyring_get_keys 00:45:47.708 18:01:29 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:45:47.708 18:01:29 keyring_linux -- keyring/linux.sh@22 -- # jq length 00:45:47.967 18:01:29 keyring_linux -- keyring/linux.sh@22 -- # (( 0 == count )) 00:45:47.967 18:01:29 keyring_linux -- keyring/linux.sh@23 -- # (( count == 0 )) 00:45:47.967 18:01:29 keyring_linux -- keyring/linux.sh@23 -- # return 00:45:47.967 18:01:29 keyring_linux -- keyring/linux.sh@84 -- # NOT bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key1 00:45:47.967 18:01:29 keyring_linux -- common/autotest_common.sh@652 -- # local es=0 00:45:47.967 18:01:29 keyring_linux -- common/autotest_common.sh@654 -- # valid_exec_arg bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key1 00:45:47.967 18:01:29 keyring_linux -- common/autotest_common.sh@640 -- # local arg=bperf_cmd 00:45:47.967 18:01:29 keyring_linux -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:45:47.967 18:01:29 keyring_linux -- common/autotest_common.sh@644 -- # type -t bperf_cmd 00:45:47.967 18:01:29 keyring_linux -- common/autotest_common.sh@644 -- # case "$(type -t "$arg")" in 00:45:47.967 18:01:29 keyring_linux -- common/autotest_common.sh@655 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key1 00:45:47.967 18:01:29 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key1 00:45:48.226 [2024-12-06 18:01:29.994202] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 421:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:45:48.226 [2024-12-06 18:01:29.994962] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xcd3930 (107): Transport endpoint is not connected 00:45:48.226 [2024-12-06 18:01:29.995954] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xcd3930 (9): Bad file descriptor 00:45:48.226 [2024-12-06 18:01:29.996967] nvme_ctrlr.c:4206:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 0] Ctrlr is in error state 00:45:48.226 [2024-12-06 18:01:29.996985] nvme.c: 709:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 127.0.0.1 00:45:48.226 [2024-12-06 18:01:29.996998] nvme.c: 895:nvme_dummy_attach_fail_cb: *ERROR*: Failed to attach nvme ctrlr: trtype=TCP adrfam=IPv4 traddr=127.0.0.1 trsvcid=4420 subnqn=nqn.2016-06.io.spdk:cnode0, Operation not permitted 00:45:48.226 [2024-12-06 18:01:29.997012] nvme_ctrlr.c:1110:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 0] in failed state. 00:45:48.226 request: 00:45:48.226 { 00:45:48.226 "name": "nvme0", 00:45:48.226 "trtype": "tcp", 00:45:48.226 "traddr": "127.0.0.1", 00:45:48.226 "adrfam": "ipv4", 00:45:48.226 "trsvcid": "4420", 00:45:48.226 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:45:48.226 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:45:48.226 "prchk_reftag": false, 00:45:48.226 "prchk_guard": false, 00:45:48.226 "hdgst": false, 00:45:48.226 "ddgst": false, 00:45:48.226 "psk": ":spdk-test:key1", 00:45:48.226 "allow_unrecognized_csi": false, 00:45:48.226 "method": "bdev_nvme_attach_controller", 00:45:48.226 "req_id": 1 00:45:48.226 } 00:45:48.226 Got JSON-RPC error response 00:45:48.226 response: 00:45:48.226 { 00:45:48.226 "code": -5, 00:45:48.226 "message": "Input/output error" 00:45:48.226 } 00:45:48.226 18:01:30 keyring_linux -- common/autotest_common.sh@655 -- # es=1 00:45:48.226 18:01:30 keyring_linux -- common/autotest_common.sh@663 -- # (( es > 128 )) 00:45:48.226 18:01:30 keyring_linux -- common/autotest_common.sh@674 -- # [[ -n '' ]] 00:45:48.226 18:01:30 keyring_linux -- common/autotest_common.sh@679 -- # (( !es == 0 )) 00:45:48.226 18:01:30 keyring_linux -- keyring/linux.sh@1 -- # cleanup 00:45:48.226 18:01:30 keyring_linux -- keyring/linux.sh@38 -- # for key in key0 key1 00:45:48.226 18:01:30 keyring_linux -- keyring/linux.sh@39 -- # unlink_key key0 00:45:48.226 18:01:30 keyring_linux -- keyring/linux.sh@31 -- # local name=key0 sn 00:45:48.226 18:01:30 keyring_linux -- keyring/linux.sh@33 -- # get_keysn :spdk-test:key0 00:45:48.226 18:01:30 keyring_linux -- keyring/linux.sh@16 -- # keyctl search @s user :spdk-test:key0 00:45:48.226 18:01:30 keyring_linux -- keyring/linux.sh@33 -- # sn=608868883 00:45:48.226 18:01:30 keyring_linux -- keyring/linux.sh@34 -- # keyctl unlink 608868883 00:45:48.226 1 links removed 00:45:48.226 18:01:30 keyring_linux -- keyring/linux.sh@38 -- # for key in key0 key1 00:45:48.226 18:01:30 keyring_linux -- keyring/linux.sh@39 -- # unlink_key key1 00:45:48.226 18:01:30 keyring_linux -- keyring/linux.sh@31 -- # local name=key1 sn 00:45:48.226 18:01:30 keyring_linux -- keyring/linux.sh@33 -- # get_keysn :spdk-test:key1 00:45:48.226 18:01:30 keyring_linux -- keyring/linux.sh@16 -- # keyctl search @s user :spdk-test:key1 00:45:48.226 18:01:30 keyring_linux -- keyring/linux.sh@33 -- # sn=496209784 00:45:48.226 18:01:30 keyring_linux -- keyring/linux.sh@34 -- # keyctl unlink 496209784 00:45:48.226 1 links removed 00:45:48.226 18:01:30 keyring_linux -- keyring/linux.sh@41 -- # killprocess 494563 00:45:48.226 18:01:30 keyring_linux -- common/autotest_common.sh@954 -- # '[' -z 494563 ']' 00:45:48.226 18:01:30 keyring_linux -- common/autotest_common.sh@958 -- # kill -0 494563 00:45:48.226 18:01:30 keyring_linux -- common/autotest_common.sh@959 -- # uname 00:45:48.226 18:01:30 keyring_linux -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:45:48.226 18:01:30 keyring_linux -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 494563 00:45:48.226 18:01:30 keyring_linux -- common/autotest_common.sh@960 -- # process_name=reactor_1 00:45:48.226 18:01:30 keyring_linux -- common/autotest_common.sh@964 -- # '[' reactor_1 = sudo ']' 00:45:48.226 18:01:30 keyring_linux -- common/autotest_common.sh@972 -- # echo 'killing process with pid 494563' 00:45:48.226 killing process with pid 494563 00:45:48.226 18:01:30 keyring_linux -- common/autotest_common.sh@973 -- # kill 494563 00:45:48.227 Received shutdown signal, test time was about 1.000000 seconds 00:45:48.227 00:45:48.227 Latency(us) 00:45:48.227 [2024-12-06T17:01:30.066Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:45:48.227 [2024-12-06T17:01:30.066Z] =================================================================================================================== 00:45:48.227 [2024-12-06T17:01:30.066Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:45:48.227 18:01:30 keyring_linux -- common/autotest_common.sh@978 -- # wait 494563 00:45:48.485 18:01:30 keyring_linux -- keyring/linux.sh@42 -- # killprocess 494465 00:45:48.485 18:01:30 keyring_linux -- common/autotest_common.sh@954 -- # '[' -z 494465 ']' 00:45:48.485 18:01:30 keyring_linux -- common/autotest_common.sh@958 -- # kill -0 494465 00:45:48.485 18:01:30 keyring_linux -- common/autotest_common.sh@959 -- # uname 00:45:48.485 18:01:30 keyring_linux -- common/autotest_common.sh@959 -- # '[' Linux = Linux ']' 00:45:48.485 18:01:30 keyring_linux -- common/autotest_common.sh@960 -- # ps --no-headers -o comm= 494465 00:45:48.485 18:01:30 keyring_linux -- common/autotest_common.sh@960 -- # process_name=reactor_0 00:45:48.485 18:01:30 keyring_linux -- common/autotest_common.sh@964 -- # '[' reactor_0 = sudo ']' 00:45:48.485 18:01:30 keyring_linux -- common/autotest_common.sh@972 -- # echo 'killing process with pid 494465' 00:45:48.485 killing process with pid 494465 00:45:48.485 18:01:30 keyring_linux -- common/autotest_common.sh@973 -- # kill 494465 00:45:48.485 18:01:30 keyring_linux -- common/autotest_common.sh@978 -- # wait 494465 00:45:49.051 00:45:49.051 real 0m5.054s 00:45:49.051 user 0m10.104s 00:45:49.051 sys 0m1.593s 00:45:49.051 18:01:30 keyring_linux -- common/autotest_common.sh@1130 -- # xtrace_disable 00:45:49.051 18:01:30 keyring_linux -- common/autotest_common.sh@10 -- # set +x 00:45:49.051 ************************************ 00:45:49.051 END TEST keyring_linux 00:45:49.051 ************************************ 00:45:49.051 18:01:30 -- spdk/autotest.sh@311 -- # '[' 0 -eq 1 ']' 00:45:49.051 18:01:30 -- spdk/autotest.sh@315 -- # '[' 0 -eq 1 ']' 00:45:49.051 18:01:30 -- spdk/autotest.sh@319 -- # '[' 0 -eq 1 ']' 00:45:49.051 18:01:30 -- spdk/autotest.sh@324 -- # '[' 0 -eq 1 ']' 00:45:49.051 18:01:30 -- spdk/autotest.sh@333 -- # '[' 0 -eq 1 ']' 00:45:49.051 18:01:30 -- spdk/autotest.sh@338 -- # '[' 0 -eq 1 ']' 00:45:49.051 18:01:30 -- spdk/autotest.sh@342 -- # '[' 0 -eq 1 ']' 00:45:49.051 18:01:30 -- spdk/autotest.sh@346 -- # '[' 0 -eq 1 ']' 00:45:49.051 18:01:30 -- spdk/autotest.sh@350 -- # '[' 0 -eq 1 ']' 00:45:49.051 18:01:30 -- spdk/autotest.sh@355 -- # '[' 0 -eq 1 ']' 00:45:49.051 18:01:30 -- spdk/autotest.sh@359 -- # '[' 0 -eq 1 ']' 00:45:49.051 18:01:30 -- spdk/autotest.sh@366 -- # [[ 0 -eq 1 ]] 00:45:49.051 18:01:30 -- spdk/autotest.sh@370 -- # [[ 0 -eq 1 ]] 00:45:49.051 18:01:30 -- spdk/autotest.sh@374 -- # [[ 0 -eq 1 ]] 00:45:49.052 18:01:30 -- spdk/autotest.sh@378 -- # [[ '' -eq 1 ]] 00:45:49.052 18:01:30 -- spdk/autotest.sh@385 -- # trap - SIGINT SIGTERM EXIT 00:45:49.052 18:01:30 -- spdk/autotest.sh@387 -- # timing_enter post_cleanup 00:45:49.052 18:01:30 -- common/autotest_common.sh@726 -- # xtrace_disable 00:45:49.052 18:01:30 -- common/autotest_common.sh@10 -- # set +x 00:45:49.052 18:01:30 -- spdk/autotest.sh@388 -- # autotest_cleanup 00:45:49.052 18:01:30 -- common/autotest_common.sh@1396 -- # local autotest_es=0 00:45:49.052 18:01:30 -- common/autotest_common.sh@1397 -- # xtrace_disable 00:45:49.052 18:01:30 -- common/autotest_common.sh@10 -- # set +x 00:45:50.955 INFO: APP EXITING 00:45:50.955 INFO: killing all VMs 00:45:50.955 INFO: killing vhost app 00:45:50.955 INFO: EXIT DONE 00:45:52.332 0000:88:00.0 (8086 0a54): Already using the nvme driver 00:45:52.332 0000:00:04.7 (8086 0e27): Already using the ioatdma driver 00:45:52.332 0000:00:04.6 (8086 0e26): Already using the ioatdma driver 00:45:52.332 0000:00:04.5 (8086 0e25): Already using the ioatdma driver 00:45:52.332 0000:00:04.4 (8086 0e24): Already using the ioatdma driver 00:45:52.332 0000:00:04.3 (8086 0e23): Already using the ioatdma driver 00:45:52.332 0000:00:04.2 (8086 0e22): Already using the ioatdma driver 00:45:52.332 0000:00:04.1 (8086 0e21): Already using the ioatdma driver 00:45:52.332 0000:00:04.0 (8086 0e20): Already using the ioatdma driver 00:45:52.332 0000:80:04.7 (8086 0e27): Already using the ioatdma driver 00:45:52.332 0000:80:04.6 (8086 0e26): Already using the ioatdma driver 00:45:52.332 0000:80:04.5 (8086 0e25): Already using the ioatdma driver 00:45:52.332 0000:80:04.4 (8086 0e24): Already using the ioatdma driver 00:45:52.332 0000:80:04.3 (8086 0e23): Already using the ioatdma driver 00:45:52.332 0000:80:04.2 (8086 0e22): Already using the ioatdma driver 00:45:52.332 0000:80:04.1 (8086 0e21): Already using the ioatdma driver 00:45:52.332 0000:80:04.0 (8086 0e20): Already using the ioatdma driver 00:45:53.705 Cleaning 00:45:53.705 Removing: /var/run/dpdk/spdk0/config 00:45:53.705 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-0 00:45:53.705 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-1 00:45:53.705 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-2 00:45:53.705 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-3 00:45:53.705 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-1-0 00:45:53.705 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-1-1 00:45:53.705 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-1-2 00:45:53.705 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-1-3 00:45:53.705 Removing: /var/run/dpdk/spdk0/fbarray_memzone 00:45:53.705 Removing: /var/run/dpdk/spdk0/hugepage_info 00:45:53.705 Removing: /var/run/dpdk/spdk1/config 00:45:53.705 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-0-0 00:45:53.706 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-0-1 00:45:53.706 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-0-2 00:45:53.706 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-0-3 00:45:53.706 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-1-0 00:45:53.706 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-1-1 00:45:53.706 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-1-2 00:45:53.706 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-1-3 00:45:53.706 Removing: /var/run/dpdk/spdk1/fbarray_memzone 00:45:53.706 Removing: /var/run/dpdk/spdk1/hugepage_info 00:45:53.706 Removing: /var/run/dpdk/spdk2/config 00:45:53.706 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-0-0 00:45:53.706 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-0-1 00:45:53.706 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-0-2 00:45:53.706 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-0-3 00:45:53.706 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-1-0 00:45:53.706 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-1-1 00:45:53.706 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-1-2 00:45:53.706 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-1-3 00:45:53.706 Removing: /var/run/dpdk/spdk2/fbarray_memzone 00:45:53.706 Removing: /var/run/dpdk/spdk2/hugepage_info 00:45:53.706 Removing: /var/run/dpdk/spdk3/config 00:45:53.706 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-0-0 00:45:53.706 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-0-1 00:45:53.706 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-0-2 00:45:53.706 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-0-3 00:45:53.706 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-1-0 00:45:53.706 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-1-1 00:45:53.706 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-1-2 00:45:53.706 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-1-3 00:45:53.706 Removing: /var/run/dpdk/spdk3/fbarray_memzone 00:45:53.706 Removing: /var/run/dpdk/spdk3/hugepage_info 00:45:53.706 Removing: /var/run/dpdk/spdk4/config 00:45:53.706 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-0-0 00:45:53.706 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-0-1 00:45:53.706 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-0-2 00:45:53.706 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-0-3 00:45:53.706 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-1-0 00:45:53.706 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-1-1 00:45:53.706 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-1-2 00:45:53.706 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-1-3 00:45:53.706 Removing: /var/run/dpdk/spdk4/fbarray_memzone 00:45:53.706 Removing: /var/run/dpdk/spdk4/hugepage_info 00:45:53.706 Removing: /dev/shm/bdev_svc_trace.1 00:45:53.706 Removing: /dev/shm/nvmf_trace.0 00:45:53.706 Removing: /dev/shm/spdk_tgt_trace.pid110959 00:45:53.706 Removing: /var/run/dpdk/spdk0 00:45:53.706 Removing: /var/run/dpdk/spdk1 00:45:53.706 Removing: /var/run/dpdk/spdk2 00:45:53.706 Removing: /var/run/dpdk/spdk3 00:45:53.706 Removing: /var/run/dpdk/spdk4 00:45:53.706 Removing: /var/run/dpdk/spdk_pid109273 00:45:53.706 Removing: /var/run/dpdk/spdk_pid110016 00:45:53.706 Removing: /var/run/dpdk/spdk_pid110959 00:45:53.706 Removing: /var/run/dpdk/spdk_pid111280 00:45:53.706 Removing: /var/run/dpdk/spdk_pid111967 00:45:53.706 Removing: /var/run/dpdk/spdk_pid112109 00:45:53.706 Removing: /var/run/dpdk/spdk_pid112829 00:45:53.706 Removing: /var/run/dpdk/spdk_pid112914 00:45:53.706 Removing: /var/run/dpdk/spdk_pid113121 00:45:53.706 Removing: /var/run/dpdk/spdk_pid114419 00:45:53.706 Removing: /var/run/dpdk/spdk_pid115341 00:45:53.706 Removing: /var/run/dpdk/spdk_pid115652 00:45:53.706 Removing: /var/run/dpdk/spdk_pid115853 00:45:53.706 Removing: /var/run/dpdk/spdk_pid116066 00:45:53.706 Removing: /var/run/dpdk/spdk_pid116262 00:45:53.706 Removing: /var/run/dpdk/spdk_pid116420 00:45:53.706 Removing: /var/run/dpdk/spdk_pid116591 00:45:53.706 Removing: /var/run/dpdk/spdk_pid116886 00:45:53.706 Removing: /var/run/dpdk/spdk_pid117078 00:45:53.706 Removing: /var/run/dpdk/spdk_pid119570 00:45:53.706 Removing: /var/run/dpdk/spdk_pid119732 00:45:53.706 Removing: /var/run/dpdk/spdk_pid119898 00:45:53.706 Removing: /var/run/dpdk/spdk_pid120021 00:45:53.706 Removing: /var/run/dpdk/spdk_pid120319 00:45:53.706 Removing: /var/run/dpdk/spdk_pid120333 00:45:53.706 Removing: /var/run/dpdk/spdk_pid120631 00:45:53.706 Removing: /var/run/dpdk/spdk_pid120760 00:45:53.706 Removing: /var/run/dpdk/spdk_pid120929 00:45:53.706 Removing: /var/run/dpdk/spdk_pid121022 00:45:53.706 Removing: /var/run/dpdk/spdk_pid121217 00:45:53.706 Removing: /var/run/dpdk/spdk_pid121234 00:45:53.706 Removing: /var/run/dpdk/spdk_pid121611 00:45:53.706 Removing: /var/run/dpdk/spdk_pid121832 00:45:53.706 Removing: /var/run/dpdk/spdk_pid122087 00:45:53.706 Removing: /var/run/dpdk/spdk_pid124205 00:45:53.706 Removing: /var/run/dpdk/spdk_pid126844 00:45:53.706 Removing: /var/run/dpdk/spdk_pid133862 00:45:53.706 Removing: /var/run/dpdk/spdk_pid134267 00:45:53.706 Removing: /var/run/dpdk/spdk_pid136900 00:45:53.706 Removing: /var/run/dpdk/spdk_pid137181 00:45:53.706 Removing: /var/run/dpdk/spdk_pid140208 00:45:53.706 Removing: /var/run/dpdk/spdk_pid143933 00:45:53.706 Removing: /var/run/dpdk/spdk_pid146120 00:45:53.706 Removing: /var/run/dpdk/spdk_pid152552 00:45:53.706 Removing: /var/run/dpdk/spdk_pid157823 00:45:53.706 Removing: /var/run/dpdk/spdk_pid159024 00:45:53.706 Removing: /var/run/dpdk/spdk_pid159691 00:45:53.706 Removing: /var/run/dpdk/spdk_pid170066 00:45:53.706 Removing: /var/run/dpdk/spdk_pid172478 00:45:53.706 Removing: /var/run/dpdk/spdk_pid227623 00:45:53.706 Removing: /var/run/dpdk/spdk_pid230826 00:45:53.706 Removing: /var/run/dpdk/spdk_pid235258 00:45:53.706 Removing: /var/run/dpdk/spdk_pid239533 00:45:53.706 Removing: /var/run/dpdk/spdk_pid239653 00:45:53.706 Removing: /var/run/dpdk/spdk_pid240196 00:45:53.706 Removing: /var/run/dpdk/spdk_pid240848 00:45:53.706 Removing: /var/run/dpdk/spdk_pid241502 00:45:53.706 Removing: /var/run/dpdk/spdk_pid241901 00:45:53.706 Removing: /var/run/dpdk/spdk_pid241909 00:45:53.706 Removing: /var/run/dpdk/spdk_pid242050 00:45:53.706 Removing: /var/run/dpdk/spdk_pid242184 00:45:53.706 Removing: /var/run/dpdk/spdk_pid242202 00:45:53.706 Removing: /var/run/dpdk/spdk_pid242847 00:45:53.706 Removing: /var/run/dpdk/spdk_pid243497 00:45:53.706 Removing: /var/run/dpdk/spdk_pid244101 00:45:53.706 Removing: /var/run/dpdk/spdk_pid244505 00:45:53.706 Removing: /var/run/dpdk/spdk_pid244559 00:45:53.706 Removing: /var/run/dpdk/spdk_pid244703 00:45:53.706 Removing: /var/run/dpdk/spdk_pid245718 00:45:53.963 Removing: /var/run/dpdk/spdk_pid246446 00:45:53.963 Removing: /var/run/dpdk/spdk_pid251771 00:45:53.963 Removing: /var/run/dpdk/spdk_pid279959 00:45:53.963 Removing: /var/run/dpdk/spdk_pid282890 00:45:53.963 Removing: /var/run/dpdk/spdk_pid284065 00:45:53.963 Removing: /var/run/dpdk/spdk_pid285999 00:45:53.963 Removing: /var/run/dpdk/spdk_pid286137 00:45:53.963 Removing: /var/run/dpdk/spdk_pid286280 00:45:53.963 Removing: /var/run/dpdk/spdk_pid286420 00:45:53.963 Removing: /var/run/dpdk/spdk_pid286866 00:45:53.963 Removing: /var/run/dpdk/spdk_pid288179 00:45:53.963 Removing: /var/run/dpdk/spdk_pid288938 00:45:53.963 Removing: /var/run/dpdk/spdk_pid289357 00:45:53.963 Removing: /var/run/dpdk/spdk_pid290960 00:45:53.963 Removing: /var/run/dpdk/spdk_pid291362 00:45:53.963 Removing: /var/run/dpdk/spdk_pid291827 00:45:53.963 Removing: /var/run/dpdk/spdk_pid294217 00:45:53.963 Removing: /var/run/dpdk/spdk_pid297612 00:45:53.963 Removing: /var/run/dpdk/spdk_pid297613 00:45:53.963 Removing: /var/run/dpdk/spdk_pid297614 00:45:53.963 Removing: /var/run/dpdk/spdk_pid299737 00:45:53.963 Removing: /var/run/dpdk/spdk_pid301920 00:45:53.963 Removing: /var/run/dpdk/spdk_pid305451 00:45:53.963 Removing: /var/run/dpdk/spdk_pid328719 00:45:53.963 Removing: /var/run/dpdk/spdk_pid331430 00:45:53.963 Removing: /var/run/dpdk/spdk_pid335316 00:45:53.964 Removing: /var/run/dpdk/spdk_pid336260 00:45:53.964 Removing: /var/run/dpdk/spdk_pid337355 00:45:53.964 Removing: /var/run/dpdk/spdk_pid338320 00:45:53.964 Removing: /var/run/dpdk/spdk_pid341148 00:45:53.964 Removing: /var/run/dpdk/spdk_pid343661 00:45:53.964 Removing: /var/run/dpdk/spdk_pid346128 00:45:53.964 Removing: /var/run/dpdk/spdk_pid350838 00:45:53.964 Removing: /var/run/dpdk/spdk_pid350878 00:45:53.964 Removing: /var/run/dpdk/spdk_pid353669 00:45:53.964 Removing: /var/run/dpdk/spdk_pid353816 00:45:53.964 Removing: /var/run/dpdk/spdk_pid354050 00:45:53.964 Removing: /var/run/dpdk/spdk_pid354322 00:45:53.964 Removing: /var/run/dpdk/spdk_pid354332 00:45:53.964 Removing: /var/run/dpdk/spdk_pid355418 00:45:53.964 Removing: /var/run/dpdk/spdk_pid356717 00:45:53.964 Removing: /var/run/dpdk/spdk_pid357894 00:45:53.964 Removing: /var/run/dpdk/spdk_pid359068 00:45:53.964 Removing: /var/run/dpdk/spdk_pid360244 00:45:53.964 Removing: /var/run/dpdk/spdk_pid361428 00:45:53.964 Removing: /var/run/dpdk/spdk_pid365238 00:45:53.964 Removing: /var/run/dpdk/spdk_pid365656 00:45:53.964 Removing: /var/run/dpdk/spdk_pid366972 00:45:53.964 Removing: /var/run/dpdk/spdk_pid367709 00:45:53.964 Removing: /var/run/dpdk/spdk_pid371429 00:45:53.964 Removing: /var/run/dpdk/spdk_pid373393 00:45:53.964 Removing: /var/run/dpdk/spdk_pid377299 00:45:53.964 Removing: /var/run/dpdk/spdk_pid380759 00:45:53.964 Removing: /var/run/dpdk/spdk_pid387261 00:45:53.964 Removing: /var/run/dpdk/spdk_pid391610 00:45:53.964 Removing: /var/run/dpdk/spdk_pid391689 00:45:53.964 Removing: /var/run/dpdk/spdk_pid404378 00:45:53.964 Removing: /var/run/dpdk/spdk_pid404906 00:45:53.964 Removing: /var/run/dpdk/spdk_pid405313 00:45:53.964 Removing: /var/run/dpdk/spdk_pid405717 00:45:53.964 Removing: /var/run/dpdk/spdk_pid406299 00:45:53.964 Removing: /var/run/dpdk/spdk_pid406706 00:45:53.964 Removing: /var/run/dpdk/spdk_pid407116 00:45:53.964 Removing: /var/run/dpdk/spdk_pid407635 00:45:53.964 Removing: /var/run/dpdk/spdk_pid410114 00:45:53.964 Removing: /var/run/dpdk/spdk_pid410394 00:45:53.964 Removing: /var/run/dpdk/spdk_pid414695 00:45:53.964 Removing: /var/run/dpdk/spdk_pid414754 00:45:53.964 Removing: /var/run/dpdk/spdk_pid418107 00:45:53.964 Removing: /var/run/dpdk/spdk_pid420715 00:45:53.964 Removing: /var/run/dpdk/spdk_pid427570 00:45:53.964 Removing: /var/run/dpdk/spdk_pid428022 00:45:53.964 Removing: /var/run/dpdk/spdk_pid430402 00:45:53.964 Removing: /var/run/dpdk/spdk_pid430672 00:45:53.964 Removing: /var/run/dpdk/spdk_pid433215 00:45:53.964 Removing: /var/run/dpdk/spdk_pid436984 00:45:53.964 Removing: /var/run/dpdk/spdk_pid439033 00:45:53.964 Removing: /var/run/dpdk/spdk_pid445511 00:45:53.964 Removing: /var/run/dpdk/spdk_pid451213 00:45:53.964 Removing: /var/run/dpdk/spdk_pid452390 00:45:53.964 Removing: /var/run/dpdk/spdk_pid453049 00:45:53.964 Removing: /var/run/dpdk/spdk_pid463199 00:45:53.964 Removing: /var/run/dpdk/spdk_pid465356 00:45:53.964 Removing: /var/run/dpdk/spdk_pid467350 00:45:53.964 Removing: /var/run/dpdk/spdk_pid472397 00:45:53.964 Removing: /var/run/dpdk/spdk_pid472490 00:45:53.964 Removing: /var/run/dpdk/spdk_pid475418 00:45:53.964 Removing: /var/run/dpdk/spdk_pid476813 00:45:53.964 Removing: /var/run/dpdk/spdk_pid478215 00:45:53.964 Removing: /var/run/dpdk/spdk_pid478956 00:45:53.964 Removing: /var/run/dpdk/spdk_pid480959 00:45:53.964 Removing: /var/run/dpdk/spdk_pid481845 00:45:53.964 Removing: /var/run/dpdk/spdk_pid487182 00:45:53.964 Removing: /var/run/dpdk/spdk_pid487520 00:45:53.964 Removing: /var/run/dpdk/spdk_pid487910 00:45:53.964 Removing: /var/run/dpdk/spdk_pid489467 00:45:53.964 Removing: /var/run/dpdk/spdk_pid489867 00:45:53.964 Removing: /var/run/dpdk/spdk_pid490232 00:45:53.964 Removing: /var/run/dpdk/spdk_pid492595 00:45:53.964 Removing: /var/run/dpdk/spdk_pid492604 00:45:53.964 Removing: /var/run/dpdk/spdk_pid494070 00:45:53.964 Removing: /var/run/dpdk/spdk_pid494465 00:45:53.964 Removing: /var/run/dpdk/spdk_pid494563 00:45:53.964 Clean 00:45:54.222 18:01:35 -- common/autotest_common.sh@1453 -- # return 0 00:45:54.222 18:01:35 -- spdk/autotest.sh@389 -- # timing_exit post_cleanup 00:45:54.222 18:01:35 -- common/autotest_common.sh@732 -- # xtrace_disable 00:45:54.222 18:01:35 -- common/autotest_common.sh@10 -- # set +x 00:45:54.222 18:01:35 -- spdk/autotest.sh@391 -- # timing_exit autotest 00:45:54.222 18:01:35 -- common/autotest_common.sh@732 -- # xtrace_disable 00:45:54.222 18:01:35 -- common/autotest_common.sh@10 -- # set +x 00:45:54.222 18:01:35 -- spdk/autotest.sh@392 -- # chmod a+r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/timing.txt 00:45:54.222 18:01:35 -- spdk/autotest.sh@394 -- # [[ -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/udev.log ]] 00:45:54.222 18:01:35 -- spdk/autotest.sh@394 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/udev.log 00:45:54.222 18:01:35 -- spdk/autotest.sh@396 -- # [[ y == y ]] 00:45:54.222 18:01:35 -- spdk/autotest.sh@398 -- # hostname 00:45:54.222 18:01:35 -- spdk/autotest.sh@398 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk -t spdk-gp-11 -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_test.info 00:45:54.480 geninfo: WARNING: invalid characters removed from testname! 00:46:26.618 18:02:06 -- spdk/autotest.sh@399 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -a /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_base.info -a /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_test.info -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:46:29.154 18:02:10 -- spdk/autotest.sh@400 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info '*/dpdk/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:46:32.472 18:02:13 -- spdk/autotest.sh@404 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info --ignore-errors unused,unused '/usr/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:46:35.027 18:02:16 -- spdk/autotest.sh@405 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info '*/examples/vmd/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:46:38.325 18:02:19 -- spdk/autotest.sh@406 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info '*/app/spdk_lspci/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:46:41.621 18:02:22 -- spdk/autotest.sh@407 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info '*/app/spdk_top/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:46:44.163 18:02:25 -- spdk/autotest.sh@408 -- # rm -f cov_base.info cov_test.info OLD_STDOUT OLD_STDERR 00:46:44.163 18:02:25 -- spdk/autorun.sh@1 -- $ timing_finish 00:46:44.163 18:02:25 -- common/autotest_common.sh@738 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/timing.txt ]] 00:46:44.163 18:02:25 -- common/autotest_common.sh@740 -- $ flamegraph=/usr/local/FlameGraph/flamegraph.pl 00:46:44.163 18:02:25 -- common/autotest_common.sh@741 -- $ [[ -x /usr/local/FlameGraph/flamegraph.pl ]] 00:46:44.163 18:02:25 -- common/autotest_common.sh@744 -- $ /usr/local/FlameGraph/flamegraph.pl --title 'Build Timing' --nametype Step: --countname seconds /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/timing.txt 00:46:44.163 + [[ -n 16232 ]] 00:46:44.163 + sudo kill 16232 00:46:44.174 [Pipeline] } 00:46:44.185 [Pipeline] // stage 00:46:44.189 [Pipeline] } 00:46:44.200 [Pipeline] // timeout 00:46:44.204 [Pipeline] } 00:46:44.218 [Pipeline] // catchError 00:46:44.222 [Pipeline] } 00:46:44.235 [Pipeline] // wrap 00:46:44.240 [Pipeline] } 00:46:44.250 [Pipeline] // catchError 00:46:44.257 [Pipeline] stage 00:46:44.259 [Pipeline] { (Epilogue) 00:46:44.271 [Pipeline] catchError 00:46:44.273 [Pipeline] { 00:46:44.284 [Pipeline] echo 00:46:44.285 Cleanup processes 00:46:44.289 [Pipeline] sh 00:46:44.577 + sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:46:44.577 506817 sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:46:44.590 [Pipeline] sh 00:46:44.876 ++ sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:46:44.876 ++ grep -v 'sudo pgrep' 00:46:44.876 ++ awk '{print $1}' 00:46:44.876 + sudo kill -9 00:46:44.876 + true 00:46:44.889 [Pipeline] sh 00:46:45.171 + jbp/jenkins/jjb-config/jobs/scripts/compress_artifacts.sh 00:46:57.376 [Pipeline] sh 00:46:57.666 + jbp/jenkins/jjb-config/jobs/scripts/check_artifacts_size.sh 00:46:57.666 Artifacts sizes are good 00:46:57.682 [Pipeline] archiveArtifacts 00:46:57.691 Archiving artifacts 00:46:58.243 [Pipeline] sh 00:46:58.530 + sudo chown -R sys_sgci: /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:46:58.546 [Pipeline] cleanWs 00:46:58.556 [WS-CLEANUP] Deleting project workspace... 00:46:58.556 [WS-CLEANUP] Deferred wipeout is used... 00:46:58.564 [WS-CLEANUP] done 00:46:58.566 [Pipeline] } 00:46:58.583 [Pipeline] // catchError 00:46:58.596 [Pipeline] sh 00:46:58.881 + logger -p user.info -t JENKINS-CI 00:46:58.890 [Pipeline] } 00:46:58.904 [Pipeline] // stage 00:46:58.910 [Pipeline] } 00:46:58.924 [Pipeline] // node 00:46:58.929 [Pipeline] End of Pipeline 00:46:58.965 Finished: SUCCESS